Building a Quick Estimation Template When You Have Almost Nothing to Go On
The 4 PM Estimation Crisis
Last month, my PM dropped 73 backlog items on my desk at 4 PM. "Need estimates by tomorrow's planning meeting." Sound familiar?
Here's what I did: I grabbed the scariest item—a machine learning integration nobody on the team had touched before. I rated it:
- Innovation: 8 (we've never done ML)
- Scope: 7 (new infrastructure, data pipelines, APIs)
- People: 7 (data scientists, backend devs, DevOps, plus external vendors)
That's 22 points. With my multiplier of 5 for high complexity, I got 110 → mapped to 89 story points.
Took me 90 seconds. By 6 PM, I'd estimated all 73 items. Here's the framework I built and how you can adapt it for your team.
The Framework
I needed something simple: rate each factor on a 1-10 scale, add them up, apply a multiplier. That's it.
Addition gives a manageable range (3-30) vs. multiplication (1-1000). I can explain it in 30 seconds and calculate it in my head.
(Innovation + Scope + People) × Multiplier = Estimate
Task: Rate 8+7+7 = Sum 22, Apply ×5 = Result 110 points
The Three Core Factors (Deep Dive)
Innovation: Have we done this before?
- 1-3: Routine work with templates (login forms, standard bug fixes)
- 4-6: Similar work needing adaptation (new payment provider, different framework)
- 7-10: New territory requiring R&D (first ML model, blockchain integration)
Ask: "How much Googling vs. copy-pasting from previous work?"
Scope: How big is the work?
- 1-3: Small, isolated change (button color, single validation)
- 4-6: Multiple components (dashboard page, API + form)
- 7-10: System-wide changes (database migration, entire microservice)
Ask: "How many files/systems will we touch?"
People: Coordination complexity?
- 1-3: Solo or single team
- 4-6: 2-3 different skill sets with handoffs
- 7-10: Many teams, diverse skills, external vendors
Ask: "How many different specialties and integration points?"
Multipliers Based on Complexity
Complexity doesn't scale linearly. A sum of 5 might double in effort, but a sum of 22 explodes due to context switching, integration testing, stakeholder delays, and communication overhead.
| Sum Range | Multiplier | Why |
|---|---|---|
| 3-9 | ×2.5 | Small tasks, minimal overhead |
| 10-15 | ×4 | Moderate coordination |
| 16-20 | ×5 | Significant complexity |
| 21+ | ×6 | Red flag - consider splitting |
Mapping to Different Formats
The framework generates consistent relative values that map to any output format. Working in Agile? Map to Fibonacci. T-shirt sizes? Create bands. Need hours? Use a conversion factor (but keep it hidden from stakeholders).
The mapping can be a simple lookup table or nested IF statements. Let AI write these tedious formulas for you.
Fibonacci Mapping
Example conversion:
- Sum 3-6 → 3 points
- Sum 7-10 → 5 points
- Sum 11-14 → 8 points
- Sum 15-18 → 13 points
- Sum 19-22 → 21 points
Mapping to Output Formats
The raw score maps to any format your team uses:
- Fibonacci: 3-6→3, 7-10→5, 11-14→8, 15-18→13, 19-22→21
- T-Shirt: XS (3-6), S (7-9), M (10-13), L (14-17), XL (18-21), XXL (22+)
Getting Your Team On Board
Week 1: Estimate your next sprint solo. Track actuals and calculate variance.
Week 2: Share results with the team. Walk through examples and ask what they would've rated differently.
Week 3: Have the team rate 5 items independently, then compare and discuss differences.
Common Pushback:
🗨️ "This is too simple"
→ "Try it for one sprint. Track accuracy. If it works, keep using it. If not, we'll adjust."
🗨️ "My tasks are too unique"
→ "Then rate them as high Innovation. The framework accommodates uniqueness—that's why we have a 1-10 scale."
Week 2: Share results with the team. Walk through examples and ask what they would've rated differently.
Week 3: Have the team rate 5 items independently, then compare and discuss differences.
Common Pushback:
- "This is too simple" → Try it for one sprint and track accuracy
- "My tasks are unique" → Rate them as high Innovation, that's what the scale is for
- "People will game it" → Calibration exposes inflated ratings when actuals don't match
- "We use Planning Poker" → Use this for initial estimates, then validate with Planning Poker
The Critical Lesson
Don't rate based on the task description. Rate based on YOUR TEAM'S current reality. The framework is only as good as your honesty when rating. If your expert left, adjust Innovation. If "simple" hides complexity, adjust Scope. If coordination is messy, adjust People.
Focus on Defendable, Not Perfect
The goal isn't perfect estimates—it's estimates you can explain and adjust. When someone challenges your numbers, you can show:
- The specific Innovation, Scope, and People ratings
- Why you rated each factor that way
- How changing any rating affects the estimate
- Which completed work validates your approach
This transparency builds trust even when estimates are wrong.
Python Implementation
Here's a complete working implementation you can adapt:
def estimate_task(innovation, scope, people):
"""
Calculate task estimate using Innovation, Scope, People framework.
Args:
innovation (int): 1-10, have we done this before?
scope (int): 1-10, how big is the work?
people (int): 1-10, coordination complexity?
Returns:
dict: Contains sum, multiplier, raw_score, and fibonacci_points
"""
# Calculate sum
sum_factors = innovation + scope + people
# Apply progressive multipliers
if sum_factors <= 9:
multiplier = 2.5
complexity = "Low"
elif sum_factors <= 15:
multiplier = 4
complexity = "Medium"
elif sum_factors <= 20:
multiplier = 5
complexity = "High"
else:
multiplier = 6
complexHistory
Work backwards from 5-10 completed items:
- Rate a completed item (Innovation + Scope + People)
- Calculate: Actual Effort ÷ Sum = Implied Multiplier
- Average across several items
- That's your team's baseline multiplier
No history? Start with ×3 and adjust after a few completions
How This Compares to Other Methods
| Method | Setup Time | Estimation Speed | Team Buy-in | Accuracy | Best For |
|---|---|---|---|---|---|
| Planning Poker | 30 min | 5 min/item | High | Medium | Small backlogs, team building |
| This Framework | 10 min | 90 sec/item | Medium | Medium-High | Large backlogs, rapid estimation |
| Expert Judgment | None | Varies | Low | Low-Medium | Quick guesses only |
| Story Points (Gut) | 5 min | 2 min/item | Low | Low | When nothing else works |
| Historical Analysis | 2+ hours | 10 min/item | Medium | High | Similar, well-documented work |
Key insight: This framework sits in the sweet spot—faster than Planning Poker, more structured than gut feel, and doesn't require extensive historical data.
After six months of 10% accuracy, I got cocky. A "simple" mobile redesign: Innovation 3 (done mobile before), Scope 5 (just UI), People 2 (one team). Sum of 10, ×4 multiplier = 40 hours.
It took 6 weeks.
What I missed:
- Our mobile dev had left—should've been Innovation 7, not 3
- "UI changes" required rewriting state management—Scope 8, not 5
- "One team" was designer + junior dev + consultant—People 5, not 2
Rate based on your team's current reality, not the task description. Key insight: This framework sits in the sweet spot—faster than Planning Poker, more structured than gut feel, and doesn't require extensive historical data.
Build in Automatic Reality Checks
Add a column for "actual effort" and calculate variance automatically. Use conditional formatting to highlight when estimates are off by more than 50%.
Track patterns:
- Are innovative tasks always underestimated?
- Does your team consistently underrate People complexity?
- Do certain types of work always blow up?
These patterns help you adjust ratings going forward, not just multipliers. The framework improves as you learn your team's blindspots.
Keep Cognitive Load Low
SpDefendable Over Perfect
The goal isn't perfect estimates—it's estimates you can explain. When challenged, show the specific Innovation/Scope/People ratings and why you chose them. Transparency builds trust even when you' | Use AI to Handle Tedious Parts
AI assistants excel at:
- Generating Excel formulas for your ISP logic
- Creating validation rules and conditional formatting
- Building reference tables and lookup functions
- Writing calibration calculations
- Producing multiple format outputs from the same data
Don't waste time on formula syntax. Describe what you want and let AI write the implementation. See the LLM Prompts section below for specific examples.
Embrace Continuous Refinement
Your first version will be wrong. The framework gives you something to be wrong WITH, which is infinitely better than being wrong without structure.
After each sprint/milestone/project:
- Compare estimates to actuals
- Look for systematic bias in Innovation, Scope, or People ratings
- Adjust either rating definitions or multipliers
- Document what you learned
Within 3-4 cycles, your accuracy improves dramatically. The framework evolves with your team's reality.
The Psychology Matters
Having a framework changes the conversation from "how did you guess that?" to "let's discuss these ratings." It moves you from defending random numbers to discussing specific complexity factors.
Stakeholders can engage with "I rated People complexity as 7 because we need frontend, backend, DBA, and DevOps coordination" in a way they can't with "my gut says 3 weeks."
Transparency is a Feature, Not a Bug
Be upfront that this is a rapid estimation tool based on simple math. Don't pretend it's more sophisticated than it is. When presenting estimates, show the Innovation + Scope + People ratings openly and explain it takes 30 seconds per item. Honesty builds credibility.
LLM Prompts to Get You Started
AI Prompts to Build Your Framework
Copy these prompts to generate implementation details:
Initial Framework:
"Create an Excel formula that takes 3 factors (Innovation, Scope, People) rated 1-10, adds them, applies progressive multipliers, and maps to Fibonacci."
Spreadsheet Template:
"Generate a Python script using openpyxl for an estimation template with headers, data validation (1-10), automatic calculations, and conditional formatting."
Calibration Formula:
"Write an Excel formula that calculates implied multiplier from actuals and provides an average recommendation."
| Method | Speed | Accuracy | Best For |
|---|---|---|---|
| Planning Poker | 5 min/item | Medium | Small backlogs |
| This Framework | 90 sec/item | Medium-High | Large backlogs |
| Expert Judgment | Varies | Low-Medium | Quick guesses |
| Historical Analysis | 10 min/item | High | Similar work only |
Tips for Success
Track patterns: Are innovative tasks always underestimated? Do you underrate People complexity? Adjust future ratings based on blindspots.
Keep it fast: More than 2 minutes per estimate means it's too complex. You should estimate 50 items in under an hour.
Use AI for formulas: Let AI generate Excel formulas, validation rules, and lookup tables. Don't waste time on syntax.
Refine continuously: Your first version will be wrong. Compare estimates to actuals after each sprint and adjust. Within 3-4 cycles, accuracy improves dramatically.
Be transparent: Show the ratings openly. It changes conversations from "how did you guess?" to "let's discuss these factors."
Real Calculation Examples
Let's walk through exactly how the math works with concrete numbers:
Example 1: Simple Bug Fix
- Innovation: 2 (we fix similar bugs weekly)
- Scope: 2 (single component affected)
- People: 1 (one developer, no coordination)
- Sum: 5
- Multiplier: ×2 (low complexity)
- Raw Score: 10
- Result: → 8 points
Example 2: Customer Portal Feature
- Innovation: 5 (done similar, has new elements)
- Scope: 6 (multiple screens, DB, API)
- People: 4 (frontend, backend, UX, PO)
- Sum: 15
- Multiplier: ×4 (medium-high)
- Raw Score: 60
- Result: → 55 points
Example 3: ML Integration
Copy these prompts to generate implementation details:
Initial Framework:
"Create an Excel formula that takes 3 factors (Innovation, Scope, People) rated 1-10, adds them, applies progressive multipliers, and maps to Fibonacci."
Spreadsheet Template:
"Generate a Python script using openpyxl for an estimation template with headers, data validation (1-10), automatic calculations, and conditional formatting."
Calibration Formula:
"Write an Excel formula that calculates implied multiplier from actuals and provides an average recommendation.
What I've Learned
The perfect estimation framework doesn't exist. But having a simple, transparent, adjustable approach beats guessing every time. I built mine with AI help, implemented it immediately, and keep refining it with every project.
Here's what surprised me: You don't need much information to start estimating systematically. You just need a consistent way to capture what little you know. Three factors, simple math, and regular calibration have gotten me far.
(Innovation + Scope + People) × Multiplier = Estimate
Remember: The goal isn't to predict the future perfectly. It's to be consistently wrong in a way you can measure and correct. ul.mb-4 li Overcomplicating factors: Don't split Innovation into subcategories. Keep it simple. li Confusing Scope with People: A large migration (Scope 8) might only need one DBA (People 2). li Rating by industry standards: Innovation means "have WE done this?"—not whether it exists. li Using multiplication: Innovation × Scope × People gives 1-1000. Addition gives 3-30 (much better). li Ignoring skill diversity: Three backend devs = low People. Backend + frontend + DBA = high People.
Example Calculations
| Task | I | S | P | Sum | Mult | Result |
|---|---|---|---|---|---|---|
| Bug Fix | 2 | 2 | 1 | 5 | ×2.5 | 13 |
| Portal Feature | 5 | 6 | 4 | 15 | ×4 | 60 |
| ML Integration | 8 | 7 | 7 | 22 | ×5 | 110 |
| OAuth | 8 | 5 | 6 | 19 | ×5 | 95 |