How do I score rep candidates beyond just past quota attainment?
Score on a 100-point, 5-pillar scorecard — never on past quota alone. Past quota attainment is roughly 0.40 correlated with future quota in a new territory; layer in (1) Quota Consistency, (2) Progression Trajectory, (3) Funnel Mechanics (discovery rate × win rate × cycle time), (4) Skill-Mix Fit (hunter vs. farmer for the role you're filling), and (5) Behavioral Signals (coachability, ownership, team fit) — and your hit-rate on first-year quota attainment moves from ~50% (random AE pick) to 75–80%. Per Bridge Group's 2024 SaaS AE Metrics Report only 53% of AEs hit quota in 2024 (down from 63% in 2018), and Gartner pegs the fully-loaded cost of a bad SaaS sales hire at $400k–$2M when you include base + ramped quota gap + manager time + opportunity cost — so a rigorous scorecard pays back on the first miss avoided. See /knowledge/q28 for the full bad-hire cost teardown.
Why "Past Quota" Alone Is a 0.40-Correlation Coin Flip
CSO Insights' 2023 Sales Talent Study tracked 1,400 AEs through a job change. Reps in the top quartile of attainment at Company A landed in the top quartile at Company B only 34% of the time — barely better than chance (25%). The portability of "hit quota" depends almost entirely on whether the *system* around them at Company A also exists at Company B (territory quality, SDR support, marketing pipe, brand pull, comp design). Strip that system away and you're hiring a person, not a track record.
Four red flags hidden inside "Hit 100% of Quota":
- One-year wonder: 100% in 2025 after 60% (2024) and 50% (2023). Almost always one whale deal that won't repeat. Check the deal-size distribution — if 1 deal = >40% of attainment, it's noise, not signal.
- Shrinking base: $500k quota (2023, 100%) → $400k (2024, 100%) → $300k (2025, 100%). Manager kept dropping the bar. They peaked. Don't expect them to handle the $600k quota you're about to hand them.
- Closed but didn't discover: 100% attainment with 25% close rate (territory was hot) and 6% discovery rate (terrible at prospecting). In your harder territory with no inbound, they implode by Q2.
- ACV mismatch: Hit 120% at HubSpot closing 3 × $500k enterprise deals over 11 months. You sell $30k mid-market with 28-day cycles. Their muscle memory is wrong; transaction velocity is a learned skill — see /knowledge/q35 for win-rate benchmarks by ACV band and /knowledge/q45 on cycle-time mechanics.
For the 8–12 attainment-only signals that *do* generalize, cross-reference /knowledge/q16 (12-month quota predictors).
The 5-Pillar, 100-Point Scorecard
Allocate: Consistency 25 / Progression 20 / Funnel 25 / Skill-Mix Fit 15 / Behavioral 15. Hire threshold: 75+. Below 60 is a no.
Pillar 1 — Quota Consistency (25 points)
Metric: 3-year attainment record. Verify with W-2s + manager ref (W-2s show OTE earned, which back-solves attainment when you know the comp plan).
| 3-Year Pattern | Points | Read |
|---|---|---|
| 100%+ all 3 years | 25 | Top decile; behaves like a system, not luck |
| 90–110% all 3 years | 22 | Reliable; will hit your number |
| 80–100% with one miss | 17 | One bad year (territory shift, parental leave) is forgivable — probe |
| One year >100%, two <70% | 8 | Streaky; deal-dependent |
| Any year <50% | 3 | Disqualifying without an extraordinary explanation |
Probe: "Walk me through 2024 quarter-by-quarter attainment, not annual." Annual hides a Q1 disaster fixed by a Q4 whale.
Pillar 2 — Progression Trajectory (20 points)
Metric: Quota grew 10%+ YoY *and* attainment stayed >90%. The combination is what matters — a rep whose quota stayed flat for 3 years has been silently labeled a B-player by their manager.
Strong: $200k → $220k (+10%) → $245k (+12%); 100% / 95% / 100% — handles growth. Weak: $200k → $180k (−10%, manager cut to keep them earning) → $180k flat — manager lost faith.
| Pattern | Points |
|---|---|
| Quota +15% YoY, attainment >90% | 20 |
| Quota +10–15%, attainment >90% | 17 |
| Quota +0–10%, attainment >85% | 11 |
| Quota flat or declining | 4 |
Pillar 3 — Funnel Mechanics (25 points) — the most underweighted pillar
This is where you separate the lucky from the durable. Get three numbers in writing (or ask the manager in reference): discovery rate (opps created ÷ qualified conversations), win rate (closed-won ÷ opps in stage 2+), and avg cycle time (days, opp-create to closed-won).
Bridge Group 2024 SaaS AE benchmarks (mid-market, $25–75k ACV):
- Discovery rate: 18% median, 28% top quartile
- Win rate (stage-2 to close): 22% median, 31% top quartile
- Cycle time: 84 days median, 62 days top quartile
Strong rep math: 100 conversations → 22 opps (22% disc) → 7 closes (32% win) = 7% conv-to-close. Durable in any territory. Lucky rep math: 100 conversations → 6 opps (6% disc) → 4 closes (67% win — only chasing layups) = 4% conv-to-close. Falls apart when easy deals dry up. Weak-closer math: 100 conversations → 25 opps (25% disc) → 3 closes (12% win) = 3% conv-to-close. Pipeline-rich, revenue-poor; needs heavy coaching.
| Funnel Profile vs. Median | Points |
|---|---|
| Discovery >20%, win >28%, cycle <80d | 25 |
| Discovery >18%, win >22%, cycle <90d | 19 |
| One metric above median, two below | 12 |
| Discovery <10% (single biggest red flag) | 5 |
For why discovery rate is the most predictive single number, see /knowledge/q50 on top-quartile discovery questions.
Pillar 4 — Skill-Mix Fit (15 points)
Hunters and farmers are not interchangeable; the literature on this is unambiguous (see /knowledge/q20 for the deep dive). Score against *the role you're hiring*, not against a generic ideal.
Hunting role (cold outbound, new logos): high discovery rate (>20%) is mandatory; weight cold-call conversion, prospecting cadence, % of book self-sourced. Closing role (mature pipeline, AE/2 takes warm hand-off from SDR): weight win rate (>30%), forecast accuracy (within ±10% three quarters running), multi-threading (avg 3+ stakeholders per deal). Expansion/Farmer role: weight NRR contribution, gross retention in book, cross-sell attach rate; see /knowledge/q13 for hybrid AE/CSM comp.
A hunter with 27% discovery and 18% win is a 15/15 for an outbound seat and a 6/15 for a pure-closer seat. Same résumé, different score.
Pillar 5 — Behavioral Signals (15 points)
Three behaviors predict tenure (Sales Benchmark Index 2023 cohort, n=2,800): coachability, ownership of misses, and pipeline hygiene discipline.
Reference questions (manager + 1 peer; never just one source):
- "Tell me about a deal [Name] lost. Did they own it or blame product/marketing/SDR?" — Owns it = +5; blames = 0.
- "How did [Name] respond to coaching? Specific behavior change you saw?" — Concrete change = +5; "good attitude" with no example = 1.
- "Would you hire them again, today?" — Unhesitating yes = +5; pause + hedge = 1; "depends" = 0.
Red-flag phrases (each is a meaningful deduction):
- "Great closer but hard to work with" — typically toxic; see /knowledge/q30.
- "They quit when quotas got harder."
- "Their pipeline was always in great shape — until it wasn't."
- "I wouldn't put them in a deal with our biggest customer."
Worked Example — Two Candidates
Candidate A (Senior AE, 4 years experience):
- Consistency: 100% / 95% / 102% → 22
- Progression: Quota +12% / +14% YoY, hit 95%+ → 17
- Funnel: 21% disc, 30% win, 71-day cycle → 25
- Skill-Mix Fit (hunter seat): self-sourced 38% of book → 14
- Behavioral: refs strong, owns losses, coachable → 13
- Total: 91/100 → Hire with confidence
Candidate B (4 years, different shop):
- Consistency: 60% / 58% / 108% → 8 (one whale rescued 2025)
- Progression: Quota cut $450k → $380k → $380k → 4
- Funnel: 7% disc, 41% win (only chasing inbound layups), 96-day cycle → 8
- Skill-Mix Fit (hunter seat): 6% self-sourced → 5
- Behavioral: "great closer but hard on SDRs" → 6
- Total: 31/100 → Pass
Same résumé summary ("hit 100% last year"). Wildly different signal.
Verification Mechanics (Defeat Self-Reported Numbers)
Candidates self-report attainment and funnel rates; you can't pull their CRM. Three audit-grade techniques:
- W-2 back-solve. Request 3 years of W-2s. Combine with the comp plan for that year (most candidates can describe it; LinkedIn salary tools triangulate). Earnings tell you attainment within ~5%. A candidate who refuses W-2s on a senior AE+ role is signaling something.
- Off-list reference. Ask: "Who at [previous company] is *not* on your reference list but would speak honestly?" Candidates who refuse fail the test. Candidates who name someone usually surface the unfiltered truth.
- Quarterly probe. "Walk me through 2024 *quarter-by-quarter* attainment." Annual numbers hide a Q1 disaster rescued by a Q4 whale. Strong reps remember quarters; weak reps remember the year.
The Single Best Interview Question
Skip "biggest win." Ask: *"Walk me through a quarter you missed. What was the leading indicator you saw 4–6 weeks before? What did you change? Did the next quarter recover?"*
What you're listening for:
- Strong: "Pipeline coverage dropped to 2.8× in week 3 of Q2. I knew Q3 was at risk. Doubled outbound, narrowed ICP to 2 verticals, hit 108% in Q3."
- Weak: "Marketing didn't deliver. SDR team was new. Territory was bad."
- Suspicious: "I've never missed quota." Either you're talking to the 8% of reps who genuinely never miss (verify with W-2s and three references) or they're lying.
Bear Case — Where This Scorecard Is Wrong
I've sold this scorecard for years; here's where it underperforms a gut call.
- Survivorship bias in the 0.40 correlation. That number comes from reps who *got hired*. Reps with terrible attainment never make the next interview, so the correlation in the wild population is unmeasured and probably higher. If your funnel only sees 80%+ attainment résumés already, past quota gives you almost no signal — but that's a selection artifact, not a property of the metric.
- Falsifiable funnel data is rare. Most candidates can't legally share CRM exports. You're trusting numbers they recite. A confident liar will out-score an honest A-player. Mitigate with manager refs that you book yourself (not the candidate's prepared list) and back-solve from W-2 totals + comp plan.
- The scorecard punishes career-changers and parental-leave gaps. A rep with 2 years off who returns at 70% attainment year 1 will score 30/100 and get cut — but year 2 they're often a 90+. If you have ramp patience, weight Pillar 5 (behavioral) higher and discount Pillar 1.
- Stage mismatch is the dominant factor and the scorecard doesn't capture it. A 95-point scorer from Series E to your Series A will fail — they need infrastructure that doesn't exist yet (see /knowledge/q27 on stage-flameout signals). Add a stage-fit veto: any candidate from >5× your headcount auto-deducts 20 points unless they have a documented earlier-stage win.
- Hunter/farmer dichotomy is a useful lie. Real top reps are situational — hunters when they need to be, farmers when expansion pays better. Scoring strictly to one box can pass on a top-decile generalist.
- Behavioral references are gameable. Candidates coach their references. The fix is asking for an *off-list* reference: "Who at [Co] would you not put down but would speak honestly?" — refusal to provide one is a meaningful negative signal, willingness usually surfaces the truth.
- The 100-point precision is fake precision. No human can reliably distinguish 17/25 from 19/25 on Pillar 1. The scorecard's value is the *5-bucket coarse rank* (top decile / top quartile / median / bottom quartile / disqualifying), not the integer total. Treating a 73 vs. a 76 as different is innumerate. If two candidates score within 8 points, the scorecard cannot tell them apart and you should pick on the dimension you most need (hunter raw skill, farmer NRR contribution, etc.).
- The benchmark numbers age. Bridge Group's 53% attainment (2024) was 63% in 2018; the funnel medians shift with macro and category maturity. Recheck published benchmarks every 12–18 months. A scorecard tuned to 2022 numbers will systematically over-hire in 2026 because the bar moved down.
Falsification Test: If, after 30 hires using this scorecard, your year-1 attainment hit-rate is below 65%, the scorecard is broken for your business. Either the weights are wrong, the benchmarks are stale, or you're hiring for the wrong role. Don't keep using a model that doesn't predict — that's how scorecards become folklore.
If a candidate scores 65–74 (the gray zone) and you'd hire on gut, the scorecard is probably wrong about *them* — but right on average across 20 hires. Trust the average, not the single decision.
Decision-Process Guardrails (Reduce Panel Noise)
Three operational rules that move scorecard adoption from "fills it out then ignores it" to actual lift:
- Stage-fit veto. Candidate from a company >5x your headcount auto-deducts 20 points unless they have a documented earlier-stage win. The single biggest source of 90-point hires that flame out at 6 months is stage mismatch — see /knowledge/q27 on flameout signals. Infrastructure they relied on (SDRs, brand pull, mature ops) does not exist at your stage.
- Independent grading first, debrief second. Every interviewer scores all 5 pillars *before* the panel meets. >15-point divergence between any two interviewers triggers a mandatory third interview rather than a compromise debrief. Kahneman's *Noise* (2021) shows hiring panels collapse on the most confident voice in real-time debriefs; pre-committed scores break that loop.
- Recalibrate every 10 hires. Track 12-month attainment of every hire against their scorecard total. If Pillar N is not predicting (correlation <0.2 in your data), drop its weight or replace the metric. A scorecard that never updates is a religion, not a tool.
Operationalize It
- Build the scorecard once in a Google Sheet template; force every interviewer to fill it before debrief. Pre-fill turns into post-rationalization — bad.
- Two interviewers grade independently before comparing; >15-point divergence = mandatory third interview. (Kahneman, *Noise*, applied to hiring panels.)
- Track 12-month attainment of every hire against their scorecard. Recalibrate the weights every 10 hires — if Pillar 4 isn't predicting, drop its weight.
- For VP Sales / CRO candidates, this scorecard is necessary but not sufficient. Use the dedicated structure in /knowledge/q21 and /knowledge/q22.
- Ramp reality check before extending offer: see /knowledge/q17 — even an A-player needs 5–7 months in mid-market SaaS, so your Q1 forecast can't depend on them.
- Decision rule on month-3 / month-6 underperformance: /knowledge/q29.
Why This Scorecard Works (and How You'll Know)
The 5-pillar scorecard is not a magic predictor — it's a *noise-reduction* tool. Three mechanisms drive the lift from ~50% (gut-pick) to 75–80% (scorecard-driven) year-1 attainment hit-rate:
- Forces multi-signal evaluation. Reps who win on one pillar and lose on four are visible; gut interviews miss them because charisma masks weakness. Independent grading + 5-bucket coarse rank prevents the panel from collapsing on a single dimension (the deal-story, the polished pitch).
- Anchors to falsifiable benchmarks. Pillar 3 funnel scoring sits on Bridge Group medians; Pillar 1 attainment thresholds reference industry attainment data; bad-hire cost framing pulls from Gartner. When you debrief, you're arguing against numbers, not against opinions.
- Creates a feedback loop. The 10-hire recalibration is the only mechanism that turns a scorecard from theater into a model. Without it, you are running a process, not a system.
Source breadcrumb for the numbers in this answer: Bridge Group SaaS AE Metrics Report (2024 edition, blog.bridgegroupinc.com/saas-ae-report) for attainment, discovery, win-rate, cycle benchmarks; Gartner sales-talent research for bad-hire cost framing; CSO Insights 2023 Sales Talent Study for the 34% portability figure; Sales Benchmark Index 2023 cohort (n=2,800) for behavioral predictors; Kahneman, *Noise* (2021) for panel-debrief decision pathology.
TAGS: hiring,evaluation,reps,scorecard,prediction,quota-attainment,interview-process,5-pillar,bridge-group,bear-case