← All Blog Articles

Weighted Scoring Is Subjectivity Wrapped in a Veneer of Objectivity

Why Bespoke Criteria Usually Produce Score-Gaming Theatre

(updated Jan 24, 2026)
Weighted Scoring Is Subjectivity Wrapped in a Veneer of Objectivity

This is one of RoadmapOne ’s articles on Objective Prioritisation frameworks .

Weighted Scoring sounds rigorous. You define criteria that matter to your organisation—Strategic Fit, Customer Value, Technical Feasibility, Revenue Impact. You assign weights reflecting relative importance—Strategic Fit 30%, Customer Value 25%, and so on. You score each objective against each criterion, multiply by weights, sum the totals. Out comes a ranked backlog, customised to your unique context.

It’s subjectivity wrapped in a veneer of objectivity.

I’ve watched this play out repeatedly—in product prioritisation, in RFP evaluations, in strategic planning workshops. The moment you introduce adjustable weights and bespoke criteria, you create a system optimised for gaming rather than clarity.

TL;DR

Weighted Scoring appeals to organisations that want “customised” prioritisation, but the customisation usually adds complexity without adding rigour. Teams spend weeks debating whether Strategic Fit should be 30% or 35% while the backlog rots unscored. PMs learn which levers to pull to inflate their projects’ scores. And when the final ranking doesn’t “feel right,” leadership overrides it anyway—at which point you’ve produced the same result as a 30-minute conversation with extra steps. For most product teams, RICE , BRICE , or ICE provide better prioritisation with less overhead.

How Weighted Scoring Works

The framework follows a standard pattern:

  1. Define criteria. What dimensions matter for prioritisation? Common choices include Strategic Alignment, Customer Value, Revenue Impact, Technical Feasibility, Risk, Time-to-Market.

  2. Assign weights. How important is each criterion relative to the others? Weights typically sum to 100%. Strategic Alignment might get 30%, Customer Value 25%, Revenue Impact 20%, Technical Feasibility 15%, Risk 10%.

  3. Score each objective. Rate every backlog item against each criterion, typically on a 1-5 or 1-10 scale.

  4. Calculate weighted scores. For each objective: (Criterion1 Score × Weight1) + (Criterion2 Score × Weight2) + … = Total Score.

  5. Rank by total score. Highest scores win capacity allocation.

The appeal is obvious: you can tailor the framework to your specific context. A regulated industry might weight Compliance Risk heavily. A growth-stage startup might weight Time-to-Market. The customisation feels strategic.

A Worked Example

Imagine three objectives competing for capacity:

Objective Strategic Fit (30%) Customer Value (25%) Revenue (20%) Feasibility (15%) Risk (10%) Total
Enterprise SSO 5 4 5 3 4 4.25
Mobile Dark Mode 2 5 2 5 5 3.45
API Rate Limiting 4 3 3 4 3 3.50

Enterprise SSO wins with 4.25. The weighted scoring has spoken.

Except… everyone in the room already knew Enterprise SSO was the priority. The exercise took two hours and produced the same ranking a quick conversation would have. The numbers create an illusion of precision—4.25 vs 3.50—but those decimals are fiction. They’re built on subjective 1-5 scores multiplied by debatable weights.

When Weighted Scoring Goes Wrong

Score Gaming Theatre

The moment you introduce weighted scoring, people reverse-engineer it. If “Strategic Alignment” carries 30% weight, suddenly every PM discovers their pet project is “highly strategically aligned.” Impact scores inflate. Risk scores deflate. The framework becomes a game where the winner is whoever best understands the scoring incentives.

The worst case: leadership sets the weights, PMs game the scoring, and leadership overrides the final ranking because “the numbers don’t feel right.” At that point you’ve wasted everyone’s time with theatre that produced the same result as trusting leadership’s judgement in the first place.

I’ve seen the same dynamic in RFP evaluations. You get vendor responses, score them against weighted criteria, and one vendor scores 72, another 70, another 68. So the evaluation committee takes only the 72-scorer to the next round—as if a 2-point difference on a subjective scale is meaningful. Then someone notices the “wrong” vendor won, so they adjust weights and tweak scores until the preferred vendor comes out on top.

It’s bad layered on bad. The weights and scores create false precision that doesn’t survive contact with actual preferences.

The Weight-Debating Trap

I’ve seen leadership teams burn entire quarterly planning cycles debating whether “Customer Satisfaction” should be weighted 25% or 30%, while the actual backlog sits unscored and teams have no idea what to work on next.

The irony is that the difference between 25% and 30% weighting rarely changes the final ranking. If your prioritisation outcome flips because of a 5% weight adjustment, your items are so close in value that you should just pick one and move on. The debate is displacement activity—it feels like strategic work but produces nothing.

Vanity Criteria

Teams love adding criteria that sound important but can’t be objectively measured:

  • “Innovation Score” — What’s a 4 vs a 5 on innovation?
  • “Brand Impact” — Nobody can define what this means
  • “Technical Excellence” — Just a proxy for “engineers like this”
  • “Strategic Alignment” — Without defined strategy, this is pure vibes

These criteria are proxies for “I like this project.” They add columns to the spreadsheet without adding rigour to the decision.

When Weighted Scoring Actually Works

Despite the above, I’ve seen weighted scoring work in specific contexts.

Platform Teams and Internal Products

A platform team focused on developer experience used weighted scoring with three dimensions: Developer Velocity Impact, Technical Debt Reduction, and Incident Risk Reduction. Those criteria made sense for their context—they weren’t shipping customer-facing features, they were improving internal systems.

BRICE would have felt forced because “Reach” and “Business Importance” don’t map naturally to internal platform work. The team needed criteria specific to their mission, and weighted scoring provided that flexibility.

Cultural Transformation Initiatives

If your organisation is explicitly investing in cultural change—improving DORA metrics, sunsetting legacy technology, shifting engineering practices—those goals don’t map cleanly to RICE dimensions. A weighted model with “Cultural Impact” as a dimension can surface initiatives that improve how teams work, not just what they ship.

Early-Stage Products with North Star Bets

A very early-stage startup pre-product-market-fit had three north star hypotheses: “10x faster than competitors,” “zero configuration required,” and “enterprise-grade security.” They scored every feature idea against those three dimensions.

It wasn’t sophisticated, but it kept them focused on differentiation rather than building whatever customers asked for. The criteria were specific to their stage and strategy—not generic “Strategic Fit” nonsense.

What Made These Work?

In every successful case:

  • Criteria were specific to context, not generic placeholders
  • Limited to 3-4 dimensions, not 8-criterion monsters
  • Everyone understood exactly what each criterion meant
  • No “Brand Impact” or “Innovation Score” vagueness

Notice anything? Those successful weighted scoring models look suspiciously like RICE or ICE with slightly different labels. Every time I’ve seen a team strip their weighted model down to “criteria that actually mean something,” they end up with 4-5 dimensions that could have been RICE from the start.

The Criteria That Actually Matter

If you strip away the vanity criteria, what’s left? The dimensions that can be meaningfully scored:

Useful Criteria Why It Works
Revenue/Business Impact Quantifiable in £/ARR
Reach Number of users/customers affected—countable
Effort Person-weeks, validated by engineering
Confidence What evidence supports our estimates?
Time Sensitivity Is there a deadline that makes delay costly?
Risk Reduction Does this mitigate a specific, identified risk?

These are basically RICE and BRICE dimensions. If your weighted scoring model uses these criteria, you’ve reinvented RICE with extra steps.

The worst offender is “Strategic Fit” as a standalone criterion. If you can’t articulate which strategy and how it fits, you’re just adding a column where everyone scores their own projects highly. That’s exactly what Business Importance in BRICE solves—it forces you to define the 2-3 strategic priorities before scoring, so “strategic fit” becomes objective rather than vibes.

Weighted Scoring vs RICE/BRICE

Dimension Weighted Scoring RICE/BRICE
Criteria Custom, often vague Predefined, battle-tested
Weights Debatable, gameable Built into the formula
Time to implement Weeks of criteria/weight debates Score in an afternoon
Gaming potential High—adjust criteria and weights to get desired result Lower—dimensions are standardised
Precision False precision from arbitrary weights Honest about estimation uncertainty

For most product teams shipping to customers, weighted scoring is fundamentally inferior to RICE, BRICE, or ICE. It creates the illusion of customisation while adding complexity that invites gaming.

Practical Advice

If your organisation insists on weighted scoring:

Timebox the weight-setting to 60 minutes maximum. Pick round numbers (20%, 30%, 50%), accept that it’s imperfect, score your backlog, and ship the roadmap. You’ll learn more from executing a “good enough” roadmap than from perfecting weights you’ll probably change next quarter anyway.

Limit to 4-5 criteria maximum. If your model has more than 5 dimensions, you’ve already lost. Complexity doesn’t equal rigour—it equals hiding behind spreadsheets.

Define criteria concretely. “Revenue Impact” needs a rubric: 5 = >£1M ARR, 4 = £500K-£1M, 3 = £100K-£500K, etc. Without concrete definitions, scores are meaningless.

Calibrate against reality. If last quarter’s “High Strategic Fit” items didn’t actually deliver strategic value, your calibration is broken. Use historical outcomes to validate that your scoring predicts actual results.

Consider whether you need this at all. If your weighted scoring model ends up with Revenue, Reach, Effort, and Confidence as the main criteria… just use RICE.

The Bottom Line

Weighted Scoring is subjectivity wrapped in a veneer of objectivity. The customisation appeals to organisations that want to feel sophisticated, but the extra complexity rarely produces better prioritisation than simpler frameworks.

For platform teams, cultural transformation, or early-stage products with specific north star dimensions—weighted scoring has a place. For everyone else, it’s usually theatre.

Use RICE when you have reach and impact data. Use BRICE when you need strategic alignment. Use ICE when you’re moving fast with incomplete data. Use PIE when you’re running experiments. These frameworks have predefined dimensions that are harder to game, faster to implement, and produce scores that mean something.

If weighted scoring appeals because you think your context is unique—it probably isn’t. The criteria that matter for product prioritisation are well-understood. Reach, Impact, Confidence, Effort, Business Importance. Pick a framework that uses them and get on with delivery.

References