The Color Team Trap: When Evaluation Ratings Don't Support the Award
Evaluation ratings show scores, but if you can't explain why that offeror won, your award decision won't survive a protest.
You're three weeks from award. The evaluation is done. Consensus is final. The ratings chart is clean: Offeror A earned Blue under Technical, Green under Past Performance. Offeror B got Green and Yellow. The numbers line up. The colors make sense. Then you sit down to draft the Source Selection Decision Document and realize something is badly wrong.
The narrative doesn't support the ratings. The strengths sound generic. The weaknesses aren't tied to anything in the RFP. And when you try to explain why Offeror A won, the best you can come up with is "because they were rated higher." That's not a tradeoff analysis. That's circular reasoning. And if this goes to protest, it will crumble under scrutiny.
This is the color team trap: when evaluation ratings become decoupled from the underlying logic that's supposed to justify them. It's one of the most common and least discussed failure modes in federal source selection. And it usually reveals itself at the worst possible moment—right before award, when there's no time to start over.
This article walks you through the reverse-engineering method that lets you audit your evaluation record, identify where the logic chain breaks down, and repair the disconnect without re-evaluating proposals. It's not about changing scores. It's about making your existing record coherent, traceable, and legally defensible.
The Ratings-Narrative Disconnect Explained
Here's what the problem looks like in practice. Your evaluation team produces a summary matrix with color-coded ratings for each offeror under each factor. Blue means Excellent. Green means Good. Yellow means Acceptable. The chart is tidy and consensus was smooth.
But when you open the actual evaluation narratives, the documented strengths and weaknesses don't clearly connect to the ratings. A strength might say "the offeror demonstrated relevant experience," but that's not a discriminator—it's table stakes. A weakness might say "the approach lacked detail in certain areas," but it doesn't cite which RFP criteria were affected or how significantly.
Worse, when you try to draft the tradeoff analysis comparing Offeror A and Offeror B, you find yourself restating the colors instead of explaining differences. "Offeror A was rated Blue and Offeror B was rated Green, therefore Offeror A provides better value." That's not analysis. That's tautology.
The real warning sign is this: you cannot answer the question "why not Offeror B?" without just pointing back to the ratings chart. If a protester asks why their Green rating wasn't good enough, and your only answer is "because someone else got Blue," you don't have a defensible record.
This disconnect happens because consensus meetings focus on agreement, not documentation. Evaluators debate the merits, settle on a rating, and move on. The narrative write-up comes later, often rushed, and it becomes a summary of conclusions rather than a traceable chain of reasoning.
Why This Disconnect Happens
Evaluation teams are typically led by technical experts, not acquisition professionals. Their job is to assess whether proposals meet the government's needs. They're very good at that. But they're not trained to build legally defensible documentation that will survive GAO review.
So what usually happens is this: the team reads proposals, discusses them in meetings, and assigns ratings based on collective judgment. Those ratings reflect real conclusions. But the reasoning behind them stays mostly verbal. It lives in the room, not on the page.
When it's time to write the evaluation narratives, evaluators often treat ratings as shorthand. They'll write "the offeror's past performance was strong," assign a Green, and assume that's enough. They don't realize that the contracting officer will later need to defend why "strong" meant Green instead of Blue, or why one offeror's Green was better than another offeror's Green.
Add time pressure, and the problem compounds. Teams are pushed to finalize consensus and move toward award. Documentation becomes a checkbox. Narratives get written quickly, often by copying and pasting proposal content with evaluative adjectives sprinkled in.
By the time the KO inherits the evaluation record, the ratings are locked. Consensus is final. And the KO is left trying to reverse-engineer a defensible decision from incomplete documentation, often with a protest clock already ticking.
What GAO and COFC Actually Scrutinize
The legal standard for a defensible source selection is straightforward: the decision must be reasonable, consistent with the solicitation's stated evaluation criteria, and documented in the record. What matters is the logic chain, not the aesthetics of your SSDD.
GAO doesn't care that your ratings chart is color-coded. They care whether the documented findings under each factor are meaningful, specific, and tied to RFP language. A strength that says "the offeror proposed a solid approach" is legally meaningless. A strength that says "the offeror's use of automated data validation reduces manual review time, which aligns with the RFP's emphasis on efficiency" is defensible.
GAO also traces how evaluation findings connect to the award decision. If your technical factor was weighted more heavily than price, and Offeror A's higher rating under technical drove the award, the record must show what specific advantages Offeor A offered and why those advantages justified the price premium.
Comparative analysis is where most protests succeed. If you rated Offeror B lower and didn't select them, you must explain what discriminators led to that rating difference. You can't just restate the rating. You have to show what Offeror A did better, link it to RFP priorities, and explain why that mattered to the government's decision.
Think of it like a math proof. You can't just write the answer. You have to show your work. Every step must follow logically from the one before it. If GAO can't trace your reasoning from solicitation language to evaluation finding to award rationale, the decision is legally vulnerable.
The Reverse-Engineering Method
The good news is that you don't necessarily have to start over. The goal is not to re-evaluate proposals or change scores. The goal is to make the existing record coherent and traceable. You're not rewriting history. You're making implicit reasoning explicit.
The reverse-engineering method treats the SSDD as a logic test. You start with the award decision and work backward, checking every link in the chain. Can the decision be explained by the comparative analysis? Can the comparative analysis be explained by the individual ratings? Can the ratings be explained by documented findings? Can the findings be tied to RFP evaluation criteria?
If any link is broken or invisible, that's where you focus repair efforts. Sometimes the findings are there but poorly worded. Sometimes the reasoning exists in meeting notes but never made it into the narratives. Sometimes the discriminators are real but weren't documented at all.
The three layers that must connect are: solicitation language, evaluation findings, and award rationale. If you can trace a clear line from "the RFP said we'd evaluate X" to "Offeror A's proposal did X better than Offeror B because of Y" to "we selected Offeror A based on that advantage," your record is defensible.
You'll know the record can be saved if evaluators can articulate the reasoning when you ask them directly and it's consistent with what's already documented. You'll know you need to re-evaluate if the ratings were based on unstated assumptions, improper considerations, or if the findings directly contradict the scores.
Step 1 – Audit Your Evaluation Narratives
Start by pulling out every documented strength and weakness for each offeror under each evaluation factor. List them separately. Don't summarize. You need to see exactly what the record says.
Now test each finding with this question: Does it tie directly to RFP evaluation criteria language? If your RFP said "the government will assess the offeror's risk mitigation strategy," and your documented strength is "the offeror proposed a strong risk mitigation strategy," that's too vague. What made it strong? How did it differ from other offerors?
Flag any findings that are conclusory. Phrases like "demonstrates understanding," "shows capability," or "reflects experience" are placeholders, not discriminators. They could apply to anyone. If you removed the offeror's name, could the finding describe a different proposal? If yes, it's not specific enough.
Check your weaknesses carefully. Are they actual deficiencies—gaps or flaws that create performance risk? Or are they just "not as strong as others"? A weakness must be tied to specific RFP criteria and explain what's missing or inadequate.
Look for missing findings. If evaluators spent twenty minutes in consensus debating a particular proposal element, but there's no corresponding strength or weakness documented, that's a gap. The reasoning happened, but it's not in the record.
Finally, check whether any finding relies on information not in the proposal or not permitted by the RFP. If you're citing past performance that wasn't submitted, or penalizing an offeror for something outside the evaluation criteria, that's a legal vulnerability.
Step 2 – Map Findings to Ratings
Now take the findings you've audited and map them to the ratings. For each subfactor, ask: do the documented strengths and weaknesses logically support the assigned rating?
If an offeror has three documented strengths and one minor weakness, does that justify a Green rating or should it be Blue? If two offerors both have Green ratings under the same subfactor, are their documented findings comparable in significance? Or did one offeror earn Green with strong findings and the other earn Green with thin documentation?
This is where you identify gaps. Sometimes a rating exists but the findings are sparse or generic. That tells you the evaluators had reasoning that didn't make it onto the page. You'll need to go back and reconstruct it from notes or evaluator recollection.
Check for internal consistency. If Offeror A got a Yellow because of two documented weaknesses, and Offeror B got a Green with one documented weakness, can you explain why B's weakness was less significant? If not, the ratings may reflect evaluator intuition rather than documented facts.
This step isn't about challenging the evaluators' judgment. It's about verifying that the record reflects their judgment accurately. If the logic is sound but poorly documented, you can fix it. If the ratings don't align with the findings at all, you may need to revisit consensus.
Step 3 – Build the Comparative Tradeoff Thread
This is where you move from individual ratings to cross-offeror comparison. The tradeoff analysis is the heart of your SSDD, and it must explain discriminators, not just restate colors.
Identify where offerors differed in ways that matter. Did Offeror A propose a more experienced project manager? Did Offeror B's technical approach include automation that others didn't? Did Offeror C have stronger past performance references but a weaker transition plan? These are discriminators.
Translate those discriminators into clear comparative statements. Not "Offeror A was better," but "Offeror A proposed daily status briefings while Offeror B proposed weekly briefings, and the government prioritized frequent communication due to the program's fast pace and high stakeholder visibility."
Verify that your tradeoff respects the RFP's stated weights. If Past Performance was less important than Technical Approach, you can't award to an offeror solely because of stronger references. The highest-weighted factors must drive the decision, and your narrative must show that.
Test your tradeoff logic by asking: if a lower-rated offeror had won, could I explain why based on this record? If a lower-priced offeror wasn't selected, can I show what the government got in exchange for paying more? If you can't answer those questions without repeating adjectives, the thread isn't complete.
Common Red Flags That Signal a Vulnerable Record
Certain patterns show up repeatedly in protest-vulnerable SSDDs. If you see any of these, pause and repair the record before moving to award.
- Strengths that describe minimum requirements rather than discriminators. "The offeror met the experience requirement" is not a strength. It's baseline compliance.
- Weaknesses that aren't tied to specific RFP criteria or reflect evaluator preference rather than solicitation priorities. "We would have preferred a different approach" is not a valid weakness unless the RFP specified that approach.
- Tradeoff statements that just restate ratings. "Offeror A was rated Blue and therefore provides the best value to the government" doesn't explain anything.
- Evaluation narratives that summarize what the offeror proposed but never assess it against RFP standards. Description is not evaluation.
- Ratings that rely on assumptions discussed verbally but not captured in writing. If it's not documented, it doesn't exist for protest purposes.
- Comparative analysis that ignores lower-rated factors or treats all factors as equally important when the RFP assigned different weights.
Think of your SSDD like a bridge. If any of these red flags are present, you've got a structural weakness. It might hold up under normal conditions, but the moment someone applies pressure—like a protest—it's going to fail.
Repairing the Record Without Re-Evaluating
Once you've identified where the logic chain breaks down, you can often repair it without re-opening the evaluation. The key is distinguishing between clarifying existing reasoning and creating new justifications after the fact.
If the findings exist but are poorly worded, you can revise the narrative to make them more specific. If the reasoning exists in consensus notes or technical reports but wasn't included in the SSDD, you can incorporate it. This is making implicit reasoning explicit, not changing the substance of the evaluation.
You can write a supplemental memo that connects dots the original narratives left unconnected. You can ask evaluators for clarifying statements, as long as you're not leading them toward a predetermined answer. The question is "what did you mean when you wrote this?" not "can you justify this rating?"
The bright line is this: if the evaluators can't explain their reasoning because they didn't actually have clear reasoning, or if the findings don't support the ratings no matter how you reframe them, you can't paper over the gap. That's when you need to consider corrective action or re-evaluation.
But in many cases, the evaluation was sound and the ratings were justified. The problem is just that the documentation didn't capture the team's reasoning. Fixing that is legitimate and necessary. You're not changing the decision. You're making the record match what actually happened.
Practical Application – A Real-World Scenario
Let's walk through an example. You're the KO for a professional services contract with three offerors. Technical Approach is weighted more heavily than Past Performance, and price is evaluated but not rated. Offeror B was selected with a Blue rating under Technical and Green under Past Performance. Offeror A had Green and Green, with a lower price.
Your draft SSDD says "Offeror B's superior technical rating justified the price premium." That's the problem right there. It's circular. A protester will ask: what made Offeror B's technical approach superior? What did the government get for the extra money?
You go back and audit the Technical Approach findings. Offeror B has three documented strengths: automation of reporting, a dedicated quality assurance lead, and prior experience with the same IT platform. Offeror A has two strengths: experienced staff and a proven methodology. Both have minor weaknesses.
Now you check the RFP. It emphasized reducing administrative burden and minimizing transition risk. Offeror B's automation directly addresses administrative burden. Their platform experience reduces transition risk. Those aren't just strengths—they're discriminators aligned with stated priorities.
You rebuild the tradeoff narrative: "While both offerors received favorable ratings, Offeror B's automated reporting approach and prior use of the government's IT platform provided distinct advantages under the highest-weighted factor. Given the RFP's emphasis on reducing administrative burden and minimizing transition risk, these discriminators justified the associated price premium."
Now the thread is complete. You can trace it from solicitation priorities to documented findings to the award decision. If Offeror A protests and asks "why not us?", you have a clear answer grounded in the record.
Why This Matters
Ratings are not the decision. The documented rationale is the decision. A clean chart with color-coded boxes is meaningless if the narrative underneath it can't withstand scrutiny.
A protest-vulnerable SSDD doesn't just risk the award. It risks the entire procurement timeline, the evaluation team's credibility, and the government's ability to bring the selected contractor onboard when needed. Corrective action after a sustained protest can add months or years to an acquisition.
Contracting officers are the last line of defense before award. You're the one who has to stand behind the record and defend it under protest. Catching the ratings-narrative disconnect before it becomes a legal problem is part of the job, and it's one of the most important things you can do to protect the procurement.
The reverse-engineering method gives you a repeatable process to audit and repair evaluation records under tight deadlines. It's not about perfection. It's about traceability. Can someone unfamiliar with your procurement read the SSDD and understand why you made the decision you made? If yes, you're in good shape. If no, you've got work to do.
And once you build the habit of checking for traceability, your future source selections get stronger from the start. You learn to challenge vague findings during consensus instead of after. You train technical evaluators to document discriminators, not just conclusions. You catch the disconnect early, when it's easy to fix, instead of discovering it three weeks before award when the stakes are highest.
%20(1).png)
.png)