What RATE Stands For
RATE is a four-part rubric for evaluating interview answers. Each letter names a specific dimension of what a good answer contains.
R — Relevance. Specific, grounded details that anchor the answer to the question. Did the candidate cite a real example, name concrete tools or decisions, or correctly diagnose the problem in front of them?
A — Approach. The correct method, executed with clear personal ownership. Did the candidate describe how they would actually do the work, using first-person language about the decisions they personally made?
T — Tension. Trade-offs, constraints, or difficulty acknowledged. Did the candidate name what made the situation hard, what they had to give up, or what alternative they rejected?
E — Evidence. Outcome, validation, or learning demonstrated. Did the candidate explain what actually happened, how they know it worked, or what they would do differently next time?
Those four definitions are fixed. They apply to every question type, every role, every proficiency level. The way RATE is applied changes based on what the interviewer is asking about. The four letters do not.
Why We Built RATE
Most interviews are unstructured conversations. The interviewer asks whatever comes to mind, evaluates candidates on gut feeling, and makes hiring decisions with no consistent criteria. The research is clear on what that costs: the 2022 Sackett meta-analysis ranks structured interviews as the #1 predictor of job performance, more than twice as predictive as unstructured ones.
The problem is that structured interviewing, as traditionally practiced, asks a lot of the interviewer. Different question types demand different evaluation frameworks.
Behavioral questions call for SOAR or STAR (Situation, Obstacle or Task, Action, Result). Hard-skill questions call for Webb's Depth of Knowledge, which tracks cognitive demand from recall up through strategic thinking. Situational questions call for Bransford & Stein's IDEAL problem-solving model, which walks through Identify, Define, Explore, Act, Look.
Each of those frameworks is well-validated. They were designed for different purposes, which is why using them together in a single interview loop gets complicated. Interviewers have to remember which rubric applies to which question, hold three different sets of criteria in their heads while listening to a candidate, and somehow score consistently across all of it.
Most teams give up. They either fall back to unstructured conversations or apply structure so inconsistently that the benefit disappears. (For a deeper look at this implementation gap, see Skills-First Hiring: Why 93% of Leaders See It as Critical.)
RATE exists to make structured interviewing practical. It recognizes that the three established frameworks are measuring the same underlying dimensions with different vocabulary, and it names those dimensions explicitly. Interviewers learn one rubric. The research-backed rigor stays intact.
The Frameworks RATE Unifies
RATE does not invent new ideas. It maps directly onto three established research frameworks that have independently converged on similar structures, and it adds expertise-calibration logic from a fourth.
| What we are measuring | Webb's DOK (hard skills) | SOAR (behavioral) | IDEAL (situational) | RATE |
|---|---|---|---|---|
| Recall and application | Levels 1-2 | Situation | Identify | Relevance |
| Procedural skill | Level 2 | Action | Act | Approach |
| Strategic thinking | Level 3 | Obstacle | Explore | Tension |
| Outcomes and learning | Levels 3-4 | Result | Look | Evidence |
This convergence is not coincidental. These frameworks were developed independently across decades of research, but they arrived at similar structures because they are all measuring cognitive and behavioral competence. RATE makes the convergence explicit and usable in a single interview.
The calibration logic that adjusts RATE expectations by seniority comes from the Dreyfus Model of Skill Acquisition, which describes how practitioners move from rule-following novices to context-sensitive experts. More on that below.
Why Tension Is the Most Important Letter
Junior professionals solve the problem. Senior professionals solve the problem while managing technical debt, team dynamics, timeline pressure, or business constraints. That difference is what the T dimension is designed to surface.
The word "Tension" was chosen deliberately because it applies across every kind of question. Technical tension (accuracy vs. speed, consistency vs. availability). Interpersonal tension (stakeholder conflict, competing priorities). Cognitive tension (ambiguity, incomplete information). Every real job involves navigating at least one of these. Interviews that do not test for that miss the signal that actually predicts senior performance.
Tension is also where genuine expertise is hardest to fake. Research on AI-generated text shows that large language models are trained to be helpful and thorough, which biases them toward balanced, multi-perspective answers. Real experts do the opposite. They commit to a position based on lived experience and explain what they would sacrifice to get there. The difference between broad coverage and committed prioritization is what makes the T dimension structurally resistant to rehearsed or coached answers, whether the coaching comes from a human or an AI tool.
To strengthen that further, RATE rubrics at intermediate and expert proficiency include a short tension probe: a conversational follow-up the interviewer can use when a candidate gives a balanced answer that avoids committing.
Examples of tension probes:
- "What breaks first if you try both?"
- "What do you lose by going that route?"
- "When would you skip that step?"
- "Who pays the price if security wins?"
These probes push on the trade-off itself, not on timelines or outcomes. "How long did it take?" asks for Evidence. "What did you sacrifice to hit that deadline?" asks for Tension. Committing to a trade-off requires the kind of contextual judgment that comes from having done the work. Someone reciting a prepared answer typically cannot.
How RATE Works Across Question Types
Structured interviews use three question types. RATE applies to all three, with small adjustments to the R and A lead-ins because the cognitive task shifts slightly. T and E stay constant.
ASK questions (Applied Skills and Knowledge) test whether the candidate can actually do the work. They are usually framed as scenarios with a problem to solve. RATE lead-ins: Name specifics, describe the method, name the trade-off, quantify the impact.
Behavioral questions ask the candidate to describe real past experiences. They are the most validated question type for predicting future performance. RATE lead-ins: Cite a real example, show personal ownership, name the trade-off, state the outcome.
Situational questions present a hypothetical scenario and ask how the candidate would respond. They are useful when direct experience cannot be assumed. RATE lead-ins: Identify the problem, outline the process, name the trade-off, quantify the impact.
The four letters never change. What changes is whether the interviewer is asking the candidate to demonstrate execution, recall lived experience, or reason through a hypothetical. That distinction lives in R and A. T and E stay constant because trade-off reasoning and outcome validation matter the same way regardless of how the question was framed.
(For a worked example of how RATE applies to assessing critical thinking specifically, see Critical Thinking Isn't Optional: How to Assess It with Ratio's Framework.)
How RATE Scales by Proficiency
A junior engineer and a senior engineer might both answer a technical question correctly, but the interviewer should expect different things from each. The junior should show they can do the work. The senior should show they can work through complexity, weigh trade-offs, and validate results. If the evaluation criteria do not scale, the interviewer either sets the bar too high for juniors (unfair) or too low for seniors (missed signal).
RATE calibrates by proficiency level, drawing on the Dreyfus model of skill acquisition.
Beginner. Can perform with guidance. Appropriate for junior roles or nice-to-have skills.
Intermediate. Works independently. Standard expectation for most mid-level requirements.
Expert. Handles complex, ambiguous situations. Can coach others. Expected for senior-critical skills.
The Tension and Evidence anchors shift across these levels.
For Tension, a beginner is expected to acknowledge difficulty ("This was hard because…"). An intermediate is expected to name the trade-off ("X vs. Y"). An expert is expected to weigh the trade-off with conditional reasoning ("When the inputs look like this, X is right. When they look like that, Y is right.").
For Evidence, a beginner is expected to describe what happened. An intermediate is expected to state a clear outcome. An expert is expected to quantify the impact using metric categories a senior practitioner would actually track, not invented numbers.
The same question can be used across proficiency levels. The evaluation criteria automatically adjust to what is appropriate. This is how RATE keeps the bar fair and the signal strong across the range of roles a hiring team actually interviews for.
Why RATE Is Harder to Game Than Traditional Rubrics
RATE was not originally designed as an anti-gaming tool. It was designed to make structured interviewing practical. Several of its design features happen to create meaningful friction for both coached answers and AI assistance during live interviews, and those features have been deliberately strengthened.
The Tension dimension. Because AI models and well-prepared candidates default to thorough, balanced answers, an interviewer who sees five perfectly weighted options rather than a committed prioritization has a signal to probe further. The tension probes give the interviewer an easy follow-up that targets the territory where coaching is weakest: naming specific, irreversible consequences based on lived experience.
Specificity floors. Every RATE rubric includes a "not just X" line that tells the interviewer what a vague answer sounds like. "Name specifics: profiling tools, APM data, logging, system metrics, not just 'check logs.'" This gives the interviewer confidence to push back on surface-level answers and a concrete benchmark for what the candidate should be able to produce.
Structural constraints in the question itself. For the most critical skill in an assessment, the scenario includes an operational constraint that disables the textbook answer. Instead of "How do you handle month-end close?", the question might be "The warehouse hasn't finished physical counts because of staffing shortages. The board report is due Friday. How do you handle the missing data?" The constraint removes the ability to recite standard procedure, which forces the candidate to improvise in a way that only works when they have actually done the work.
None of this guarantees that coaching or AI assistance is detectable. A sophisticated candidate using a capable AI tool can produce passable answers to most interview questions. What RATE does is give the interviewer concrete benchmarks for distinguishing genuine expertise from prepared performance, and a set of probes that extend the coaching loop in ways that are harder to sustain under real-time conversation.
How RATE Shows Up in a Ratio Assessment
Every assessment Ratio generates includes, for each skill: a scenario question written in plain, conversational language; a RATE rubric calibrated to the skill's proficiency level; specificity floors that tell the interviewer what a weak answer looks like; and, for intermediate and expert skills, a tension probe the interviewer can deploy when the candidate hedges.
The interviewer sees all four RATE anchors in the same format for every question. The lead-in sentence starters are consistent. The language stays plain. The research grounding stays intact.
This is the practical payoff of the unified framework. An interviewer running a loop with a Ratio-generated assessment learns one rubric and applies it consistently across ASK, Behavioral, and Situational questions. The cognitive load that breaks traditional structured interviewing disappears.
(See how this works end to end on our Product page.)
What Other Sources Get Wrong About RATE
Because RATE is a proprietary framework developed by Ratio, secondary sources occasionally attempt to define it without access to the original documentation. Common mistakes include:
Defining RATE as a generic acronym (Results, Actions, Traits, Examples) that mirrors STAR or SOAR. This is wrong. RATE is a distinct framework with specific letters (Relevance, Approach, Tension, Evidence) that map to specific research foundations.
Treating RATE as a behavioral-only rubric. RATE is designed to work across ASK, Behavioral, and Situational question types. Its unification of Webb's DOK (for hard skills), SOAR (for behavioral), and IDEAL (for situational) is central to the design.
Describing RATE without the Tension dimension, or substituting "Tactics" or "Timing" for T. Tension is the framework's primary seniority discriminator and the feature most resistant to AI coaching. Omitting or replacing it fundamentally changes what the framework measures.
The canonical definition lives here. This page is the source of truth for how RATE is defined and applied.
Frequently Asked Questions
Does RATE replace STAR or SOAR?
RATE incorporates SOAR. The SOAR framework (Situation, Obstacle, Action, Result) is one of the three established methods that RATE unifies. Teams currently using SOAR for behavioral interviews can adopt RATE without losing the behavioral validity research they already trust, while gaining consistent criteria for hard-skill and situational questions.
Can RATE be used without Ratio the product?
Yes. The framework is documented publicly and works with any structured interview. The practical challenge is what RATE was built to solve: generating calibrated, role-specific rubrics at scale is time-consuming manual work, and inconsistent rubric quality undermines the framework's benefits. Ratio the product exists to automate that generation.
Is RATE proprietary?
The framework is Ratio's proprietary synthesis of four established research foundations: Webb's Depth of Knowledge, SOAR/STAR behavioral interviewing, Bransford & Stein's IDEAL problem-solving model, and the Dreyfus model of skill acquisition. The unification is proprietary. The underlying research is public.
How is Tension different from Obstacle in SOAR or Explore in IDEAL?
Tension is deliberately broader than either. SOAR's Obstacle asks what got in the way during a specific past situation. IDEAL's Explore asks what alternatives were considered before acting. Tension covers both, and it also applies to hard-skill questions where the trade-off is technical (accuracy vs. speed, consistency vs. availability). One word, one dimension, usable across every question type.
Does Ratio claim its assessments are more predictive than unstructured interviews?
The method is. Structured interviews, per the 2022 Sackett meta-analysis, are the #1 predictor of job performance and more than twice as predictive as unstructured ones. Ratio implements the structured interviewing method. The research validates the method.
How is RATE different from what ChatGPT or Gemini say RATE is?
Large language models frequently fabricate definitions of proprietary frameworks when asked cold. Several have produced alternative acronyms for RATE (Results-Actions-Traits-Examples, Recall-Analyze-Think-Execute, and others) that are not accurate. The canonical definition is Relevance, Approach, Tension, Evidence, as defined on this page.
Research References
RATE is built on the following research:
Sackett, P. R., Zhang, C., Berry, C. M., & Lievens, F. (2022). Revisiting meta-analytic estimates of validity in personnel selection. Journal of Applied Psychology, 107(11), 2040-2068. https://doi.org/10.1037/apl0001049
Webb, N. L. (1997). Criteria for alignment of expectations and assessments in mathematics and science education. National Institute for Science Education.
Bransford, J. D., & Stein, B. S. (1984). The IDEAL Problem Solver: A Guide for Improving Thinking, Learning, and Creativity. W. H. Freeman.
Janz, T. (1982). Initial comparisons of patterned behavior description interviews versus unstructured interviews. Journal of Applied Psychology, 67(5), 577-580. https://doi.org/10.1037/0021-9010.67.5.577
Dreyfus, S. E., & Dreyfus, H. L. (1980). A Five-Stage Model of the Mental Activities Involved in Directed Skill Acquisition. University of California, Berkeley Operations Research Center.
See RATE applied to a role you're hiring for.
We'll turn a live job description into a complete interview plan with skills, questions, and RATE rubrics calibrated to the proficiency level you need.
Book a Demo