Cut Scores and Assessment Defensibility

A man smiles at his scored test on a laptop with his hands raised in excitement, showing the value of an appropriate exam cut score in making assessments defensible.

When administering a certification exam or licensing assessment, someone must decide what “passing” looks like. That decision, where exactly to draw the line between qualified and not qualified, is the cut score. It sounds simple, but in practice, it’s a highly significant and legally scrutinized component of the assessment development process.

Here’s what you need to know.

What Is a Cut Score?

A cut score (also called a passing score or passing standard) is the minimum score a candidate must achieve to demonstrate competency to pass an assessment. It’s the numerical threshold that separates two decisions: pass or fail.

For example, if a professional certification exam is scored on a 100-point scale and the cut score is set at 72, any candidate scoring 71 or below does not pass.

Cut scores aren’t arbitrary numbers; they represent a judgment about what level of performance is necessary for protecting public safety, demonstrating job readiness, and ensuring minimum competency in a regulated field.

How Are Cut Scores Established?

There are several established methodologies for setting cut scores, and choosing the right one matters. The most widely used approaches include:

The Angoff Method/Modified Angoff Method is the most commonly used study for many credentialing and licensure programs. SMEs review each test item and estimate the probability that a minimally competent candidate (someone who just barely qualifies) would answer the item correctly. Those probabilities are averaged across items to produce a recommended cut score.

The Bookmark Method presents items in order of difficulty. SMEs place a metaphorical “bookmark” at the point where items shift from what a minimally competent candidate would know to what they likely wouldn’t.

Borderline methods, such as the Borderline Group or Contrasting Groups methods, use actual candidate performance data and expert examiners rather than SME predictions. They’re often used in clinical or performance-based assessments where observable behavior from borderline candidates can be categorized.

Each approach has its appropriate context. The key is that the method must be systematic, structured, and documented.

What Does “Defensibility” Mean in This Context?

Defensibility refers to the ability to justify any aspect of the assessment development process, including the cut score, if they are ever challenged. Challenges can come from candidates who believe they were unfairly failed, from employers who question the validity of the credential, or from regulatory bodies and courts that scrutinize whether the assessment serves a legitimate purpose without creating unlawful barriers.

A defensible cut score is one you can explain clearly, trace back to a rigorous process, and show was informed by a panel of experts who determined what a minimum level of competency was needed for safe and effective practice.

In high-stakes contexts, defensibility isn’t optional. It’s a legal and ethical requirement.

Why SMEs Are Central to Defensibility

Subject matter experts are the starting point of a defensible standard-setting process. Regulatory agencies don’t just want to know what the cut score is, they want to know who set it and why those people were qualified to do so.

This is why SME selection matters as much as the methodology itself. Panelists should represent the relevant profession, hold current credentials or experience, and reflect diversity in geography, practice setting, employer type, and demographics. A cut score set by a homogeneous panel of experts from a single region or sector is harder to defend than one developed by a broadly representative group.

Documentation is equally critical. Every step of the standard-setting process should be recorded: who participated, what training they received, what instructions were given, how discussions were facilitated, and how final scores were calculated and decided. If a challenge arises three years after the assessment is launched, your standard setting report is your defense.

The Ongoing Relationship Between Cut Scores and Assessment Defensibility

A cut score isn’t set once and forgotten. Best practice calls for a standard setting for the first form after a new job task analysis or there are significant changes to eligibility requirements. Subsequent forms that are developed after the standard setting form should be equated back to the standard setting form to adjust for any differences in difficulty between forms.

Looking for guidance on setting your exam’s cut score?

reach out