Key TakeawaysCriterion validity (orcriterion-related validity) examines how well a measurement tool corresponds to other established and valid measures of the same concept.It includes concurrent validity (existing criteria) and predictive validity (future outcomes).Criterion validity is important because, without it, tests would not be able to accurately measure in a way consistent with other validated instruments.

Key Takeaways

criterion validity

Criterion validity assesses how well a test predicts or relates to a specific outcome or criterion. It includesconcurrent validity(correlation with existing measures) andpredictive validity(predicting future outcomes).

This approach emphasizespractical applicationsand focuses on demonstrating that the test scores are useful for predicting or estimating a particular outcome.

For example, when measuring depression with a self-report inventory, a researcher can establish criterion validity if scores on the measure correlate with external indicators of depression such as clinician ratings, number of missed work days, or length of hospital stay.

Types of Criterion Validity

Predictive

Predictive validity demonstrates that a test score can predict future performance on another criterion (Cohen & Swerdik, 2005).

Good predictive validity is important when choosing measures for employment or educational purposes, as it increases the likelihood of selecting individuals who will perform well.

The correlation between scores on standardized tests like the SAT or ACT and a student’s first-year GPA is often used as evidence for the predictive validity of these tests.

These tests aim to predict future academic performance, and a strong positive correlation between test scores and subsequent GPA would support their ability to do so.

Concurrent

This can be shown when scores on a new test correlate highly with scores on an established test measuring similar constructs (Barrett et al., 1981).

While correlational analyses are most common, researchers may also use regression.

Validation methods include comparing responses between new and established measures given to the same group, or comparing responses to expert judgments (Fink, 2010).

Note that concurrent validity does not guarantee predictive validity.

How to measure criterion validity

Identify a well-established, validated measure (criterion) that assesses the same construct as the new measure you want to validate.

This criterion measure should have demonstrated reliability and validity, serving as a benchmark for comparison.

Establishing Concurrent Validity:

Establishing Predictive Validity:

Intelligent tests

Researchers developing a new, shorter intelligence test might administer it concurrently alongside a well-established test, such as the Stanford-Binet.

If there is a high correlation between the scores from the two tests, it suggests the new test measures the same construct (intelligence), supporting its concurrent validity.

Risk assessment and dental treatment

Bader et al. (2005) studied the predictive validity of a subjective method for dentists to assess patients’ caries risk.

They analyzed data from practices that had used this method for several years to see if the risk categorization predicted the subsequent need for caries-related treatment.

Their findings showed that patients categorized as high-risk were four times more likely to receive treatment than those categorized as low-risk, while those categorized as moderate-risk were twice as likely.

This supports the predictive validity of this assessment method.

Minnesota Multiphasic Personality Inventory

The initial validation of theMMPIinvolved identifying items that differentiated between individuals with specific psychiatric diagnoses and those without, contributing to the development of scales for various psychopathologies.

This method of establishing validity, where the test is compared to an existing criterion measured at the same time, exemplifies concurrent validity.

What is the difference between criterion and construct validity?Criterion validity examines the relationship between test scores and a specific external criterion the test aims to measure or predict.This criterion is a separate, independent measure of the construct of interest.This approach emphasizespractical applicationsand focuses on demonstrating that the test scores are useful for predicting or estimating a particular outcome.Construct validityseeks to establish whether the test actually measures the underlying psychological construct it is designed to measure.It goes beyond simply predicting a criterion and aims to understand the test’stheoretical meaning.How do you increase criterion validity?There are several ways to increase criterion validity, including (Fink, 2010):– Making sure the content of the test is representative of what will be measured in the future– Using well-validated measures– Ensuring good test-taking conditions– Training raters to be consistent in their scoring

What is the difference between criterion and construct validity?Criterion validity examines the relationship between test scores and a specific external criterion the test aims to measure or predict.This criterion is a separate, independent measure of the construct of interest.This approach emphasizespractical applicationsand focuses on demonstrating that the test scores are useful for predicting or estimating a particular outcome.Construct validityseeks to establish whether the test actually measures the underlying psychological construct it is designed to measure.It goes beyond simply predicting a criterion and aims to understand the test’stheoretical meaning.

What is the difference between criterion and construct validity?

Criterion validity examines the relationship between test scores and a specific external criterion the test aims to measure or predict.This criterion is a separate, independent measure of the construct of interest.This approach emphasizespractical applicationsand focuses on demonstrating that the test scores are useful for predicting or estimating a particular outcome.Construct validityseeks to establish whether the test actually measures the underlying psychological construct it is designed to measure.It goes beyond simply predicting a criterion and aims to understand the test’stheoretical meaning.

Criterion validity examines the relationship between test scores and a specific external criterion the test aims to measure or predict.

This criterion is a separate, independent measure of the construct of interest.

Construct validityseeks to establish whether the test actually measures the underlying psychological construct it is designed to measure.

It goes beyond simply predicting a criterion and aims to understand the test’stheoretical meaning.

How do you increase criterion validity?There are several ways to increase criterion validity, including (Fink, 2010):– Making sure the content of the test is representative of what will be measured in the future– Using well-validated measures– Ensuring good test-taking conditions– Training raters to be consistent in their scoring

How do you increase criterion validity?

There are several ways to increase criterion validity, including (Fink, 2010):– Making sure the content of the test is representative of what will be measured in the future– Using well-validated measures– Ensuring good test-taking conditions– Training raters to be consistent in their scoring

There are several ways to increase criterion validity, including (Fink, 2010):

– Making sure the content of the test is representative of what will be measured in the future– Using well-validated measures– Ensuring good test-taking conditions– Training raters to be consistent in their scoring

Aboraya, A., France, C., Young, J., Curci, K., & LePage, J. (2005). The validity of psychiatric diagnosis revisited: the clinician’s guide to improve the validity of psychiatric diagnosis.Psychiatry (Edgmont),2(9), 48.

Bader, J. D., Perrin, N. A., Maupomé, G., Rindal, B., & Rush, W. A. (2005). Validation of a simple approach to caries risk assessment.Journal of public health dentistry,65(2), 76-81.

Barrett, G. V., Phillips, J. S., & Alexander, R. A. (1981). Concurrent and predictive validity designs: A critical reanalysis.Journal of Applied Psychology,66(1), 1.

Conte, J. M. (2005). A review and critique of emotional intelligence measures.Journal of Organizational Behavior,26(4), 433-440.

Fink, A. Survey Research Methods. In McCulloch, G., & Crook, D. (2010).The Routledge international encyclopedia of education. Routledge.

Prince, M. Epidemiology. In Wright, P., Stern, J., & Phelan, M. (Eds.). (2012).Core Psychiatry EBook. Elsevier Health Sciences.

Schmidt, F. L. (2012). Cognitive tests used in selection can have content validity as well as criterion validity: A broader research review and implications for practice.International Journal of Selection and Assessment,20(1), 1-13.

Swerdlik, M. E., & Cohen, R. J. (2005). Psychological testing and assessment: An introduction to tests and measurement.

Print Friendly, PDF & Email

Saul McLeod, PhD

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Olivia Guy-Evans, MSc

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

Charlotte NickersonResearch Assistant at Harvard UniversityUndergraduate at Harvard UniversityCharlotte Nickerson is a student at Harvard University obsessed with the intersection of mental health, productivity, and design.

Charlotte NickersonResearch Assistant at Harvard UniversityUndergraduate at Harvard University

Charlotte Nickerson

Research Assistant at Harvard University

Undergraduate at Harvard University

Charlotte Nickerson is a student at Harvard University obsessed with the intersection of mental health, productivity, and design.