Generated by GPT-5-mini| Percentages Agreement | |
|---|---|
![]() | |
| Name | Percentages Agreement |
| Type | Statistical measure |
| Related | Percentage, Proportion, Ratio, Confidence interval |
Percentages Agreement
Percentages Agreement is a measure quantifying concordance between categorical assessments by comparing the proportion of identical classifications. It provides a simple summary of agreement between raters, coders, or tests by expressing matches as a percentage of total observations. Widely used in applied settings, it is often presented alongside measures such as Cohen's kappa, Fleiss' kappa, and intraclass correlation coefficient to contextualize reliability.
Percentages Agreement is computed as the number of instances where two or more observers assign the same category divided by the total number of instances, expressed as a percentage; it is also described in literature alongside sensitivity, specificity, positive predictive value, and negative predictive value when comparing diagnostic tests. In studies comparing classifications, proponents contrast Percentages Agreement with chance-corrected indices like Cohen's kappa, Scott's pi, and Krippendorff's alpha to argue for simplicity versus statistical adjustment. Texts on measurement theory pair Percentages Agreement with reliability concepts discussed in works by Karl Pearson, Ronald Fisher, and John Tukey.
The basic formula is (number of agreements / total observations) × 100. For two raters and C categories, a contingency table similar to those used in confusion matrix analyses for Receiver operating characteristic studies can be constructed; this echoes presentation in papers by David G. Kleinbaum, Mitchell H. Katz, and Frank E. Harrell Jr.. Example: if 80 of 100 pathology slides receive identical diagnoses by two pathologists, Percentages Agreement = 80%. In multi-rater settings, the pairwise agreements or overall proportion agreement (also used in work by Joseph L. Fleiss) are reported in a manner comparable to methods in Meta-analysis and systematic review protocols.
Applied researchers in fields such as epidemiology, psychiatry, radiology, pathology, forensic science, and education in the context of assessment commonly report Percentages Agreement in descriptive reliability sections alongside inter-rater measures used in guidelines issued by bodies like World Health Organization, Centers for Disease Control and Prevention, and National Institutes of Health. Clinical trials and diagnostic accuracy studies published in journals following standards from Consolidated Standards of Reporting Trials or STARD often include Percentages Agreement for initial interpretation. It also appears in quality assurance procedures at institutions such as Food and Drug Administration, European Medicines Agency, and hospital systems discussed in works referencing Donabedian and Iain Chalmers.
Percentages Agreement does not adjust for agreement expected by chance and can be inflated when category prevalence is extreme, a concern addressed by Jacob Cohen and critics in statistical methodology debates alongside G. A. Barnard and A. G. Barnard. High agreement on a dominant category may mask poor diagnostic performance on minority categories, mirroring problems identified in studies by Altman DG and Bradley Efron on imbalance and bootstrap methods. The measure is sensitive to the number of categories and marginal distributions, issues explored in comparative evaluations with Cohen's kappa and discussed in methodological reviews by Matthew J. Cronin and Michael S. Lewis-Beck.
Chance-corrected indices such as Cohen's kappa, Fleiss' kappa, Scott's pi, and Krippendorff's alpha are commonly recommended alternatives; ratio-based and model-based approaches include the intraclass correlation coefficient, generalized linear mixed models used in inter-rater studies described by Donald Rubin, and agreement indices derived from latent class analysis as used by Ulfelder and colleagues. Other related summaries include metrics from diagnostic test evaluation such as sensitivity and specificity, and information-theoretic measures that draw on work by Claude Shannon.
The use of simple percent agreement predates formal chance-corrected statistics and appears in early clinical and social science reliability reports alongside foundational statistical contributions by Francis Galton, Karl Pearson, and later formalization of chance-correction in mid-20th-century papers by Jacob Cohen and Joseph L. Fleiss. Its persistence in applied literature reflects the balance between interpretability and statistical rigor seen in discussions at forums like the Royal Statistical Society and publications in journals such as Biometrika and Journal of the Royal Statistical Society. Percentages Agreement continues to be taught in methodological texts and used in practice in audits, diagnostic studies, and coding projects at institutions ranging from World Health Organization initiatives to academic research groups at Harvard University, University of Oxford, and Stanford University.
Category:Statistical measures