top of page

STARD 2015: How to Report Diagnostic Accuracy Studies with Clarity and Rigor

  • Writer: Mayta
    Mayta
  • May 12, 2025
  • 3 min read

Introduction

Diagnostic tests form the cornerstone of clinical reasoning. But understanding whether a test is truly accurate requires high-quality research—research that is not only methodologically sound but also transparently reported. This is where STARDSTAndards for Reporting Diagnostic Accuracy Studies—steps in.

STARD 2015 is a reporting guideline, not a quality checklist. It ensures that authors provide sufficient information for readers, peer reviewers, and decision-makers to judge the trustworthiness and applicability of a study. In this article, we’ll walk through the full logic of the STARD checklist: what each item means, why it matters, and how to implement it, using novel examples for clarity.

🧩 1. What is STARD, and Why Does It Matter?

Core Purpose:

  • Promote transparency and completeness in reporting diagnostic accuracy studies.

  • Help users identify bias, applicability, and methodologic strengths/weaknesses.

What It Is Not:

  • STARD is not a tool to appraise study quality (like QUADAS-2).

  • It does not tell you how to conduct a study—but rather how to report it.

Problem Addressed:

  • Many diagnostic studies omit key data on methods, patients, or test interpretation—leading to misleading or irreproducible findings.

🧠 2. Key Concepts in Diagnostic Accuracy Studies

  • Diagnostic accuracy is not fixed; it varies across settings, populations, and test thresholds.

  • Bias can be introduced by choices in:

    • Study design (e.g., cross-sectional vs. case-control)

    • Data collection (e.g., blinding, verification)

    • Analysis methods (e.g., handling indeterminate results)

🧾 3. The STARD 2015 Checklist: 30 Reporting Items

Grouped across 7 domains, the STARD checklist guides the researcher from title to funding:

A. Title & Abstract (Items 1–2)

  • Clearly state that the study is a diagnostic accuracy study.

  • Include key metrics like sensitivity, specificity, PPV, NPV, and AUC.

Example: “Evaluation of Saliva Antigen Test Accuracy for COVID-19 Detection: A Prospective Cross-Sectional Study”

B. Introduction (Items 3–4)

  • Item 3: Give clinical context, including:

    • Intended use (screening, monitoring, staging)

    • Clinical role (triage, add-on, replacement)

  • Item 4: State objectives and hypotheses:

    • Is the test being compared against another?

    • Are there predefined accuracy thresholds?

C. Methods (Items 5–18)

Study Design (Item 5)

  • Specify if data collection was prospective or retrospective.

Participants (Items 6–9)

  • Eligibility criteria must match the test's intended use.

  • Clarify recruitment strategy (e.g., consecutive sampling).

  • Report setting and dates to assess external validity.

Example: Recruited all adult outpatients presenting with suspected UTI at 3 urban clinics over 12 months.

Test Methods (Items 10–13)

  • Describe both index and reference tests in enough detail for replication.

  • Report who interpreted the tests and whether they were blinded.

  • State the rationale for reference standard selection.

Cut-offs (Items 12a–12b)

  • Pre-specify thresholds to avoid overfitting.

  • Avoid choosing cut-offs based solely on maximizing AUC after seeing data.

Analysis (Items 14–18)

  • Describe statistical methods for calculating Se/Sp, likelihood ratios, etc.

  • Clarify how indeterminate or missing data were handled.

  • Report whether sample size was calculated a priori and how.

D. Results (Items 19–25)

Flow Diagram (Item 19)

  • Use STARD-style flow charts showing inclusion/exclusion.

Participant Characteristics (Items 20–21)

  • Report demographics, disease severity spectrum, and alternative diagnoses in those without disease.

Timing (Item 22)

  • Describe time interval between tests—important for diseases that change rapidly (e.g., infectious disease).

Data Presentation (Items 23–24)

  • Include 2×2 cross-tabulations and confidence intervals for metrics.

Harms (Item 25)

  • Report any adverse effects from tests (e.g., radiation from CT).

E. Discussion (Items 26–27)

  • Study limitations: Reflect on biases, generalizability, and uncertainty.

  • Implications for practice: Should the test replace current practice? For whom?

F. Other Information (Items 28–30)

  • State registration number, protocol availability, and funding sources.

📌 Key Pitfalls STARD Helps Prevent

Pitfall

STARD Protection

Missing participant flow info

Item 19 – Flow diagram

No clarity on test positivity threshold

Item 12 – Cut-off reporting

Hidden verification bias

Item 5 – Study design + Item 13 – Blinding

Overfitting by post-hoc threshold tuning

Item 12 – Require pre-specification

Unclear setting and generalizability

Items 8, 20, 21 – Context & demographics


🔍Clinical Example: Applying STARD

Imagine evaluating a novel blood biomarker to detect early-stage pancreatic cancer. A good STARD-compliant report would:

  • Recruit patients based on clinical presentation (e.g., unexplained weight loss + epigastric pain).

  • Describe how blood samples were collected, who processed them, and who interpreted the results.

  • Use endoscopic ultrasound with biopsy as the reference standard and explain why.

  • Clearly state the cut-off value used to classify the biomarker as positive.

  • Report 2×2 tables and CI for all diagnostic metrics.

  • Describe whether pathologists were blinded to the blood test result.


✅ Key Takeaways

  • STARD 2015 ensures that diagnostic accuracy studies are transparent, reproducible, and interpretable.

  • Proper reporting improves the credibility and clinical utility of your findings.

  • The checklist supports a methodologically sound narrative—one that tells the full story of your study.

  • It aligns well with bias avoidance (see: QUADAS-2) and strengthens peer-review defense.

Recent Posts

See All
Internal Validation vs Instability

Pocket note “The concept depends on which dataset you compare the model against (i.e., where you evaluate it).” Why it feels like “same data but different view” Think of data as wearing different hat

 
 
 

Comments

Rated 0 out of 5 stars.
No ratings yet

Add a rating
Post: Blog2_Post

​Message for International and Thai Readers Understanding My Medical Context in Thailand

Message for International and Thai Readers Understanding My Broader Content Beyond Medicine

bottom of page