Skip to Main Content Skip to main content

Home

Evidence-Based Practice in Rehabilitation

Appraising the Literature

Critical appraisal is a systematic analysis of research articles to determine the strength of the evidence in reference to your clinical question. Systematically review the different aspects (below) of each study to answer the following questions:

  1. What were the strengths of the study and the overall rigor of the study in terms of credibility?
    • Quality of evidence is based on the level and strength of the study.
      1. Level of evidence: Various scales have been developed to rank; e.g., Pedro scale
      2. Strength of evidence is based on the methodological limitations and/or threats to validity that may affect interpretation of findings and generalization of results.
  2. How are the results of this study relevant to your clinical question? How might the results influence clinical practice?
    • Applicability of evidence is based on its relevance to your question
    • How are the results of this study relevant to your clinical question? How might the results influence clinical practice?

Study Designs & Level of Evidence

Setting & Participants

Setting: Was the setting appropriate to answer the research question?

Participants/Sample:

  • How were participants recruited?
  • What were the inclusion/exclusion criteria?
  • How many participants participated? How many were lost through attrition?
  • What were the participant demographics?
  • If more than one group, was there similarity between groups?

Intervention & Outcome Measures

Intervention(s):

  • Was the intervention clearly described?
  • Who delivered the intervention and how often?
  • Was there cross-contamination between interventions?
  • Was there a break-in period?

Outcome measure(s):

  • What instruments or methods were used to measure the variables? Examples include participant observation, interviews, focus groups, instruments, devices & questionnaires.
  • Did the authors use measures with documented evidence of validity and reliability?
  • Was the procedural reliability documented?
  • How frequent were the participants measured?

Results & Conclusions

Main Results or Key Findings:

  • What were the results?
  • Was there statistical difference? What was the effect size?
  • How were the results analyzed and were the analysis methods appropriate?

Authors interpretation/conclusion: What was the clinical relevance of the study and the results?

Critical Appraisal

  • What are the strengths and weaknesses of the study?
    • Were interventions delivered and data collected systematically, objectively and with fidelity?
    • What were the potential threats to internal validity?
      • Examples of potential threats to internal validity: history, maturation, testing, instrumentation, statistical regression, selection, mortality, interactions with selection, ambiguity about causal influence, and diffusion of intervention.
    • What were the potential threats to external validity?
      • Examples of potential threats to external validity: interaction of testing and treatment, interaction of selection and treatment, interaction of setting and treatment, interaction of history and treatment, multiple-treatment interference.
      • Description of external validity
  • Were there limitations of the study to answer your clinical question?
  • Rate the study quality – this may be rated on a scale of 1-5 or 1-3.
    • Level of evidence is based on the study design(s)
    • Quality of evidence is based on the methodological strengths and weaknesses.
    • Applicability of evidence is based on its relevance to your question
      • Various scales have been developed to rank; e.g., Pedro scale
  • Your overall summary and interpretation of the study:
    • What were the strengths of the study and the overall rigor of the study in terms of credibility?
    • How are the results of this study relevant to your clinical question? How might the results influence clinical practice?