Validation of a detailed scoring checklist for use during advanced cardiac life support certification

Matthew D. McEvoy, Jeremy C. Smalley, Paul J. Nietert, Larry C. Field, Cory M. Furse, John W. Blenko, Benjamin G. Cobb, Jenna L. Walters, Allen Pendarvis, Nishita S. Dalal, John J. Schaefer

Research output: Contribution to journalArticlepeer-review

26 Scopus citations

Abstract

Introduction: Defining valid, reliable, defensible, and generalizable standards for the evaluation of learner performance is a key issue in assessing both baseline competence and mastery in medical education. However, before setting these standards of performance, the reliability of the scores yielding from a grading tool must be assessed. Accordingly, the purpose of this study was to assess the reliability of scores generated from a set of grading checklists used by nonexpert raters during simulations of American Heart Association (AHA) Megacodes. Methods: The reliability of scores generated from a detailed set of checklists,when used by 4 nonexpert raters, was tested by grading team leader performance in 8 Megacode scenarios. Videos of the scenarios were reviewed and rated by trained faculty facilitators and a group of nonexpert raters. The videos were reviewed "continuously" and "with pauses." The grading made by 2 content experts served as the reference standard, and 4 nonexpert raters were used to test the reliability of the checklists. Results: Our results demonstrate that nonexpert raters are able to produce reliable grades when using the checklists under consideration, demonstrating excellent intrarater reliability and agreement with a reference standard. The results also demonstrate that nonexpert raters can be trained in the proper use of the checklist in a short amount of time, with no discernible learning curve thereafter. Finally, our results show that a single trained rater can achieve reliable scores of team leader performance during AHA Megacodes when using our checklist in a continuous mode because measures of agreement in total scoring were very strong [Lin's (Biometrics 1989;45:255Y268) concordance correlation coefficient, 0.96; intraclass correlation coefficient, 0.97]. Conclusions: We have shown that our checklists can yield reliable scores, are appropriate for use by nonexpert raters, and are able to be used during continuous assessment of team leader performance during the review of a simulated Megacode. This checklist may be more appropriate for use by advanced cardiac life support instructors during Megacode assessments than the current tools provided by the AHA.

Original languageEnglish (US)
Pages (from-to)222-235
Number of pages14
JournalSimulation in Healthcare
Volume7
Issue number4
DOIs
StatePublished - Aug 2012
Externally publishedYes

Keywords

  • ACLS
  • Checklist
  • Education
  • Reliability
  • Simulation

ASJC Scopus subject areas

  • Epidemiology
  • Medicine (miscellaneous)
  • Education
  • Modeling and Simulation

Fingerprint

Dive into the research topics of 'Validation of a detailed scoring checklist for use during advanced cardiac life support certification'. Together they form a unique fingerprint.

Cite this