Workplace-Based Assessments Using Pediatric Critical Care Entrustable Professional Activities

MD, MHPE,
PhD,
MD, and
PhD, MHPE
Online Publication Date: 01 Aug 2019
Page Range: 430 – 438
DOI: 10.4300/JGME-D-18-01006.1
Save
Download PDF

ABSTRACT

Background

Workplace-based assessment (WBA) is critical to graduating competent physicians. Developing assessment tools that combine the needs of faculty, trainees, and governing bodies is challenging but imperative. Entrustable professional activities (EPAs) are emerging as a clinically oriented framework for trainee assessment.

Objective

We sought to develop an EPA-based WBA tool for pediatric critical care medicine (PCCM) fellows. The goals of the tool were to promote learning through benchmarking and tracking entrustment.

Methods

A single PCCM EPA was iteratively subdivided into observable practice activities (OPAs) based on national and local data. Using a mixed-methods approach following van der Vleuten's conceptual model for assessment tool utility and Messick's unified validity framework, we sought validity evidence for acceptability, content, internal structure, relation to other variables, response process, and consequences.

Results

Evidence was gathered after 1 year of use. Items for assessment were based on correlation between the number of times each item was assessed and the frequency professional activity occurred. Phi-coefficient reliability was 0.65. Narrative comments demonstrated all factors influencing trust, identified by current literature, were cited when determining level of entrustment granted. Mean entrustment levels increased significantly between fellow training years (P = .001). Compliance for once- and twice-weekly tool completion was 50% and 100%, respectively. Average time spent completing the assessment was less than 5 minutes.

Conclusions

Using an EPA-OPA framework, we demonstrated utility and validity evidence supporting the tool's outcomes. In addition, narrative comments about entrustment decisions provide important insights for the training program to improve individual fellow advancement toward autonomy.

Introduction

Workplace-based assessment (WBA), the process of observing and assessing trainee performance in the clinical environment, is critical to the development and certification of competent physicians. Directly observing trainees in the workplace provides opportunities for immediate performance feedback to guide further learning.1 By situating the learning within daily clinical duties and focusing on assessment that improves performance, WBAs present the greatest rewards for all stakeholders.24

The challenge has been identifying a WBA framework that balances holistic and longitudinal judgements of trainee performance with appropriate granularity for frontline clinicians. Based on the requirement to upload milestone levels for trainees every 6 months, the Accreditation Council for Graduate Medical Education (ACGME) Competencies and Milestones served as an initial framework for WBA tools.58 Unfortunately, many behaviors in the Pediatrics Milestones, such as “reflection in” versus “reflection on” action, cannot be directly observed, and tools based on the ACGME Competencies and Milestones deconstructed the complex interactions needed to make clinical decisions. This mismatch between the assessments by clinical faculty and the way clinicians function in practice leads to construct misalignment which creates unreliable assessment outcomes.9

Entrustable professional activities (EPAs), descriptions of daily professional activities trainees should be capable of performing independently, emerged as an alternative WBA framework.10 EPAs provide a holistic approach to defining workplace outcomes and having improved construct alignment.1012 While not required by the ACGME as part of trainee assessment, EPAs developed by specialty societies have provided a mechanism for converting theoretical concepts into meaningful clinical curricula and WBA tools by standardizing language for the decisions made by clinical faculty about trusting and supervising trainees.1015 The scope of some EPAs impacts their usefulness as an assessment for learning tool, because faculty feel uncomfortable making decisions of entrustment or supervision over such a broad statement.13 To address this concern, Warm and colleagues introduced the concept of observable practice activities (OPAs).14,15 Narrowing the scope of broad EPAs to specific practice activities provides the granularity and contextualization needed for direct observation, formative assessments, and translation to a useful assessment tool. As such, using a combined EPA-OPA framework to assess more granular, yet frequent, measurements may create useful WBA tools to ensure independent practice.

To address a paucity of meaningful assessment data for performance improvement and advancement decisions due to construct misalignment coming from current assessment tools based on the ACGME Competencies and Milestones in our PCCM fellowship, we sought to develop a novel WBA tool using an EPA-OPA framework. This study examines the development and implementation of that tool, hypothesizing that the tool provides sufficient validity evidence to infer utility as a formative assessment.

Methods

Setting and Participants

Participants included 15 faculty supervisors and 12 PCCM fellows practicing in a 28-bed medical-surgical pediatric intensive care unit at an academic, quaternary freestanding children's hospital. Data were gathered over a single academic year (2015–2016). The clinical service model paired a fellow and a faculty member together as a clinical care team for 7 days.

Tool Development

To ensure the tool overcomes current assessment barriers and meets the needs of users, PCCM fellows and faculty participated in surveys, focus groups, and individual discussions. These identified perceptions of current assessment tools and guided tool development by articulating goals for the assessment tool. Needs assessment outcomes demonstrated the previous tool's length and complex, educationally-based language as key deterrents to timely completion and provision of narrative feedback, an important step in using the assessments to enhance learning. To create the assessment items, the PCCM EPA, “acute management of the critically ill patient, including those with underlying chronic disease,” served as the foundation for the WBA tool (figure 1).16 Initial meetings with faculty discussing tool development revealed unwillingness to provide entrustment decisions, because the scope of this EPA was too broad. In considering the literature defining EPAs, including number and breadth, and van der Vleuten's definition of utility, the PCCM EPA was subsequently divided into a list of sub-EPAs and OPAs.13,17 Creation of the initial list of sub-EPAs, based on patient management tasks, utilized the American Board of Pediatrics PCCM subspecialty examination content, the opinions of PCCM faculty with more than 5 years of experience, and the frequency of admission pathophysiology determined by sampling 3 months of admission diagnoses (figure 1). An acceptable level of granularity was still unmet for faculty agreement to assess based on the sub-EPAs. A third round of refinement provided OPAs based on more granular management tasks within each sub-EPA, utilizing the mechanism described by Warm and colleagues (figure 1).14 These were defined as sub-EPAs and OPAs rather than specifications or nested activities from larger EPAs, because they continued to constitute a distinct and independent unit of work that a fellow must be entrusted to perform and, within themselves, included several specifications such as interpreting vital signs and laboratory data, managing patient-assistive technology, and recognizing when to escalate or deescalate therapies.13 The final tool included 20 sub-EPAs with individualized OPAs. Full details on tool development, faculty and fellow training, and implementation are available as online supplemental material.

figure 1. Development of Pediatric Critical Care Medicine Entrustable Professional Activity–Based Assessment Toolfigure 1. Development of Pediatric Critical Care Medicine Entrustable Professional Activity–Based Assessment Toolfigure 1. Development of Pediatric Critical Care Medicine Entrustable Professional Activity–Based Assessment Tool
figure 1 Development of Pediatric Critical Care Medicine Entrustable Professional Activity–Based Assessment Tool

Citation: Journal of Graduate Medical Education 11, 4; 10.4300/JGME-D-18-01006.1

To disseminate the tool, a password-protected, conditional branching SurveyMonkey form was developed, pilot tested by faculty and fellows, and downloaded to division-provided cell phones. Tool completion required selection of the EPA-OPA (jointly chosen by faculty-fellow dyads) and faculty members to provide the level of entrustment (table) and a narrative explanation for the choice and requirements for advancement to the next level. Finally, 90-minute sessions were held separately for faculty and fellows to introduce the new tool.

table Levels of Entrustment and Definitions Provided to Faculty and Fellows

            
              table

Data Analysis for Validity Evidence

A mixed-methods analysis was conducted using sources of evidence outlined in Messick's validity framework (content, response process, internal structure, relations to other variables evidence, and consequences) in addition to acceptability as described in van der Vleuten's model of assessment tool utility.1719 Participant demographic characteristics were obtained. Acceptability evaluation included analysis of feasibility and compliance. Time required for tool completion, the main complaint for previous assessment tools, served as the feasibility evaluation. Compliance was determined, for faculty and fellows, using each person's expected number of completed assessments based on service requirements. Because WBAs provide unequal numbers of observations per trainee from different evaluators, reliability was estimated using an unbalanced random-effects model of generalizability theory with EPAs crossed with learners, where learners were the object of measurement, and variance components of the assessment looking for phi-coefficient > 0.3 and a person variance > 5%, respectively.2022 Analysis of variance was used to determine the significance between mean entrustment levels using SPSS Statistics 23 (IBM Corp, Armonk, NY).

A qualitative analysis of faculty-identified rationale behind entrustment decisions was conducted using iterative thematic coding processes. A hybrid deductive and inductive approach began with 5 factors influencing entrustment decisions identified in the literature.23,24 Themes corresponding to these factors as well as inductively and iteratively identified subthemes were coded in faculty narrative comments describing the rationale behind the entrustment level assessed. Narrative comments were independently coded with discussion and refinement of differences in coding. Coding was iterative, allowing for review and recoding of previous narratives until full agreement was reached. The analysis was completed by 2 authors, 1 from the implementation institution (A.R.E.) and 1 from outside the institution (S.S.).

The study was approved by the Institutional Review Boards of Washington University School of Medicine and the University of Illinois at Chicago.

Results

Demographic Characteristics

Of the 12 fellows included in the analysis, 5 (42%) were first-year fellows, 3 (25%) were second-year fellows, and 4 (33%) were third-year fellows. Of the 15 faculty, 4 (27%) were instructors, 10 (67%) were assistant professors, and 1 (7%) was an associate professor. Faculty were 60% (9 of 15) male and fellows were 42% (5 of 12) male.

Acceptability Evidence

Acceptability was determined using feasibility, based on the time required to complete the tool and compliance with twice-weekly assessment tool completion. On average, full completion of the tool (initiation of the assessment to faculty signature) required 4.6 minutes, with a range of 1 to 14.3 minutes. To create dual responsibility over assessment, the tool was downloaded on division-provided cell phones for the fellows. As such, compliance was determined for individual faculty and fellows as the percentage of assessments completed compared to those expected. Faculty were 50% to 100% compliant with twice-weekly completion. All faculty completed assessments at least once each service week. Fellows demonstrated wider variability with 28% to 100% compliance.

Content Validity Evidence

Content validity evidence sought to demonstrate that the items selected for assessment were representative.18,25 Twenty sub-EPAs were chosen a priori as the assessment items using experts. A total of 171 assessments were completed. Sub-EPA assessments ranged from 0 to 24. Only 1 received no assessments with a median of 6 assessments per sub-EPA. Of these 20 sub-EPAs, 19 (95%) were assessed at least once with 14 (70%) assessed at least 5 times. When compared to admission diagnoses, the frequency of sub-EPA assessment demonstrated similar distribution suggesting appropriate choices for assessment items. Individual OPA assessments ranged from 0 to 9 with most receiving 1 assessment.

Internal Structure Validity Evidence

Evidence for internal structure validity sought to demonstrate reliability of trainee entrustment level assigned by showing that differences in entrustment were due to the trainee, not assessors. Person variance (true variance) based on generalizability theory was 7.2% of total variance. The phi-coefficient reliability between person and score was 0.65.

Response Process Evidence

Response process validity evidence sought to demonstrate that the responses provided were accurate by understanding faculty rationale for providing a given entrustment level. Medical education literature has identified 5 factors that influence entrustment decisions: (1) trainee factors, (2) supervisor factors, (3) trainee-supervisor relationship factors, (4) task-based factors, and (5) factors regarding the context in which the entrustment decision occurs.2630 These 5 factors influencing entrustment decisions served as the initial categories for thematic analysis to determine if faculty were using the same factors to assign entrustment levels.

In total, 171 faculty comments were included for analysis. Within the thematic analysis, each of the 5 previously identified trust influencers appeared, and an additional 7 subthemes were identified within “trainee” (figure 2). Due to the complexity of select faculty narratives, multiple codes were applied to some comments. Descriptions of themes and subthemes as well as illustrative comments are included in figure 2. As multiple codes were often applied to comments, emphasis added by the authors identifies specific phrases or words leading to the application of the described code. “Trainee” was the most commonly applied code, seen in 163 narratives. “Task,” “supervisor,” “relationship,” and “context/setting” were applied with decreasing frequency.

figure 2. Organizational Model for Categories and Codes of Faculty Narratives Describing What Influenced Entrustment Decisionfigure 2. Organizational Model for Categories and Codes of Faculty Narratives Describing What Influenced Entrustment Decisionfigure 2. Organizational Model for Categories and Codes of Faculty Narratives Describing What Influenced Entrustment Decision
figure 2 Organizational Model for Categories and Codes of Faculty Narratives Describing What Influenced Entrustment Decision

Citation: Journal of Graduate Medical Education 11, 4; 10.4300/JGME-D-18-01006.1

Relations to Other Variables' Validity Evidence

Relations to other variables' evidence is derived from comparing total scores to subsets scores, correlation of novel assessment tool outcomes to a gold standard, or correlation with an expected outcome.19 Because a gold standard for WBA is absent, correlation with an expected outcome and the ability of the tool to differentiate between years of fellowship training provided evidence of relations to other variables. Mean entrustment levels (figure 3) increased significantly between training years across all EPAs (P = .001; fellowship year 1 [FY-1]: mean 2.43, SD 0.38; FY-2: mean 3.25, SD 0.17; FY-3: mean 3.51, SD 0.28). We also sought to understand the impact of task complexity, as this could be used by the program to ensure trainees were managing more complicated patients and/or that trainees were given more entrustment as training advanced. The sub-EPA “Low Compliance/Hypoxemic Respiratory Failure,” the most frequently assessed sub-EPA for all fellow years with 24 total assessments, was used to understand this educational impact. Analysis of this sub-EPA comparing average OPA complexity to year of training did not reveal significant differences. First-year fellows demonstrated a mean complexity of 2.77 out of 5 (SD 0.73) compared to 3.57 (SD 0.79) and 3.14 (SD 1.21) in FY-2s and FY-3s, respectively (P = .52; figure 4). When assessed across the same complexity level, an increase in entrustment level was seen between years of fellowship training, but it did not reach statistical significance. First-year fellows demonstrated a mean entrustment level of 2.38 (SD 0.51) compared to 3.14 (SD 0.38) in FY-2s and 3.86 (SD 0.90) in FY-3s (P = .08).

figure 3. Mean Entrustment Level Across All Sub-EPAs by Year of Fellowship Trainingfigure 3. Mean Entrustment Level Across All Sub-EPAs by Year of Fellowship Trainingfigure 3. Mean Entrustment Level Across All Sub-EPAs by Year of Fellowship Training
figure 3 Mean Entrustment Level Across All Sub-EPAs by Year of Fellowship Training

Citation: Journal of Graduate Medical Education 11, 4; 10.4300/JGME-D-18-01006.1

figure 4. Subanalysis of Impact of Average OPA Complexity on Assigned Entrustment Level Using Most Frequently Assessed Sub-EPA Low Compliance/Hypoxemic Respiratory Failure Demonstrated No Significant Differences Between Year of Fellowship Trainingfigure 4. Subanalysis of Impact of Average OPA Complexity on Assigned Entrustment Level Using Most Frequently Assessed Sub-EPA Low Compliance/Hypoxemic Respiratory Failure Demonstrated No Significant Differences Between Year of Fellowship Trainingfigure 4. Subanalysis of Impact of Average OPA Complexity on Assigned Entrustment Level Using Most Frequently Assessed Sub-EPA Low Compliance/Hypoxemic Respiratory Failure Demonstrated No Significant Differences Between Year of Fellowship Training
figure 4 Subanalysis of Impact of Average OPA Complexity on Assigned Entrustment Level Using Most Frequently Assessed Sub-EPA Low Compliance/Hypoxemic Respiratory Failure Demonstrated No Significant Differences Between Year of Fellowship Training

Citation: Journal of Graduate Medical Education 11, 4; 10.4300/JGME-D-18-01006.1

Discussion

The novel EPA-OPA tool described in this study demonstrated appropriate validity evidence to support its use as a tool to guide assessment for learning and address some of the current concerns surrounding WBA tools. The results of this study also identified how the tool might be used as trainees advance and the faculty-documented reasons behind supervision decisions that can help guide learning.

The content validity evidence supports that the tool measured important practice activities, a key factor needed to support learning. Our work demonstrated that family communication, while not a pathophysiology- or diagnosis-based assessment item, was one of the most frequently assessed activities. Adult studies have identified family-physician communication as a key factor in satisfaction with care provided and the psychological well-being of family members of critically ill patients,3133 supporting the notion that “Family Communication” is an important practice activity for PCCM trainees developing trust to practice independently. Further support for using the outcomes of the assessment tool was demonstrated by its reliability and generalizability data. With learners as the object of measurement, sound reliability estimates and variance components estimates demonstrate the tool's ability to discriminate between entrustable decisions. These findings, similar to those described in other generalizability studies for WBA where person variance ranges from 5% to 11%, support using the tool to discriminate between high and low performance.2022

Feasibility of a WBA tool directly relates to its length, complexity, ability to fit direct observation into clinical workflow, and the overall burden placed on faculty to complete it.1,2,34,35 This, in turn, influences the number of assessments completed and the ability to mitigate natural variations in performance.1,2,34,36 These are important threats to reliability18,19 that impact use of the tool for reflection and learning. The overall high number of assessments completed suggests that the tool was feasible. High response rates likely resulted from 2 decisions in tool development: (1) the choice to use the EPAs as the tool's conceptual framework capitalizes on decisions intuitively made by clinical supervisors, and (2) the choice to involve frontline users in tool development to help decrease complexity and improve buy-in. Time constraints, especially in a busy clinical environment, significantly influence perceptions of feasibility and become detractors for completing WBA.35 Considering the brief time required for completion, the evidence presented here suggests the tool should not detract from clinical workflow. The feasibility demonstrated by this tool supports development of assessment tools for the other 2 PCCM-specific EPAs: “manage and coordinate care in pediatric critical care units for optimal patient outcomes” and “management of patients at the end of life.”16

The ability to use the assessment outcomes to guide reflection and learning is also influenced by alignment with outcomes of current assessments with the same construct. Examples include comparing to a gold standard (relationship to other variables' evidence), and appropriate rationale by the person completing the assessment for the decision made (response processes evidence).18,25 Unfortunately, “gold standard” assessment tools in WBA are lacking especially as the ACGME Competencies and Milestones were “never intended to serve as a regular assessment tool.”37 However, narrative comments by faculty provided important insight into the rationale used to determine the level of entrustment afforded to a trainee. Fellows can use these narrative comments and changes in entrustment level to analyze their performance and develop a learning plan to further advance entrustment. Trainee-based factors, specifically the themes of knowledge and competence, were the most frequently cited influences of entrustment and supervision decisions reported by faculty, which mirrors the literature.2729 Using these faculty narratives, fellows could develop a learning plan to read more about specific pathologies and improve their communication around decision-making. Once specific learning plans have been devised based on the assessment outcomes, fellows can use this information to guide subsequent faculty interactions, teaching, and assessments.

One unexpected finding was that, while mean entrustment levels increased by year of training, the average acuity of the practice activity assessed did not increase. One explanation is that the highest level of entrustment, “supervises others,” was being provided to lower acuity patients. Decisions of entrustment and supervision are, by their nature, about risk acceptance.2730 When allowing a trainee to supervise someone else, faculty may mitigate that risk by selecting lower acuity patients. Further evaluation will need to explore this idea.

This study has limitations. It was conducted at a single center with small numbers of faculty and fellows, limiting generalizability to other settings and subjects. Only 1 year of data are available for evaluation leading to cross-sectional analysis rather than longitudinal analysis in tool evaluation. The study did not determine if tool use affected trainee progress or whether trainees found the information useful. The amount of time to complete the tool and adherence rates for use of the tool are not adequate surrogate measures for tool acceptability to faculty, and additional concerns may exist.

Further studies will focus on determining how the outcomes of this assessment tool guide trainee learning, influence program changes, and can serve as an assessment of learning by understanding how achievement of individual OPA entrustment influences sub-EPA trust decisions.

Conclusion

The EPA-OPA tool described here demonstrated sufficient validity evidence to justify using the outcomes to guide assessment for learning and promote the development of entrustment. The processes of developing, implementing, and evaluating this tool can be broadly applied.

Copyright: 2019
pdf
<bold>
  <sc>figure</sc>
  1
</bold>
figure 1

Development of Pediatric Critical Care Medicine Entrustable Professional Activity–Based Assessment Tool


<bold>
  <sc>figure</sc>
  2
</bold>
figure 2

Organizational Model for Categories and Codes of Faculty Narratives Describing What Influenced Entrustment Decision


<bold>
  <sc>figure</sc>
  3
</bold>
figure 3

Mean Entrustment Level Across All Sub-EPAs by Year of Fellowship Training


<bold>
  <sc>figure</sc>
  4
</bold>
figure 4

Subanalysis of Impact of Average OPA Complexity on Assigned Entrustment Level Using Most Frequently Assessed Sub-EPA Low Compliance/Hypoxemic Respiratory Failure Demonstrated No Significant Differences Between Year of Fellowship Training


Author Notes

Editor's Note: The online version of this article contains full details on tool development, faculty and fellow training, and implementation of the workplace-based assessment tool.

Funding: The authors report no external funding source for this study.

Conflict of interest: The authors declare they have no competing interests.

The results were presented at the Society of Critical Care Medicine 47th Critical Care Congress, San Antonio, Texas, February 25–28, 2018.

Corresponding author: Amanda R. Emke, MD, MHPE, Washington University in St Louis, NWT CB 8116, One Children's Place, St Louis, MO 63110, 314.454.2678, emke_a@wustl.edu
Received: 23 Jan 2019
Accepted: 29 May 2019
  • Download PDF