Group Assessment of Resident Performance: Valuable for Program Director Judgment?

MA, MD,
PhD,
MD,
PhD, MD,
MD, PhD,
PhD, and
PhD
Online Publication Date: 01 Aug 2019
Page Range: 118 – 124
DOI: 10.4300/JGME-D-18-01069
Save
Download PDF

ABSTRACT

Background

Group discussion of resident performance is an emerging assessment approach in postgraduate medical education. However, groups do not necessarily make better decisions than individuals.

Objective

This study examined how group meetings concerning the assessment of residents take place, what information is shared during the meetings, and how this influences program directors' judgment of resident performance.

Methods

In 2017, the researchers observed 10 faculty group meetings where resident performance was discussed and interviewed the program directors within a month after the meetings. We used a thematic framework analysis to identify themes from the transcribed meetings and interviews.

Results

The information shared by group members during the meetings had 2 aims: (1) forming a judgment about the residents, and (2) faculty development. Most group members shared information without written notes, most discussions were not structured by the program director, the major focus of discussions was on residents with performance concerns, and there was a lack of a shared mental model of resident performance. The program directors who benefited most from the meetings were those who thought group members were engaged and summarized the information after every discussion.

Conclusions

Unstructured discussions and a lack of a shared mental model among group members impede effective information sharing about resident performance with a developmental approach. Structured discussions with an equal amount of discussion time for every resident and creating a shared mental model about the purpose of the discussions and the assessment approach could enhance use of a developmental approach to assessing resident performance.

Introduction

Program directors are responsible for assessing resident performance.1,2 Until recently, they primarily accomplished this individually by interpreting multiple assessment data points, with or without consulting faculty.35 In recent years, the competency-based medical education (CBME) approach in medical education has emphasized group decision-making regarding resident performance as a new assessment approach in postgraduate medical education.68 Since 2013, in the United States, the Accreditation Council for Graduate Medical Education has required clinical competency committees (CCCs) to determine resident performance, and other countries are following this example.6

Group decision-making related to resident performance is based on the concept that groups make better decisions than individuals if they discuss existing data and share and integrate new information, uniquely held by members.9 Group discussion can increase detection of resident problematic performance.8,1012 Yet, the literature on group decision-making shows that reality often falls short of expectations.8,13 Biases, such as holding on to an initial opinion, believing things because others do, judgment influenced by overreliance on consensus or by emotions instead of objective data,14 and lack of discussion intensity can lead to ineffective information sharing and poor decision-making.15,16 This may jeopardize the validity of the judgments about resident performance.

Literature on CCCs includes guidelines to set up a meeting8,1721 and reviews of literature on group decision-making that offer recommendations for maximizing the effectiveness of CCC processes.8,14 One study7 found most CCCs review resident performance by identifying problems, instead of using a developmental approach.6

The majority of studies on CCCs are conducted in the United States, where programs have acquired experience with creating and operating CCCs since 2013. It is unclear whether findings and recommendations apply to other nations and cultures. This information is relevant since many other nations are establishing CCCs.

Our study explored how group decision-making about resident performance takes place in the Netherlands, where CCCs are not yet a required component of assessment in postgraduate medical education. We also sought to understand how group discussions influence program directors' judgment of resident performance, since the purpose of CCCs is to advise program directors about residents' progress. We sought answers to the following questions: (1) What kind of information about residents is shared during a CCC meeting and how is this shared? (2) How does group information sharing influence program directors' judgment about resident performance? The results provide insight into current practices and contribute to the understanding of effective group discussion about resident performance.

Methods

Background and Participants

In the Netherlands, most residents complete their postgraduate medical training in both general hospitals and university medical centers. All programs are competency-based. Residents are trained and supervised by all staff members, but program directors (PDs) are solely responsible for the assessment of residents' progress. PDs are required to have an evaluation meeting with residents at least twice a year to provide feedback on their performance.1 Group meetings regarding resident assessment are not required, but most PDs organize meetings with faculty to discuss resident performance. The intent of these meetings is comparable to the purpose of CCCs in the United States: the group makes a decision about the level of performance of residents and advises the PD.

We purposefully sampled 10 group meetings in a range of medical specialties during which resident performance was discussed.22,23 We invited PDs for participation by e-mail.

Based on ethnographic principles,2426 we combined observations of program faculty meetings with interviews with the PD to answer the second research question.

Data Collection and Analysis

Data were collected from May to December 2017. CCC meetings were observed by 1 researcher (M.E.D. or I.A.S.), who recorded observations using a semistructured scheme based on Hauer and colleagues' narrative review on group decision-making8 that scored whether and how often certain interaction occurred.27 Meetings were audio recorded and transcribed, and transcripts and field notes were analyzed by 3 researchers (M.E.D., I.A.S., I.C.M.), using a thematic framework analysis28 interaction schema.8 The researchers discussed findings using constant comparison until agreement was reached. Within a month of the group meeting we interviewed the PDs, asking them to reflect on the findings from our group observations, including the influence of the meetings on PD assessment of resident performance. Interviews were based on a semistructured interview guide, conducted by M.E.D. or I.A.S., and lasted about 60 minutes. They also were audio recorded, transcribed, and analyzed using a thematic framework analysis.28 We selected key themes by analyzing and discussing the data within the interview categories: process, content, and result. After 5 meetings and interviews and again after 10 meetings and interviews, themes were discussed in the whole research team. Saturation was reached when all data could be analyzed using the existing themes and no new themes were identified. Observation and interview protocols and tools are provided as online supplemental material.

The study was approved by the ethical board of the Netherlands Association for Medical Education. We obtained informed consent from all participants.

Results

Participants represented 10 medical specialties from 4 Dutch university medical centers (table). Team size varied from 4 to 20 members, and all group members were program faculty. The number of residents discussed ranged from 2 to 32, and the duration of meetings ranged 22 to 96 minutes. All meetings were chaired by the PD.

table Characteristics of Participants

          
            table

The results for our research questions are presented below, and are supported by illustrative quotations from meetings or from PD interviews.

Sharing Information

Shared information during the group meetings had 2 aims: (1) forming a judgment and (2) faculty development.

Forming a Judgment

PDs were looking for divergent faculty opinions to form a broader picture of resident performance. One PD stated, “Without the faculty meeting I have no foundation [. . .] I always imagine a picture of an elephant. I describe its tail, someone else describes its trunk, and a third person its feet. We need each other to paint the whole picture.” PDs also desired specific examples of performance to provide feedback to residents in semiannual evaluation meetings: “I need to gather ammunition with concrete examples of behavior.”

In many cases, faculty put a label on residents, uttering sentences like, “If it were my mother [who needed surgery], I would drive by really fast”; “She performs surgery like a rusty gate”; or “That is a good resident, because he does not bother me.”

Faculty Development

Part of the shared information was to help develop faculty by talking about the way residents should be supervised. As one PD indicated: “The meeting is a moment in which we can discuss training-related subjects.” It is also seen as an opportunity for faculty members to learn from each other: “One hears how colleagues tackle things and how they judge residents.” PDs used the meeting as an opportunity to instruct faculty. For example, we observed that when faculty mentioned an incident with a resident, PDs often insisted on completing a mini Clinical Evaluation Exercise or an Objective Structured Assessment of Technical Skills. We also observed several discussions about the assessment criteria used to judge resident performance.

Interaction During Meetings

Most meetings included jokes, and often there was a giggly atmosphere and a lot of laughter. Almost none of the faculty members brought written notes to the meetings, with most sharing information by heart and frequently repeating each other's comments. Real discussions of the performance of residents were rare. One PD expressed disappointment: “They don't listen to each other; there is no discussion at all.” Most of the time, PDs listened to the repetitive comments without interference. They also did not encourage faculty to share new information. When asked about this, a PD answered: “I don't want to play bad cop. I don't want to spoil the good atmosphere in the group.” Drawing conclusions or summarizing information was hardly ever done. The exception was one PD who repeatedly summarized information and drew conclusions: “With a summary I ask for consensus in the group. . . . this is what I have to feed back to the resident.”

Influence on Judgment of Program Directors

Faculty Engagement

During the interviews, PDs indicated that the group's judgment of resident performance in meetings depended on the effort of the group members, and that members varied in how important they consider the meetings: “They check their agendas to find out what they can skip and this meeting is always the first thing to skip.” Low turnout was a problem for many PDs: “That is the flaw in forming a judgment.” PDs put a different value on the opinions of faculty: “There are faculty members I trust and others I don't take seriously. I have more trust in colleagues who are dedicated to training residents.”

All PDs pointed out that the danger in meetings is that faculty have a tendency to echo each other's opinions. PDs saw it as their obligation to filter opinions and only use information that seemed useful. One PD noted: “When mass hysteria arises I throw all that is said in the trash!” Another said some faculty members “constantly bring up the past, mostly a very distant and dirty past.” She explained she saw it as her obligation to ignore these negative stories.

According to the majority of PDs, meetings are an opportunity for faculty to unload their feelings about residents. As one PD illustrated, “They pour out all those opinions and that is that.” Another commented, “The meeting is a fixed moment of which faculty members know: Now I can finally tell something about that resident!” PDs noted that faculty often see the meeting as a brief time of leisure. “The meeting is a form of relaxation.” “It is about residents; it is not threatening.”

Resident Performance

All PDs found the meetings more influential when “problematic residents” are discussed. “We can make jokes about the good-functioning resident, but are serious about the problematic ones.” PDs explained that meetings are uncomplicated when residents function at the expected level: “If there are not too many problems, then it is easy, we don't have to think about it. But, if there are problems, then we suddenly have to start dealing with that resident.”

This suggests residents with performance concerns received more discussion time than high-functioning residents. One PD saw this as a problem, arguing that all residents must be guided to develop to a higher level: “It is not fair, because we can also help good residents to become excellent.” While residents with performance concerns received more discussion time, this rarely concluded with a plan of action. For residents without performance, groups did not make suggestions for how these residents could develop further.

Program Directors' Focus During the Meeting

PDs indicated the main focus of the meetings was on creating a group feeling among faculty members, and that it is important to emphasize that faculty as a group were responsible for training the residents as a group. Several PDs reported they did not create much structure for the meetings due to concerns of harming the positive group feelings, adding that faculty needed an opportunity to blow off steam and express their opinions about the residents. PDs closed noting they wished the meetings were more valuable to them in forming a judgment of resident performance.

A few PDs stated the main focus of meetings was on evaluating resident performance, and that they asked faculty members to complete assessment forms prior to meetings. They found it important to create a broad picture of each resident by aggregating opinions from different faculty members and using these to provide feedback to the residents. They summarized assessment information at the end of each discussion, and reported it as useful for judging resident performance.

Discussion

In most meetings the assessment conversations were not structured by the PD and predominantly focused on residents with performance concerns. PDs who found the meetings useful were those who reported that faculty members were engaged, who summarized the information at the end of each discussion, and who formulated feedback to the resident.

Biases and low intensity discussions contributed to ineffective information sharing, similar to the findings of other studies.1416 In many cases, faculty put a label on residents that was influenced by emotions rather than objective data. More problematic, faculty members frequently repeated each other's comments, real discussions were rare, and PDs did not interfere by asking for different opinions or objective data to provide a broader perspective on a trainee. Ineffective information-sharing by groups hampers good group decision-making,15 making it essential for group leaders to structure discussions. Approaches include letting members speak in a set order, giving every member an opportunity, encouraging them to speak, summarizing information to elicit discussion and new points of view, and asking for more information.8,29 When more information is shared, better decisions are possible, and meetings are more beneficial to PDs who have to make judgments about resident performance.8,15,29 To address PDs concerns structuring meetings, an option is to have the meeting chaired by another individual.8

Meetings were more useful when faculty members were engaged, yet were often skipped by faculty. One reason could be that CCCs are not mandatory in the Netherlands. In contrast to some faculty members, PDs valued the meetings as important for their role in assessing residents and providing feedback.

We found that group members lacked a shared understanding of the purpose of the meeting. Group performance improves with shared mental models—a common understanding of the purpose of the group's work, the task to be performed, and teamwork necessary to complete it.8,3032 This makes it important to explain to group members the purpose of the discussions and the task they are expected to perform.

Group discussions largely focused on residents with performance concerns, similar to previous studies of CCCs.7 This is in keeping with a “dwell time” model of medical education, assuming that most residents will be competent at the completion of a prescribed number of years of training.7 In contrast, introduction of CBME emphasizes individualized learning plans and paths to competence.33,34 To fully embrace CBME, it is necessary to let go of the habit of just identifying problems in resident performance, use a developmental assessment model, and guide all residents on individual paths to becoming competent physicians. Highly structured discussions with an equal amount of time for every resident and clearly explaining the aims of assessment for all group members promote such a developmental approach.

Limitations of our study include a sample from 1 nation, and interview data limited to PDs, which may overemphasize the perspectives of this group. Future research should seek to gather data from multiple perspectives, including residents, and should assess the generalizability of our findings to other nations and cultures.

Conclusion

Unstructured discussions and lack of a shared mental model among group members were common attributes of CCC meetings that hinder effective information sharing and discussion of resident performance. This reduces the impact of this information on PDs' judgments of resident performance and impedes a developmental assessment approach. Structuring discussions to ensure an equal amount of time for every resident and creating a shared mental model among the group members will contribute to an effective developmental approach for assessing resident performance.

References

  • 1
    Scheele F,
    Teunissen P,
    Luijk SV,
    Heineman E,
    Fluit L,
    Mulder H,
    et al.
    Introducing competency-based postgraduate medical education in the Netherlands. Med Teach. 2008;30(
    3
    ):248253. doi:10.1080/01421590801993022.
  • 2
    Accreditation Council for Graduate Medical Education. Common Program Requirements. https://www.acgme.org/Portals/0/PFAssets/ProgramRequirements/CPRs_2017-07-01.pdf. Accessed May 24, 2019.
  • 3
    Holmboe ES,
    Sherbino J,
    Long DM,
    Swing SR,
    Frank JR.
    The role of assessment in competency-based medical education. Med Teach. 2010;32(
    8
    ):676682. doi:10.3109/0142159X.2010.500704.
  • 4
    Iobst WF,
    Sherbino J,
    Cate OT,
    Richardson DL,
    Dath D,
    Swing SR,
    et al.
    Competency-based medical education in postgraduate medical education. Med Teach. 2010;32(
    8
    ):651656. doi:10.3109/0142159X.2010.500709.
  • 5
    Carraccio C,
    Englander R,
    Van Melle E,
    Ten Cate O,
    Lockyer J,
    Chan MK,
    et al.
    Advancing competency-based medical education: a charter for clinician-educators. Acad Med. 2016;91(
    5
    ):645649. doi:10.1097/ACM.0000000000001048.
  • 6
    Andolsek K,
    Padmore J,
    Hauer KE,
    Holmboe E.
    Clinical competency committees—a guidebook for programs. 2nd ed.
    Chicago, IL
    :
    Accreditation Council for Graduate Medical Education
    ; 2017. https://www.acgme.org/Portals/0/ACGMEClinicalCompetencyCommitteeGuidebook.pdf. Accessed May 24, 2019.
  • 7
    Hauer KE,
    Chesluk B,
    Iobst W,
    Holmboe E,
    Baron RB,
    Boscardin CK,
    et al.
    Reviewing residents' competence: a qualitative study of the role of clinical competency committees in performance assessment. Acad Med. 2015;90(
    8
    ):10841092. doi:10.1097/ACM.0000000000000736.
  • 8
    Hauer KE,
    Cate OT,
    Boscardin CK,
    Iobst W,
    Holmboe ES,
    Chesluk B,
    et al.
    Ensuring resident competence: a narrative review of the literature on group decision making to inform the work of clinical competency committees. J Grad Med Educ. 2016;8(
    2
    ):156164. doi:10.4300/JGME-D-15-00144.1.
  • 9
    Michaelsen LK,
    Watson WE,
    Black RH.
    A realistic test of individual versus group consensus decision making. J App Psych. 1989;74(
    5
    ):834839.
  • 10
    Hemmer PA,
    Hawkins R,
    Jackson JL,
    Pangaro LN.
    Assessing how well three evaluation methods detect deficiencies in medical students' professionalism in two settings of an internal medicine clerkship. Acad Med. 2000;75(
    2
    ):167173.
  • 11
    Thomas MR,
    Beckman TJ,
    Mauck KF,
    Cha SS,
    Thomas KG.
    Group assessments of resident physicians improve reliability and decrease halo error. J Gen Intern Med. 2011;26(
    7
    ):759764. doi:10.1007/s11606-011-1670-4.
  • 12
    Schwind CJ,
    Williams RG,
    Boehler ML,
    Dunnington GL.
    Do individual attendings' post-rotation performance ratings detect residents' clinical performance deficiencies? Acad Med. 2004;79(
    5
    ):453457.
  • 13
    Hill GW.
    Group versus individual performance: are N + 1 heads better than one? Psychol Bull. 1982;91(
    3
    ):517539.
  • 14
    Dickey CC,
    Thomas C,
    Feroze U,
    Nakshabandi F,
    Cannon B.
    Cognitive demands and bias: challenges facing clinical competency committees. J Grad Med Educ. 2017;9(
    2
    ):162164. doi:10.4300/JGME-D-16-00411.1.
  • 15
    Schultze T,
    Mojzisch A,
    Schulz-Hardt S.
    Why groups perform better than individuals at quantitative judgment tasks: group-to-individual transfer as an alternative to differential weighting. Organ Behav Hum Decis Processes. 2012;118(
    1
    ):2436. doi:10.1016/j.obhdp.2011.12.006.
  • 16
    Schulz-Hardt S,
    Mojzisch A.
    How to achieve synergy in group decision making: lessons to be learned from the hidden profile paradigm. Eur Rev Soc Psychol. 2012;23(
    1
    ):305343.
  • 17
    Chahine S,
    Cristancho S,
    Padgett J,
    Lingard L.
    How do small groups make decisions? A theoretical framework to inform the implementation and study of clinical competency committees. Perspect Med Educ. 2017;6(
    3
    ):192198. doi:10.1007/s40037-017-0357-x.
  • 18
    Colbert CY,
    Dannefer EF,
    French JC.
    Clinical competency committees and assessment: changing the conversation in graduate medical education. J Grad Med Educ. 2015;7(
    2
    ):162165. doi:10.4300/JGME-D-14-00448.1.
  • 19
    French JC,
    Dannefer EF,
    Colbert CY.
    A systematic approach toward building a fully operational clinical competency committee. J Surg Educ. 2014;71(
    6
    ):e22e27. doi:10.1016/j.jsurg.2014.04.005.
  • 20
    Ketteler ER,
    Auyang ED,
    Beard KE,
    McBride EL,
    McKee R,
    Russell JC,
    et al.
    Competency champions in the clinical competency committee: a successful strategy to implement milestone evaluations and competency coaching. J Surg Educ. 2014;71(
    1
    ):3638. doi:10.1016/j.jsurg.2013.09.012.
  • 21
    Promes SB,
    Wagner MJ.
    Starting a clinical competency committee. J Grad Med Educ. 2014;6(
    1
    ):163164. doi:10.4300/JGME-D-13-00444.1.
  • 22
    Marshall MN.
    Sampling for qualitative research. Fam Pract. 1996;13(
    6
    ):522525.
  • 23
    Pope C,
    van Royen P,
    Baker R.
    Qualitative methods in research on healthcare quality. Qual Saf Health Care. 2002;11(
    2
    ):148152.
  • 24
    Atkinson P,
    Pugsley L.
    Making sense of ethnography and medical education. Med Educ. 2005;39(
    2
    ):228234. doi:10.1111/j.1365-2929.2004.02070.x.
  • 25
    Slootweg IA,
    Scherpbier A,
    van der Leeuw R,
    Heineman MJ,
    van der Vleuten C,
    Lombarts KM.
    Team communication amongst clinical teachers in a formal meeting of post graduate medical training. Adv Health Sci Educ Theory Pract. 2016;21(
    1
    ):207219. doi:10.1007/s10459-015-9627-8.
  • 26
    Ng S,
    Stooke R,
    Regan S,
    Hibbert K,
    Schryer C,
    Phelan S,
    et al.
    An institutional ethnography inquiry of health care work in special education: a research protocol. Int J Integr Care. 2013;13:e033.
  • 27
    Hannah DR,
    Lautsch BA.
    Counting in qualitative research: why to conduct it, when to avoid it, and when to closet it. J Manag Inq. 2011;20(
    1
    ):1422.
  • 28
    Pope C,
    Ziebland S,
    Mays N.
    Qualitative research in health care: analysing qualitative data. BMJ. 2000;320(
    7227
    ):114116. doi:10.1136/bmj.320.7227.114.
  • 29
    Mesmer-Magnus JR,
    Dechurch LA.
    Information sharing and team performance: a meta-analysis. J Appl Psychol. 2009;94(
    2
    ):535546. doi:10.1037/a0013773.
  • 30
    Jonker CM,
    Van Riemsdijk MB,
    Vermeulen B.
    Shared mental models: a conceptual analysis. In:
    De Vos M,
    Fornara N,
    Pitt JV,
    Vouros G,
    eds. Coordination, Organizations, Institutions, and Norms in Agent Systems VI. COIN 2010. Lecture Notes in Computer Science. Vol 6541.
    Berlin, Germany
    :
    Springer
    ; 2011:132151.
  • 31
    Janis IL.
    Groupthink. Psych Today. 1971;5(
    6
    ):4346.
  • 32
    Lu L,
    Yuan YC,
    McLeod PL.
    Twenty-five years of hidden profiles in group decision making: a meta-analysis. Pers Soc Psychol Rev. 2012;16(
    1
    ):5475. doi:10.1177/1088868311417243.
  • 33
    Nasca TJ,
    Philibert I,
    Brigham T,
    Flynn TC.
    The next GME accreditation system—rationale and benefits. N Engl J Med. 2012;366(
    11
    ):10511056. doi:10.1056/NEJMsr1200117.
  • 34
    Nasca TJ,
    Weiss KB,
    Bagian JP,
    Brigham TP.
    The accreditation system after the “next accreditation system.” Acad Med. 2014;89(
    1
    ):2729. doi:10.1097/ACM.0000000000000068.
Copyright: 2019
pdf

Author Notes

Editor's Note: The online version of this article contains a semistructured observation scheme and a semistructured interview guide.

Funding: The study was funded by the Dutch Federation of Medical Specialists.

Conflict of interest: The authors declare they have no competing interests.

The authors would like to thank the participants for their cooperation and sincere input.

Corresponding author: Marrigje E. Duitsman, MA, MD, Radboud Health Academy, Gerard van Swietenlaan 4, PO Box 9101, 6500 HB Nijmegen, the Netherlands, +31648495035, marloes.duitsman@radboudumc.nl
Received: 13 Dec 2018
Accepted: 23 Apr 2019
  • Download PDF