Residency Applicant Preferences of Online Systems for Scheduling Interviews
ABSTRACT
Background
Residency applicants often have difficulty coordinating interviews with multiple programs. An online scheduling system might improve this process.
Objective
The authors sought to determine applicant mean time to schedule interviews and satisfaction using online scheduling compared with manual scheduling.
Methods
An electronic survey was sent to US graduates applying to any of 6 emergency medicine programs in the 2014–2015 application cycle. Of the participant programs, 3 used an online system and 3 did not. Applicants were asked to report estimated time to schedule with the online system compared to their average time using other methods, and to rate their satisfaction with the scheduling process.
Results
Of 1720 applicants to at least 1 of the 6 programs, 856 completed the survey (49.8%). Respondents reported spending less time scheduling interviews using the online system compared to other systems (median of 5 minutes [IQR 3–10] versus 60 minutes [IQR 15–240], respectively, P < .0001). In addition, applicants preferred using the online system (93.6% versus 1.4%, P < .0001.) Applicants were also more satisfied with the ease of scheduling their interviews using the online system (91.5% versus 11.0%, P < .0001) and felt that the online system aided them in coordinating travel arrangements (74.7% versus 41.5%, P < .01.)
Conclusions
An online interview scheduling system is associated with time savings for applicants as well as higher satisfaction among applicants, both in ease of scheduling and in coordinating travel arrangements. The results likely are generalizable to other medical and surgical specialties.
Introduction
Scheduling interviews during the residency selection season can be challenging and time consuming for applicants and programs. Applicants have to coordinate interviews at multiple programs, and offers may not be extended in a way conducive to easy scheduling. Additionally, applicants are often involved in active patient care during business hours, which may prevent them from calling or reaching program coordinators in a timely fashion to secure their most desired interview date.
Several scheduling systems are available to allow applicants access to online scheduling. A prior study of a single site online scheduling system showed a 77% applicant preference for scheduling online.1 Similarly, our pilot study in 2010–2011 reported that 87.5% of emergency medicine (EM) applicants preferred an online program.2
Our primary aim was to measure EM applicant satisfaction and perceived time savings using a commercially available Internet scheduling program compared to traditional manual scheduling.
Methods
This multicenter study surveyed applicants who applied to a sample of EM residency programs during the 2014–2015 application cycle. We formed a research study group consisting of 6 EM residency programs from a broad geographic distribution (Pacific, Pacific Northwest, Midwest, South, Northeast, and mid-Atlantic). Three programs use a commercial online scheduling platform (Interview Broker [The Tenth Nerve LLC, Lewes, DE]; www.interviewbroker.com), and 3 did not use any online scheduling program but rather relied on e-mails and phone calls between applicants and programs. No participating residency program had a financial interest in the online scheduler. Each program sent an e-mail with a survey link to all applicants who applied to its residency. We used the US applicants to the 6 representative programs as a purposive sample of the national EM applicant pool. No personal applicant information was shared between programs to ensure confidentiality. While applicants may have received more than 1 link to complete the survey, the survey program prohibits more than 1 response from an applicant based on IP address. A list of unique Electronic Residency Application Service identification numbers was sent to the coordinating site to generate the list of possible respondents and eliminate redundancy. The survey was sent after Match lists were submitted but before the National Resident Matching Program rank lists were released.
The survey was previously developed for use during a pilot study2 and then modified by the primary department faculty with residency leadership experience to improve content validity; such modifications involved change in format only. It was further reviewed by 6 experienced program directors from the participating sites. The revised survey was then reviewed with representative chief residents at the primary institution to optimize response process validity.3 Finally, the survey was piloted with all EM residents within the primary institution's program to further support its validity.4 Applicants were asked to estimate the time it took to schedule with the online system compared to other methods, and to rate their satisfaction with scheduling on a 5-point anchored scale.
The study protocol was granted expedited approved by the Institutional Review Board at the primary institution and approved at each participating institution.
Data analysis followed the assumption that questions regarding subjects' experiences with and without the software should be treated as nonindependent observations, and that the modified Likert scale should not be treated as an interval variable. We dichotomized the scale to “satisfied” and “less than satisfied,” which we analyzed with McNemar's test for nonindependent observations. The choice of these statistical tests limits the hypothesis testing in many cases to only P values.
Results
Of the 1720 unique US applicants to the 6 training programs, 856 applicants (49.8%) completed the survey (table). The sample represented approximately half of all US graduates applying to EM programs. Applicants' self-reported median time to schedule interviews using the online system was 5 minutes (interquartile range [IQR] 3–10) compared to a median of 60 minutes (IQR 15–240) for those programs not using the online system (P < .0001). In addition, 783 applicants (91.5%) who used the online platform were “satisfied” with the ease of scheduling interviews compared to 94 respondents (11.0%) who were “satisfied” with the typical system (P < .0001; figure 1).



Citation: Journal of Graduate Medical Education 8, 5; 10.4300/JGME-D-16-00072.1
Among applicants who had to change interview dates, 792 of 851 respondents (93.1%) were “satisfied” or “somewhat satisfied” in their ability to change their date online. For programs not using the online system, 243 of 847 respondents (28.7%) were “satisfied” or “somewhat satisfied” with the process, whereas the majority (604, 71.3%) described themselves as “neither satisfied or unsatisfied,” “somewhat unsatisfied,” or “unsatisfied” with the process of rescheduling (P < .0001), coded dichotomously as “less than satisfied.”
Overall, 801 of the 856 respondents (93.6%) reported they preferred the online system (figure 2). Only 12 (1.4%) preferred the nononline system, and 43 (5.0%) stated they had no preference.



Citation: Journal of Graduate Medical Education 8, 5; 10.4300/JGME-D-16-00072.1
Discussion
The results of this multicenter study suggest that residency applicants prefer the online platform for scheduling interviews, consistent with prior smaller studies.2 We hypothesize that applicants prefer online scheduling because it may conform to their schedules, and is consistent with their access to and comfort with technology. An online system allows scheduling or changing interviews at any hour. Similarly, the online system may allow applicants to rapidly change interview dates to coordinate interviews at other programs. Interestingly, respondents noted that the time to schedule an interview with the online system starting from receipt of the invitation was less than 5 minutes for half of all applicants.
An online system may substantially alter the standard mechanism for scheduling interviews. We view the high level of satisfaction as indirect evidence of the fidelity of data communication in the online system.
While we did not produce a cost benefit analysis due to the high variability of coordinator salaries across the country, it is possible that the increased satisfaction and the relatively low cost of the product (approximately $2 per interview) would result in a substantial return on investment.
During the 2014–2015 interview season, approximately 400 programs, mostly in EM and internal medicine, used Interview Broker, which initially became available in the 2010–2011 application cycle (e-mail conversation with Interview Broker Accounts Team, August 2014). It is not clear how many programs are currently using online interview tools as the number of commercially available products has increased.
Further research might include vigorous cost benefit analyses of online scheduling programs and the study of program director and program coordinator satisfaction with such programs.
Limitations of this study include a relatively low response rate (49.8%). However, the overwhelming preference by participants for the online system (93.6%) suggests it is the preferred approach for applicants. The data identifying the time spent scheduling interviews were self-reported, and may be subject to recall bias.
In addition, the coordinators at programs not using the online system were not standardized in terms of customer service training or scripting. Respondents were asked to reply based on their experiences at all programs they applied to, not just the ones participating in the study. Furthermore, we could not ensure that each respondent had both types of experiences, which could influence a respondent's ability to compare.
Conclusion
The results of this multicenter study suggest that EM residency applicants prefer an online scheduling system when compared to a nononline system. They found it beneficial when initially scheduling an interview and changing interview dates, and recalled that it saved a significant amount of time.

Rated Level of Satisfaction With Each Aspect of Scheduling Among Applicants

Preference Regarding Scheduling
Author Notes
Funding: The authors report no external funding source for this study.
Conflict of interest: The authors declare they have no competing interests.



