Abstract
Purpose:
In academic plastic surgery, there is a paucity of data examining the relationship between program rank, faculty training history, and production of academic program graduates. The purpose of this study is to determine objective faculty characteristics that are associated with a high program reputation.
Methods:
Accreditation Council for Graduate Medical Education-accredited integrated Plastic and Reconstructive Surgery (PRS) programs were ranked using Doximity and divided into Top-quartile programs and Other programs. Accredited medical schools were ranked using U.S. News and World Report. Individual faculty profiles were reviewed on program websites for information on prior training.
Results:
Seventy-nine programs with 712 faculty were identified and objectively analyzed. Compared to Other PRS programs, Top-quartile programs had a higher proportion of faculty that trained at Top-quartile residency programs (P < 0.0001) and Top-quartile medical schools (P < 0.0001). Top-quartile programs also had the highest proportion of faculty that trained at the same institution for fellowship (P = 0.0001), residency (P = 0.03), medical school (P = 0.4), or any prior training (medical school, residency, or fellowship) (P = 0.002). Top-quartile programs were associated with the largest total faculty size (P < 0.0001) and the largest number of graduates entering the field of academic plastic surgery (P < 0.0001).
Conclusions:
Program reputation is associated with PRS faculty selection and production. Top-ranked programs are more likely to have faculty that previously trained at the same institution or at top-ranked programs. Top-ranked programs are more likely to graduate residents that will become academic plastic surgeons.
INTRODUCTION
Integrated plastic surgery training is recognized as one of the most competitive residency positions in the United States, with only 172 slozts accredited in 2019 by the Accreditation Council for Graduate Medical Education (ACGME). However, obtaining one of those positions is only a small part of a puzzle to securing further academic appointment as an attending physician. Obtaining an academic appointment in plastic and reconstructive surgery (PRS) remains challenging and elusive to many prospective candidates. While the total number of full-time faculty positions affiliated with PRS residency training programs is steadily increasing, the number of open positions on an annual basis is limited and therefore the caliber of credentials necessary for obtaining a full-time faculty position remains high.1,2 For candidates interested in an academic position, defining the academic background of typical faculty would be of significant value. However, there is limited literature examining this topic.
There are many factors that may influence the selection of plastic surgery faculty, including pedigree, fellowship training, research productivity, advanced degrees, and intangible likability, with “top” programs being associated to a greater extent with metrics of scholarly activity.3–6 To date, no study in the peer-reviewed literature has performed an in-depth analysis to elucidate the attributes of academic plastic surgeons. Since higher-ranked PRS programs are associated with more academically productive faculty, prior academic success might be a possible factor influencing faculty selection.4 It is also possible that higher-ranked programs may indeed both select and produce future academic plastic surgeons with prior evidence of high scholarly impact.7
Thus, the primary goal of this investigation is to clearly characterize academic plastic surgeons and secondarily to determine if there are measurable differences between surgeon characteristics and program quality using objective data. We hypothesize that graduating from a top-ranked PRS residency, having advanced degrees, and completing fellowship training might be characteristics associated with faculty at top PRS programs. The results of this study can stimulate the plastic surgery community to reflect on what applicant characteristics are more likely to make a successful academic plastic surgeon and what program characteristics might encourage trainees to pursue an academic career path.
METHODS
PRS Program and Faculty Database Collection
A list of all integrated PRS residency programs in the United States was compiled from the ACGME. To ensure accuracy and completeness, this list was cross-referenced with the Doximity Residency Navigator and the National Residency Match Program (NRMP) databases. Publicly available websites for each integrated PRS program were accessed in September 2018, and a list of full-time core faculty members was compiled. Faculty members were excluded from analysis if they held the title of professor emeritus, adjunct professor, affiliated professor, clinical instructor, or part-time professor. If a program did not have a faculty list on its website, the program coordinators were contacted for missing information. Individual faculty profiles were reviewed for information on academic rank, graduate degrees, medical school, residency program, and fellowship program attended.
Reputation Data
Integrated PRS programs were divided into two groups for analysis: (1) Top-quartile programs and (2) Other programs. based on reputation rankings determined by the Doximity Residency Navigator. Doximity uses transparent methodology and is well established in ranking residency programs (since 2014). Reputation rankings of PRS programs established by Doximity are determined by peer nominations from board certified plastic surgeons that are adjusted for self-voting. Top programs change over time as the designation is dynamic from year to year and dependent on factors such as leadership, prominence of faculty, unique/novel clinical programs (ie, for composite tissue allotransplantation, lymphedema surgery, and transgender surgery), and research output. Doximity Residency Navigator is a tool that residency applicants use to shape their opinion on programs of interest to then select an integrated plastic surgery training program. The database is one of the top three most commonly used resources to investigate residency programs of interest, utilized by medical students applying into any specialty.8 Up to 78% of students surveyed in one study consider it a valuable resource, with recent data suggesting that Doximity rankings are influential enough to cause up to half of medical students reading them to modify their residency application list.8,9 For these reasons, we felt that Doximity, although not perfect, was a good ranking system to use for this analysis. Additionally, Top-quartile-accredited U.S. medical schools were divided based on rankings, as determined by the U.S. News and World Report for Best Medical Schools–Research Methodology. Caribbean and international medical schools were excluded from analysis, as were non-plastic surgery residency programs.
Statistical Analysis
All analyses were completed on GraphPad Prism (GraphPad Software, Inc., San Diego, Calif.). Numerical variables were analyzed using two-tailed Student’s t test. The P value cutoff for statistical significance was set at 0.05. All faculty data were analyzed as “percent faculty” to account for program size variation between groups.
RESULTS
Construction of a Dataset to Compare Top-Quartile Programs to All Others
A total of 79 integrated PRS residency programs with 712 faculty meeting inclusion criteria were included for analysis. Programs were then split into two groups: Top-quartile programs (ranks 1–20) and Other programs (ranks 21–79). Integrated PRS residency programs with their corresponding rank and region are listed in Table 1.
Table 1.
U.S. Integrated Plastic and Reconstructive Surgery Residency Programs
| Rank | Program | Region |
|---|---|---|
| 1 | New York University School of Medicine | NE |
| 2 | University of Texas Southwestern Medical School | S |
| 3 | UPMC Medical Education | NE |
| 4 | Johns Hopkins University/University of Maryland | NE |
| 5 | Brigham and Women’s Hospital/Harvard Medical School | NE |
| 6 | University of Pennsylvania Health System | NE |
| 7 | University of Michigan Health System | MW |
| 8 | Stanford Health Care-Sponsored Stanford University | W |
| 9 | Emory University School of Medicine | S |
| 10 | University of Washington | W |
| 11 | Washington University/B-JH/SLCH Consortium | MW |
| 12 | University of California (San Francisco) | W |
| 13 | University of Southern California/LAC+USC Medical Center | W |
| 14 | McGaw Medical Center of Northwestern University | MW |
| 15 | University of Wisconsin Hospitals and Clinics | MW |
| 16 | UCLA David Geffen School of Medicine/UCLA Medical Center | W |
| 17 | Duke University Hospital | S |
| 18 | Zucker School of Medicine at Hofstra/Northwell | NE |
| 19 | Cleveland Clinic Foundation | MW |
| 20 | University of Chicago | MW |
| 21 | Baylor College of Medicine | S |
| 22 | New York Presbyterian Hospital (Columbia and Cornell Campus) | NE |
| 23 | MedStar Health/Georgetown University Hospital | NE |
| 24 | Ohio State University Hospital | MW |
| 25 | Yale-New Haven Medical Center | NE |
| 26 | Mayo Clinic College of Medicine and Science (Rochester) | MW |
| 27 | Montefiore Medical Center/Albert Einstein College of Medicine of Yeshiva University | NE |
| 28 | University of North Carolina Hospitals | S |
| 29 | Mayo Clinic College of Medicine and Science (Arizona) | W |
| 30 | Southern Illinois University | MW |
| 31 | University of Minnesota | MW |
| 32 | University of California (San Diego) Medical Center | W |
| 33 | Loma Linda University Health Education Consortium | W |
| 34 | University of Florida | S |
| 35 | Icahn School of Medicine at Mount Sinai | NE |
| 36 | University of Texas Medical Branch Hospitals | S |
| 37 | Penn State Milton S Hershey Medical Center | NE |
| 38 | Rutgers New Jersey Medical School | NE |
| 39 | Spectrum Health/Michigan State University | MW |
| 40 | University of California Davis Health | W |
| 41 | Indiana University School of Medicine | MW |
| 42 | University of California (Irvine) | W |
| 43 | University of South Florida Morsani | S |
| 44 | University of Utah Medical Center | W |
| 45 | Louisiana State University School of Medicine | S |
| 46 | Wake Forest University School of Medicine | S |
| 47 | University of Virginia Medical Center | S |
| 48 | University of Kentucky College of Medicine | S |
| 49 | University of Kansas School of Medicine | MW |
| 50 | University of Cincinnati Medical Center/College of Medicine | MW |
| 51 | University of Colorado School of Medicine | W |
| 52 | University of Tennessee College of Medicine | S |
| 53 | Brown University | NE |
| 54 | Albany Medical Center | NE |
| 55 | University of Rochester | NE |
| 56 | Jackson Memorial Hospital/Jackson Health System | S |
| 57 | Case Western Reserve University/University Hospitals Cleveland Medical Center | MW |
| 58 | University of Massachusetts | NE |
| 59 | Medical University of South Carolina College of Medicine | S |
| 60 | Medical College of Wisconsin Affiliated Hospitals | MW |
| 61 | Mary Hitchcock Memorial Hospital/Dartmouth-Hitchcock | NE |
| 62 | St Louis University School of Medicine | MW |
| 63 | University of Missouri–Columbia | MW |
| 64 | Virginia Commonwealth University Health System | S |
| 65 | Carilion Clinic-Virginia Tech Carilion School of Medicine | S |
| 66 | Texas A&M College of Medicine-Scott and White Medical Center (Temple) | S |
| 67 | Loyola University Medical Center | MW |
| 68 | University of New Mexico School of Medicine | W |
| 69 | Wright State University | MW |
| 70 | West Virginia University School of Medicine | S |
| 71 | Oregon Health & Science University | W |
| 72 | University of Mississippi School of Medicine | S |
| 73 | University of Nevada Las Vegas (UNLV) School of Medicine | W |
| 74 | Nassau University Medical Center | NE |
| 75 | Lehigh Valley Health Network/University of South Florida College of Medicine | NE |
| 76 | Cooper Hospital-University Medical Center | NE |
| 77 | Geisinger Health System | NE |
| 78 | Lahey Clinic | NE |
| 79 | Palmetto Health/University of South Carolina School of Medicine | S |
The solid line indicates the division between Top-quartile and other PRS programs.
NE, Northeast; MW, Midwest; S, South; W, West.
Academic Rank and Productivity
Faculty rank was used as an approximate surrogate for academic productivity given the known association between higher academic rank of plastic surgery faculty and elevated research output as determined by multiple metrics, including the Hirsch index, g-index, number of publications, and number of citations.3,5,10,11 In this study, one Top-quartile PRS program and eight Other programs did not offer information on faculty academic rank on their websites (full, associate, or assistant professor status), so those faculty members were excluded from analysis on academic rank. Of the programs included in the analysis, there was no statistically significant difference between Top-quartile and Other programs in the proportion of full professors (30% versus 29%, P = 0.8), associate professors (26% versus 27%, P = 0.9), and assistant professors (37% versus 44%, P = 0.2) (Fig. 1A–C).
Fig. 1.

Faculty by academic rank. A, full professors: Top-quartile programs 30%, Other programs 29%, P = 0.8. B, associate professors: Top-quartile programs 26%, Other programs 27%, P = 0.9. C, assistant professors: Top-quartile programs 37%, Other programs 44%, P = 0.2.
Academic Degrees
Since completion of additional postgraduate training is associated with higher scientific productivity, we aimed to investigate the relationship between program rank and additional academic training.12 We collected information on the number and type of academic degrees earned by faculty to determine whether top-ranked programs are more likely to have faculty with additional academic training. Additional degrees held by PRS faculty included Doctor of Dental Surgery (DDS), Doctor of Dental Medicine (DMD), Doctor of Pharmacy (PharmD), Doctor of Philosophy (PhD), and master’s degrees. No statistically significant difference was found in the proportion of faculty with an additional academic degree between Top-quartile and Other programs (16% versus 14%, P = 0.6) (Fig. 2).
Fig. 2.

Faculty holding an additional academic degree. Top-quartile programs 16%, Other programs 14%, P = 0.6.
Fellowship Data
Since one of the roles of fellowship training is to enhance the desirability of individuals seeking a career in academia, information on fellowship training was collected to discern whether faculty at top-ranked programs are more likely to be fellowship-trained. Eighty percent (80%) of all faculty identified were fellowship-trained. There was no significant difference in proportion of fellowship-trained faculty when comparing Top-quartile and Other programs (84% versus 79%, P = 0.2) (Fig. 3).
Fig. 3.

Faculty with fellowship training. Top-quartile programs 84%, Other programs 79%, P = 0.2.
Top-quartile Residency Training
Given the knowledge that top-ranked residencies are associated with more academically productive residents, we aimed to determine whether top-ranked programs have a larger proportion of faculty that trained at top-ranked residency programs. For each group, we collected information on how many faculty trained at top-quartile PRS residencies, as determined by Doximity. Faculty who trained at residency programs other than PRS (ie, dentistry, general surgery, and otolaryngology) were excluded from analysis. We found that faculty from Top-quartile programs were significantly more likely to have trained at a top-quartile PRS residency than faculty from Other programs (69% versus 32%, P < 0 .0001) (Fig. 4A).
Fig. 4.

Faculty with training from a top residency program or medical school. A, Top-quartile residency: Top-quartile programs 69%, Other programs 32%, P < 0.0001. B, Top-quartile medical school: Top-quartile programs 52%, Other programs: 31%, P < 0.0001.
Top-quartile Medical School Training
We also aimed to determine whether top-ranked programs have a larger proportion of faculty that trained at top-ranked medical schools, as determined by U.S. News rankings. For each group, information on how many faculty attended top-quartile medical schools was collected. Faculty who attended other graduate schools (ie, Caribbean or international medical school, dental school) were excluded from statistical analysis. We found that faculty from Top-quartile programs were significantly more likely than faculty from Other programs to have trained at a top-quartile medical school (52% versus 31%, P < 0.0001) (Fig. 4B).
Faculty Alumni
We further aimed to determine whether top-ranked programs are more likely to exhibit “inbreeding” in their faculty selection, as evidenced by a higher proportion of faculty that are alumni of the institution of their program. As such, we determined the proportion of faculty that received medical school, residency, and/or fellowship training from the same institution with which they are currently affiliated. Compared to faculty from Other programs, faculty from Top-quartile programs were significantly more likely to have trained for fellowship and residency within the same institution (fellowship: 28% versus 10%, P = 0.0001; residency: 36% versus 24%, P = 0.03) (Fig. 5A, B). Faculty from Top-quartile programs were also more likely to have trained for medical school within the same institution, although this trend was not statistically significant (12% versus 9%, P = 0.4) (Fig. 5C). When looking at combined data, we found that faculty from Top-quartile programs were more likely to have received any form of medical training from the same institution (medical school, residency, or fellowship) compared to faculty from Other programs (54% versus 35%, P = 0.002) (Fig. 5D). A list of integrated PRS residency programs with the highest proportion of faculty alumni is displayed in Table 2.
Fig. 5.

Faculty with prior training from the same affiliated institution. A, same fellowship: Top-quartile programs 28%, Other programs 10%, P = 0.0001. B, same residency: Top-quartile programs 36%, Other programs 24%, P = 0.03. C, same medical school: Top-quartile programs 12%, Other programs 9%, P = 0.4. D, same institution (any training): Top-quartile programs 54%, Other programs 35%, P = 0.002.
Table 2.
Top 10 Residency Programs by Faculty Who Previously Trained at the Same Institution (Medical School, Residency, or Fellowship)
| Position | Program | N (%) |
|---|---|---|
| 1 | Brigham and Women’s Hospital/Harvard Medical School | 14 (93) |
| 2 | UCLA David Geffen School of Medicine/UCLA Medical Center | 12 (92) |
| 3 | New York Presbyterian Hospital (Columbia and Cornell Campus) | 11 (85) |
| 4 | New York University School of Medicine | 14 (78) |
| 5 | Penn State Milton S Hershey Medical Center | 7 (78) |
| 6 | University of Tennessee College of Medicine | 7 (78) |
| 7 | University of Michigan Health System | 12 (75) |
| 8 | University of Virginia Medical Center | 6 (75) |
| 9 | Nassau University Medical Center | 5 (71) |
| 10 | University of Pennsylvania Health System | 10 (71) |
Program Size
Since large programs generally have greater depth in faculty specialization and subspecialization, we asked whether faculty size was correlated with being a top program. Compared to Other programs, Top-quartile programs were indeed significantly larger as determined by total faculty size (13 versus 7 full-time faculty, P < 0.0001) (Fig. 6).
Fig. 6.

Number of full-time faculty. Top-quartile programs 13 faculty, Other programs 7 faculty, P < 0.0001.
Program Graduates in Academia
Finally, to assess whether top-ranked programs are more likely to graduate residents that will pursue a career in academia, we conducted an analysis of program graduates who become academic plastic surgeons after graduation. For each program, the total number of program graduates that are current faculty members in the United States was collected. To account for graduating class size, this was divided by the number of yearly program graduates (using data collected from program websites and NRMP reports) to calculate the number of program graduates per graduating resident that are full-time faculty members. Compared to Other programs, Top-quartile programs were more likely to graduate residents who will become academic plastic surgeons. This trend was significant for both total number of program graduates and for number of program graduates per graduating resident (total number: 15 versus 4 graduates, P < 0.0001; relative to graduating class size: 4 versus 2 graduates, P < 0.0001) (Fig. 7A, B). A list of integrated PRS residency programs with the largest number of graduates entering academia is displayed in Table 3.
Fig. 7.

Number of program graduates entering academic plastic surgery. A, faculty grads (total number): Top-quartile programs 15 graduates, Other programs 4 graduates, P < 0.0001. B, faculty grads (relative to graduating class size): Top-quartile programs 4 graduates, Other programs 2 graduates, P < 0.0001.
Table 3.
Top-10 Residency Programs by Output of Graduates Entering Academic Plastic Surgery
| Position | Program | N (adjusted N) |
|---|---|---|
| 1 | UPMC Medical Education | 35 (9) |
| 2 | New York Presbyterian Hospital (Columbia and Cornell Campus) | 33 (9) |
| 3 | New York University School of Medicine | 27 (8.3) |
| 4 | University of California (San Francisco) | 20 (6.7) |
| 5 | University of Chicago | 20 (6.5) |
| 6 | UCLA David Geffen School of Medicine/UCLA Medical Center | 20 (6) |
| 7 | Brigham and Women’s Hospital/Harvard Medical School | 18 (5.8) |
| 8 | University of Rochester | 18 (5.5) |
| 9 | University of Michigan Health System | 16 (5) |
| 10 | University of Pennsylvania Health System | 14 (5) |
Discussion
The notion of comparing PRS residencies by program rank is not new, as many studies have investigated differential outcomes for residents and faculty in “top” programs compared to all other programs. This has been studied in relation to the scholarly output of residents and faculty, the accessibility of faculty mentorship to younger trainees, and successful residency match outcomes.13–19 For instance, the direct relationship between integrated PRS program rank and research productivity of the program’s faculty has already been established and found to be significant.4 PRS program size and National Institutes of Health funding has also been positively correlated with the scholarly impact of faculty members as measured by the Hirsch index.5,6 In this study, we aimed to further identify concrete differences associated with top-ranked PRS programs and their faculty by analyzing publicly available databases of full-time integrated academic plastic surgeons.
By comparing programs by rank, we found a statistically significant tendency of faculty from top-ranked programs to have trained at top PRS residencies. Though these results do not establish a causal relationship, they emphasize the correlation between having an “elite” training background and obtaining faculty appointments at “elite” PRS programs. Interestingly, our analysis revealed a similar association between being a faculty at a top-ranked program and having attended a top-ranked medical school. For students interested in pursuing a career in academic plastic surgery, this finding suggests that even the reputation of the student’s medical school has the potential to affect their employment prospects. Whether this is because academically oriented students actively seek out top-ranked programs to further their academic agendas or because PRS programs favor applicants with more “elite” training remains to be determined. Nevertheless, it is clearly understood that there are many highly motivated and productive plastic surgeons throughout the country who are nationally acclaimed and respected regardless of the ranking of their training programs or institutions with which they are affiliated. From the applicant perspective, we know that working in research-oriented programs and having research-oriented mentors are the two most important factors associated with the choice to pursue academic medicine.20 From the program perspective, we know that at least for residency applications, PRS programs favor academically productive applicants who come highly recommended by other surgeons in the field.7 This same bias likely applies to the selection of new faculty, where research-oriented applicants are favored by more research-oriented programs.
Furthermore, to investigate alumni bias in faculty selection, we calculated the proportion of faculty who had previously trained at the same institution as their program, for medical school, residency, or fellowship. We found a significant association between program rank and faculty with alumni status, indicating a high rate of alumni retention in Top-quartile PRS residency programs. This finding emphasizes the tendency of programs from the upper echelons of academia to “inbreed” and select prospective academic plastic surgeons who previously trained from within their own ranks. Indeed, this association can be advantageous for all, allowing more academically competitive programs to preferentially hire faculty they are more familiar with and whom they know to have received a rigorous level of surgical training. Since faculty at top programs tend to have trained at the same or similarly ranked institution at some point in their career, it can be valuable for prospective faculty interested in joining the ranks of a top PRS program to consider training there first.
No significant difference in proportion of faculty by academic rank was found between Top-quartile and Other programs. This suggests that PRS programs promote faculty to higher academic ranks at a similar pace regardless of program rank. Alternatively, it also suggests that the research productivity of academic plastic surgeons is similar between differently ranked programs, although specific measures such as the Hirsch index or total number of publications might define academic productivity more accurately than academic rank alone. Similarly, we found no significant difference in the proportion of faculty with additional academic degrees between the two groups, thus suggesting that top-ranked programs do not necessarily prefer faculty who have undergone additional academic training. For instance, research doctorates may be becoming less of a requirement in academic surgery, as more aspiring academic surgeons chose to take dedicated research years during their medical school or residency training instead of pursuing full-length research doctorate degrees. Our results also reveal that 80% of academic plastic surgeons are fellowship-trained, with no significant difference in proportion of fellowship training between Top-quartile and Other programs. As previously mentioned, this could be due to applicant preference (ie, choosing to pursue fellowship training because of its ties to the world of academia) or program preference (ie, preferentially selecting applicants who have undergone more specialized training). Either way, it is useful for aspiring academic plastic surgeons to understand that the proportion of fellowship completion among integrated PRS faculty is very high.
Analysis of program size revealed that top-ranked programs are significantly larger and have more full-time faculty than their lower-ranked counterparts. Given this finding, we accounted for differences in program size by performing all analyses as “percent faculty” of Top-quartile versus Other programs. Finally, we aimed to address the theory that top-ranked PRS programs are more likely to graduate residents who will become future academic plastic surgeons. Indeed, we found that Top-quartile programs have a higher number of program graduates who pursue a career in academic plastic surgery. To discern whether this was confounded by the larger graduating class sizes of top-ranked programs, we analyzed the number of program graduates per graduating resident and still found a statistically significant difference between the two groups, suggesting that top-ranked programs are more likely to produce academic plastic surgeons than lower-ranked programs. This further lends credibility to the notion that highly reputed programs might preferentially attract as well as encourage trainees to be scholastically productive and pursue an academic career.
In this study, we chose to focus our analysis on surveying integrated plastic surgery training programs rather than independent programs. Although similar in their educational objectives and commitment to training autonomous and skilled plastic surgeons, the independent and integrated plastic surgery training programs differ in important ways. In the integrated training model, the Residency Review Committee for Plastic Surgery (RRC-PS) supervises the entire length of residency training. In the independent training model, the RRC-PS supervises only the requisite period of residency training, while the prerequisite period is supervised by another RRC (ie, general surgery, otolaryngology, urology, orthopedics, etc.). Because of this distinction, the focus of this study was on integrated plastic surgery programs and their faculty.
Another limitation of this study is that departmental websites are not always updated with the latest information, thereby rendering some of the data collected potentially out of date. The choice to analyze faculty rank as a surrogate for academic productivity also does not come without limitations, as individual faculty productivity is not uniform across divisions or departments. Furthermore, this study did not include data on percentage of faculty members who stayed in a specific program over time. This limits our ability to gauge academic retainment as a program metric and serves as a starting point for future studies.
CONCLUSIONS
In conclusion, this study utilizes objective data to reveal that integrated PRS program reputation is associated with increased selection and production of academic plastic surgeons. Top-ranked PRS programs are more likely to exhibit reputation bias and hire faculty that previously trained at top-ranked medical schools and residencies. Top-ranked programs are more likely to exhibit alumni bias and hire faculty who previously trained at the same institution. Top PRS programs are also more likely to be biased toward training residents who will enter the field of academic plastic surgery upon graduation. As the number of available faculty positions continues to grow along with the number of ACGME-accredited integrated PRS residencies, the results of this study can help aspiring academic plastic surgeons and current residency programs understand the factors involved in the selection and production of academic physicians and to ultimately attract their best match.
ACKNOWLEDGMENT
The authors would like to extend their gratitude to Warren L. Garner, MD, for his critical review of early manuscript drafts.
Footnotes
Published online 20 January 2020.
Disclosure: The authors have no financial interest to declare in relation to the content of this article.
REFERENCES
- 1.Chen JT, Girotto JA, Kitzmiller WJ, et al. Academic plastic surgery: faculty recruitment and retention. Plast Reconstr Surg. 2014;133:393e–404e. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Liu P, Singh M, Eriksson E. Academic status of plastic surgery in the united states and the relevance of independence. Handchir Mikrochir Plast Chir. 2016;48:65–68. [DOI] [PubMed] [Google Scholar]
- 3.Ruan QZ, Ricci JA, Silvestre J, et al. Academic productivity of faculty associated with microsurgery fellowships. Microsurgery. 2017;37:641–646. [DOI] [PubMed] [Google Scholar]
- 4.Zhang JQ, Herman SB, Tepper OM, et al. Rank and research: the correlation between integrated plastic surgery program reputation and academic productivity. Ann Plast Surg. 2018;80:553–560. [DOI] [PubMed] [Google Scholar]
- 5.Therattil PJ, Hoppe IC, Granick MS, et al. Application of the h-index in academic plastic surgery. Ann Plast Surg. 2016;76:545–549. [DOI] [PubMed] [Google Scholar]
- 6.Silvestre J, Abbatematteo JM, Chang B, et al. The impact of national institutes of health funding on scholarly productivity in academic plastic surgery. Plast Reconstr Surg. 2016;137:690–695. [DOI] [PubMed] [Google Scholar]
- 7.International NRMP Results of the 2018 NRMP Program Director Survey. 2018Washington, DC: National Residency Matching Program. [Google Scholar]
- 8.Smith BB, Long TR, Tooley AA, et al. Impact of Doximity residency navigator on graduate medical education recruitment. Mayo Clin Proc Innov Qual Outcomes. 2018;2:113–118. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Rolston AM, Hartley SE, Khandelwal S, et al. Effect of Doximity residency rankings on residency applicants’ program choices. West J Emerg Med. 2015;16:889–893. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Susarla SM, Lopez J, Swanson EW, et al. Are quantitative measures of academic productivity correlated with academic rank in plastic surgery? A national study. Plast Reconstr Surg. 2015;136:613–621. [DOI] [PubMed] [Google Scholar]
- 11.Gast KM, Kuzon WM, Jr, Waljee JF. Bibliometric indices and academic promotion within plastic surgery. Plast Reconstr Surg. 2014;134:838e–844e. [DOI] [PubMed] [Google Scholar]
- 12.Lopez J, Ameri A, Susarla SM, et al. Does formal research training lead to academic success in plastic surgery? A comprehensive analysis of U.S. Academic plastic surgeons. J Surg Educ. 2016;73:422–428. [DOI] [PubMed] [Google Scholar]
- 13.Tadisina KK, Orra S, Bassiri Gharb B, et al. Applying to integrated plastic surgery residency programs: trends in the past 5 years of the match. Plast Reconstr Surg. 2016;137:1344–1353. [DOI] [PubMed] [Google Scholar]
- 14.Borsting EA, Chim JH, Thaller SR. An updated view of the integrated plastic surgery match. Ann Plast Surg. 2015;75:556–559. [DOI] [PubMed] [Google Scholar]
- 15.Sue GR, Narayan D. Generation Y and the integrated plastic surgery residency match: a cross-sectional study of the 2011 match outcomes. Plast Reconstr Surg Glob Open. 2013;1:e33. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Rinard JR, Mahabir RC. Successfully matching into surgical specialties: an analysis of national resident matching program data. J Grad Med Educ. 2010;2:316–321. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Canadian Plastic Surgery Research C Barriers and attitudes to research among residents in plastic and reconstructive surgery: a National Multicenter Cross-Sectional Study. J Surg Educ. 2017;74: 1094–1104. [DOI] [PubMed] [Google Scholar]
- 18.Nagarkar PA, Janis JE. Eliminating geographic bias improves match results: an analysis of program preferences and their impact on rank lists and results. Plast Reconstr Surg. 2018;142:82e–88e. [DOI] [PubMed] [Google Scholar]
- 19.Silvestre J, So AL, Lee BT. Accessibility of academic plastic surgeons as mentors to medical students. Ann Plast Surg. 2015;74:85–88. [DOI] [PubMed] [Google Scholar]
- 20.Borges NJ, Navarro AM, Grover A, et al. How, when, and why do physicians choose careers in academic medicine? A literature review. Acad Med. 2010;85:680–686. [DOI] [PubMed] [Google Scholar]
