Significance
Scholarly productivity impacts nearly every aspect of a researcher’s career, from their initial placement as faculty to funding and tenure decisions. Historically, expectations for individuals rely on 60 years of research on aggregate trends, which suggest that productivity rises rapidly to an early-career peak and then gradually declines. Here we show, using comprehensive data on the publication and employment histories of an entire field of research, that the canonical narrative of “rapid rise, gradual decline” describes only about one-fifth of individual faculty, and the remaining four-fifths exhibit a rich diversity of productivity patterns. This suggests existing models and expectations for faculty productivity require revision, as they capture only one of many ways to have a successful career in science.
Keywords: data analysis, career trajectory, computer science, productivity, sociology
Abstract
A scientist may publish tens or hundreds of papers over a career, but these contributions are not evenly spaced in time. Sixty years of studies on career productivity patterns in a variety of fields suggest an intuitive and universal pattern: Productivity tends to rise rapidly to an early peak and then gradually declines. Here, we test the universality of this conventional narrative by analyzing the structures of individual faculty productivity time series, constructed from over 200,000 publications and matched with hiring data for 2,453 tenure-track faculty in all 205 PhD-granting computer science departments in the United States and Canada. Unlike prior studies, which considered only some faculty or some institutions, or lacked common career reference points, here we combine a large bibliographic dataset with comprehensive information on career transitions that covers an entire field of study. We show that the conventional narrative confidently describes only one-fifth of faculty, regardless of department prestige or researcher gender, and the remaining four-fifths of faculty exhibit a rich diversity of productivity patterns. To explain this diversity, we introduce a simple model of productivity trajectories and explore correlations between its parameters and researcher covariates, showing that departmental prestige predicts overall individual productivity and the timing of the transition from first- to last-author publications. These results demonstrate the unpredictability of productivity over time and open the door for new efforts to understand how environmental and individual factors shape scientific productivity.
Scholarly publications serve as the primary mode of communication through which scientific knowledge is developed, discussed, and disseminated. The amount that an individual researcher contributes to this dialogue—their scholarly productivity—thus serves as an important measure of the rate at which they contribute units of knowledge to the field, and this measure is known to influence the placement of graduates into faculty jobs (1), the likelihood of being granted tenure (2, 3), and the ability to secure funding for future research (4).
The trajectory of productivity over the course of a researcher’s lifetime has been studied for at least 60 years, with the common observation being that a researcher’s productivity rises rapidly to a peak and then slowly declines (5–9), which has inspired the construction of mechanistic models with a similar profile (7, 9–12). These models have included factors like cognitive decline with age, career age, finite supplies of human capital, and knowledge advantages conferred by recent education, as well as skill deficits among the young (among others), and have been supported by the observation that individual productivity curves feature both long- and medium-term fluctuations (12) and are not well described by even fourth-degree polynomial models (9). Indeed, every study we found to date proposes or confirms a “rise and decline,” “curvilinear,” or “peak and tapering” productivity trajectory, regardless of whether researchers are binned by chronological age (5–8, 10–13), career age (9, 10), or (only for young researchers) years since first publication (14). The pattern may even extend to mentorship, supported by a finding that the protégés of early-career mathematicians tended to mentor more students, themselves, than protégés trained by those same faculty late in their careers (15). In fact, this conventional narrative of the life course is not restricted to academia, with similar trajectories observed in criminal behavior and artistic production in 1800s France (16) and even in productivity of food acquisition by hunter-gatherers (17).
While these past studies have firmly established that the conventional academic productivity narrative is equally descriptive across fields and time, their analyses are based on averages over hundreds or thousands of individuals (5–11, 13–17). This raises two crucial and previously unanswered questions: Is this average trajectory representative of individual faculty? And how much diversity is hidden by a focus on a central tendency over a population? To answer these questions, we combine and study two comprehensive datasets that span 40 years of productivity for nearly every tenure-track professor in a North American PhD-granting computer science department. By introducing a simple mathematical description of the shape of a scientist’s productivity over time, we map individuals’ publication histories to a low-dimensional parameter space, revealing substantial diversity in the publication trends of individual faculty and showing that only a minority follow the conventional narrative of productivity. In fact, even among the conventional trajectories, individuals exhibit large fluctuations in their productivity around the average trend. Together, these results reveal that population averages provide a dramatically inaccurate picture of intellectual contributions over time and that productivity patterns are both more diverse and less predictable than previously thought. These findings were preliminarily described in a recent review (18) which provides additional context for the results reported fully here.
Moreover, while we show that the distribution of productivity trajectories resists natural categorization, it is nevertheless possible to explore covariates that are associated with different regions of its parameter space. The literature on such associations has avoided detailed trajectories and instead focused on the complicated relationship between prestige, productivity, and hiring. Past studies have found that researchers trained at prestigious institutions are likely to remain productive (19), regardless of where they place as faculty (20). Other results link the prestige of the doctorate and the advisor to early-career productivity but not long-term productivity (21), which is at odds with other studies (22, 23) that found early-career productivity predicts long-term productivity. Disagreement about hiring exists as well, with multiple studies finding that doctoral prestige and not productivity drives the initial placement of faculty (24, 25), while recent work based on comprehensive data in multiple fields suggests that prestige alone is insufficient to fully explain faculty placement (1, 26). This, too, is complicated by hypotheses of mutual causality, where departments both select for and facilitate high productivity (27). Unfortunately, while such studies shed light on a complicated system, they tend to restrict their analyses to unusual scientists, such as Nobel laureates or faculty at elite departments, rather than typical researchers. In contrast, the data analyzed here are comprehensive, covering faculty across the prestige hierarchy, which enables us to move beyond total productivity to study publication trajectories in light of prestige, hiring, and past productivity alike.
This study exploits and combines two large datasets related to faculty productivity. The first one is a comprehensive, hand-curated collection of education and academic appointment histories for tenure-track and tenured computer science faculty (26). This dataset spans all 205 departmental or school-level academic units on the Computing Research Association’s Forsythe List of PhD-granting departments in computing-related disciplines in the United States and Canada (archive.cra.org/reports/forsythe.html).
For each department, the dataset provides a complete list of regular faculty for the 2011–2012 academic year, and for each of the 5,032 faculty in this collection, it provides partial or complete information on their education and academic appointments, obtained from public online sources, mainly résumés and homepages. Of these, we selected the 2,583 faculty who both received their PhD from and held their first assistant professorship at one of these institutions and for whom the year of that hire is known and occurred in 1970–2011. The first requirement ensured that we modeled the relatively closed North American faculty market; roughly 87% of computing faculty received their PhD from one of the Forsythe institutions, and past analysis has shown that Canada and the United States are not distinct job markets in computer science (26). A number of faculty were removed in this step because the location of their first assistant professorship was not known; these were mainly senior faculty.
The first dataset also provides a ranking of institutional prestige , derived from patterns in the PhD-to-faculty hiring network between departments. In short, is a consensus of ordinal rankings (lower is better) in which prestige is defined recursively: Prestigious departments are those whose graduates are hired as faculty in prestigious departments. Networks, code, and rankings are available in ref. 26.
The second dataset, constructed around the first one, is a complete publication history as listed in the Digital Bibliography and Library Project (DBLP; dblp.uni-trier.de), an online database that provides open bibliographic information for most journals and conference proceedings relevant to computing research, using manual name disambiguation as necessary. For each paper in a faculty’s publication history, we recorded the paper’s title, author list (preserving author order), and year of publication. By following this procedure, we collected data for 200,476 publications which covered 2,453 (95.0%) faculty in our sample. Of those, we manually collected records of all peer-reviewed conference and journal publication histories from the publicly available curricula vitae (CVs) of 109 faculty, a randomly selected 10% of the 1,091 faculty with career lengths between 10 years and 25 years, providing a benchmark dataset to evaluate the accuracy of DBLP data (Collection of CV Data).
Our combined dataset consists of the career trajectories of these 2,453 tenure-track faculty as of 2011–2012: each professor’s publicly accessible metadata, their time-stamped PhD and employment history, and the annotated time series of their publications. We note that this dataset does not include information on faculty who have retired or left academia before 2011. Implications of these data limitations for the conclusions that can be drawn from our analyses are explored in Discussion. Finally, this study was not reviewed by an institutional review board because all data used were collected from publicly available sources. All results are presented anonymously or in aggregate to avoid revealing personally identifiable information about individual scientists.
Collection of CV Data
Performing the DBLP coverage analysis and adjustment discussed in General Trends in Productivity Data required a benchmark dataset with complete coverage of the publication histories of a representative subset of researchers. This section describes the relevant details of the collection of that benchmark dataset. We manually extracted lists of publication dates from publicly available CVs belonging to a random 10% of the researchers with career lengths between 10 years and 25 years and having publications in at least 3 distinct years. This last condition ensures that the piecewise linear model can be fitted to the individual’s trajectory and excludes just 32 researchers from our analysis. Because of the high diversity of productivity trajectories, we chose 10% of individuals, uniformly at random, from each of the quadrants designated by the signs of the two slope parameters, and , as shown in Fig. 4. Specifically, names of researchers from each quadrant were randomly shuffled and then collected, in order, until reaching 10% of the total. However, individuals for whom a CV could not be found or whose publicly available CV was last updated before 2011 were skipped, and other faculty were randomly selected in their place. Success rates for this exercise ranged between 66.3% and 86.6%, measured as the number of successfully extracted publication lists vs. the total number of attempts. The majority of these failures were due to researchers having out-of-date CVs. Future studies should consider whether such partial records of researcher productivity are sufficient for analysis, as their inclusion would greatly improve success rates during collection.
Results
General Trends in Productivity.
Two broad trends characterize scholarly productivity in academic computer science. First, publication rates have been increasing over the past 45 years, and second, higher publication rates are correlated with higher prestige. These two observations are intertwined and underpin a number of subsequent analyses, so we explore them briefly in more depth.
Past studies have found that researchers at more prestigious institutions tend to be more productive (20, 21, 24, 25, 27, 28). Our data corroborate this finding, but we also find that the typical productivity advantage associated with greater prestige holds regardless of whether an institution is public or private, for both early-career publications (first 10 years; Fig. 1) and lifetime publications (Fig. S4). Regressing the median number of time-adjusted publications (below) among faculty in a department against departmental prestige indicates that the relationships between prestige and productivity are statistically indistinguishable for public and private institutions, with expected increases in the first decade of a career of roughly 2.7 publications for every 10-rank improvement in prestige. In fact, when comparing public and private institutions, neither the prestige–productivity slope nor productivity overall is significantly different (, , respectively, two-tailed test), contradicting the conventional wisdom that private universities enjoy a productivity advantage over public ones. The conventional wisdom is likely skewed by a focus on elite departments, as 8 of the top 10 computer science departments are private (26), but in fact, private institutions are distributed evenly across all ranks. Expanding this analysis to include lifetime publications increases the prestige–publication slope to 3.28 publications per 10-rank improvement in prestige but does not alter the nonsignificance of public/private status (, , two-tailed test).
Past studies have also found that publication rates have increased over time (29, 30). However, before investigating whether changes in publication rates apply to computer science, we used the manually collected CV data to probe the extent of DBLP’s coverage. Indeed, the fraction of publications indexed by DBLP is nonuniform over time, increasing linearly from around 55% in the 1980s to over 85% by 2011 ( = 0.685, , two-tailed test; Fig. S1 and General Trends in Productivity Data). Because DBLP’s coverage of the published literature varies over time, in the analyses that follow we use data from hand-collected faculty CVs whenever possible and otherwise apply a statistical correction to DBLP’s data to account for its lower coverage.
Knowing already that there are substantial differences in productivity by prestige, we separated universities by prestige into five groups of approximately equal size and investigated whether the growth of publication rate varies by prestige. We find that the average number of publications per person produced in each calendar year has been increasing at all five strata of prestige at rates between 0.72 and 1.23 publications per decade, for 45 years (Fig. S3). Because we have used data from hand-collected faculty CVs to adjust DBLP-derived paper counts for DBLP’s steadily improving coverage over time, these estimated growth rates represent a real increase in publication rates over this 40-year period. Moreover, the observed steady increase in productivity is not uniform across prestige, and the difference between production growth rates between higher- and lower-prestige departments has widened slightly but significantly over time (, two-tailed test). In other words, prestigious and nonprestigious institutions have contributed to the overall growth at different rates. Not only are there small but significant differences in productivity by prestige (Fig. 1) but also those differences are slowly growing (Fig. S3).
To investigate the productivity patterns of individual researchers and test the conventional narrative of rapidly rising productivity followed by a gradual decline, for the remainder of this paper we focus on time series of individual productivity. However, due to both the observed growth in productivity and the variability in DBLP coverage, it would be misleading to directly compare a 1975 publication with a 2011 publication. Thus, hereafter we use “adjusted” publication counts, which corrects the raw DBLP counts to account for both the changing DBLP coverage and the increasing mean publication rate over time (General Trends in Productivity Data). All publication counts are hence reported as 2011-equivalent counts.
Individual Productivity Trajectories.
Examining the productivity trajectories of individual researchers, we find that they too exhibit substantial and significant differences in their publication rates. Early studies of scholarly productivity noted profound imbalance in the number of articles published by individual researchers (31, 32). Cole (33) and Reskin (34) in the 1970s noted that about 50% of all scholarly articles were produced by about 15% of the scientific workforce. Our data reflect similar levels of imbalance, with approximately half of all contributions in the dataset authored by only 20% of all faculty. Stratifying by decade, however, the Gini coefficients for productivity imbalance have been declining, from 0.62 in the 1970s to 0.40 in the 2000s (Fig. S5). This trend persists when researchers are restricted to only the publications within the first 5 years of their careers.
There are several possible explanations for the trend of decreasing inequality in individual productivity. For instance, the lower end of the productivity distribution could have become relatively more productive over time, perhaps as more institutions shifted focus from teaching to research. Or, it may reflect a strengthening selective filter on highly productive faculty, perhaps as community expectations for continual productivity rose. It may also reflect nonuniform errors in the DBLP data, although the correction for DBLP coverage should account for these (General Trends in Productivity Data).
We now focus on testing the conventional productivity narrative that has been described in various disciplines and at many points in time (5–11, 13, 14, 16, 17): Productivity climbs to a peak and then gradually declines over the course of the researcher’s career. Across computer science faculty, we find that the average number of publications per year over a faculty career is highly stereotyped (Fig. 2), with a rising productivity that peaks after around 5 years, declines slowly for another 5 years, and then remains roughly constant for any remaining years. Although departmental prestige correlates with productivity in several ways (Fig. 1 and Fig. S3), it does not alter this stereotypical pattern, which appears essentially unchanged across departments with different levels of prestige, except for a roughly constant shift up as prestige increases (Fig. 2).
The suggestion that productivity grows in the early years of a career has intuitive appeal. Professors settle into their research environments, begin training graduate students, and build their cases for promotion and tenure. Similarly, many reasons have been suggested for why productivity might decrease after promotion, including increased service and nonresearch commitments, declining cognitive abilities, and increased levels of distraction from outside work due to health issues and childcare obligations (35). Although an average over faculty appears to confirm the stereotyped trajectory of rapid growth, peak, and slow decline, it does not reveal whether this average is representative of the many individual trajectories it averages over, nor does it show how much diversity there might be around the average and whether that diversity correlates with other factors of interest.
To characterize the productivity pattern within an individual career, we fit a simple stereotypical model of productivity over time to the number of papers published per year,
[1] |
a piecewise linear function in which is the change point between the two lines, and are the rates of change in productivity before and after the change point, respectively, and is the initial productivity (Fig. 3). We apply this model to the faculty who have been used for 10–25 years. By fitting these four parameters to each individual’s publication trajectory, we map that trajectory into a low-dimensional description of its overall pattern [fitting done by least squares; see Least-Squares Fit of f(t) for optimal numerical methods and Modeling Framework for detailed discussion of statistical models].
However, before interpreting the distributions of parameters, we subjected each trajectory to two additional tests to ensure that its best-fit parameters were meaningful. First, to avoid overfitting linear trajectories with a piecewise linear model, we performed model selection, asking whether the Akaike information criterion (AIC) with finite-size correction favored a straight line or the more complex (Model Selection). This process conservatively selected only 33.3% () of researchers who are more confidently modeled by the piecewise function.
Second, to address the possibility that a researcher’s best-fit parameters may be sensitive to small changes in the years of their publications, we conducted a sensitivity analysis in which we repeatedly refit model parameters to productivity trajectories, adding a small amount of noise to shift some publications into adjacent years (Sensitivity to Timing of Publications). This procedure places each professor’s noise-free trajectory within a distribution of nearby noisy trajectories, enabling two different (but ultimately concordant) analyses. The primary sensitivity analysis focuses on individual faculty, computing whether the parameters of each professor’s noise-free trajectory are similar to their noisy distribution. This approach revealed that a majority (77.2%) of trajectories are well represented by their noise-free parameters, each consistently falling into the same region of parameter space for over 75% of resampled trajectories. We refer to these trajectories as “stable” in subsequent analyses, meaning that their noise-free parameters are representative and interpretable. The alternative sensitivity analysis focuses on the population of faculty, combining all noise-free trajectories with their noise-added distributions into a single expanded ensemble of conceivable productivity trajectories (Sensitivity to Timing of Publications). Although this ensemble is unable to support analyses of individual faculty, we use it to corroborate the findings that follow. Combining the individual stability and AICs, we find that 32.3% () of researchers possess productivity trajectories that are both stable and nonlinear. All analyses and discussions of model parameters hereafter refer to stable, nonlinear trajectories unless otherwise noted.
The narrative of “early growth in productivity, followed by a slow decline” implies four conditions on the inferred parameters: While the conditions of growth () and decline () are straightforward, we interpret “early growth” to mean that inferred peak productivity comes within the first decade after hiring () and “slow” to mean that the slope of decline is smaller in magnitude that the slope of growth . After fitting individual trajectory models to the 1,091 faculty in our sample, we find that only 20.1% follow the stereotypical trajectory. Even dropping the aforementioned restriction on increases the fraction meeting the stereotype to only 20.3%. To ensure that these results were not sensitive to our definition of stability in the presence of noise, we generated an ensemble with 200 noise-added trajectories for each professor (Fig. S8 and Sensitivity to Timing of Publications), subjected each to the AIC for nonlinearity, and found that only 19.7% of ensemble trajectories are reliably categorized as adhering to the conventional narrative. In other words, the average trajectory, which has been held up as established fact for more than 50 years, describes the behavior of only a minority of researchers, while a large majority of researchers follow qualitatively different trajectories.
Publication trajectories can be divided into four general classes based on the signs of the two slope parameters, and , corresponding to the quadrants shown in Fig. 4. Individual trajectory shapes exhibit substantial diversity, spanning all four quadrants. Even among faculty whose publication rates grew and then declined (Fig. 4, Bottom Right quadrant, 28.6%), the conventional narrative includes only the 20.3% of individuals whose rate of growth exceeds their rate of decline (; shaded region, Fig. 4). Additionally, researchers were distributed similarly across the four quadrants, comparing parameters extracted from DBLP data vs. hand-collected CV data (, ), confirming that the dispersion shown in Fig. 4 represents the true diversity of careers.
The cloud of faculty trajectory parameters shown in Fig. 4 does not naturally separate into coherent clusters. In their absence, what are the covariates that predict which region of the plot an individual is likely to occupy? First, early-career growth rate of yearly publications is significantly correlated (, test) with the prestige of researchers’ institutions. This is particularly true for researchers at “elite” institutions, which we define as being in the top 20% of universities according to prestige rank and adjusting for number of faculty (same partitions as in Fig. 2). Specifically, researchers’ productivity grows by a median of 2.02 additional papers per year at elite institutions compared with 1.19 for others (, one-tailed Mann–Whitney test). Perhaps as a result—what goes up must come down—the slope after the point of change, , correlates significantly with prestige and is more negative for researchers at higher-ranked institutions, compared with those at lower-ranked institutions (, test). Additionally, researchers who received their doctorates from elite institutions exhibit faster early-career growth than those who trained at lower-ranked institutions (, one-tailed Mann–Whitney test).
Second, the early-career initial productivity is significantly higher for faculty who graduated from elite departments (, one-tailed Mann–Whitney test). We also find that researchers who place into elite departments or who have postdoctoral experience tend to start out more productive; however, these differences are not statistically significant (, Mann–Whitney test). These findings regarding and combine to suggest that current academic environment correlates with—and perhaps influences—productivity, while prior academic environment does not. Finally, faculty at top-ranked departments are statistically no more or less likely to be found within this triangular region, a result robust to alternative cutoffs for “top-ranked” institutions.
The relationship between trajectories and gender is more complicated. First, trajectories of male and female researchers were similarly distributed across the four quadrants (, test), and gender was uncorrelated with the likelihood of meeting the four criteria of the canonical narrative (, test). Further, within this canonical subset, the women’s initial productivity grew at a rate indistinguishable from the men’s (, Mann–Whitney test) and peaked in similar years (, Kolmogorov–Smirnov test). Women’s initial productivity, however, was 0.46 publications lower than the men’s (, Mann–Whitney test) in general, and this difference exists despite the fact that men and women in this subset trained and were hired at similarly ranked institutions (, Kolmogorov–Smirnov test) and completed postdoctoral training at similar rates (, test).
The change point within a career may indicate regime shifts in productivity, regardless of which type of trajectory an individual may follow. While the change-point parameter does not correlate with the other parameters of , its distribution reveals that for most faculty, the inferred change point in productivity rates occurs at approximately year 5. Fig. 5 translates each selected faculty member’s career length and inferred change point into an ordered pair, creating a heat map of career change points. Shown in the accompanying marginal distribution, the modal value for is year 5 with the median at 6 years, closely preceding tenure decisions at most institutions. Nevertheless, there is still rich diversity in career transitions, and the average remains misleading as the descriptor of a majority of individuals. In particular, faculty at the top 20% of institutions have significantly earlier than the remaining 80%, with medians of 4.1 years and 6.4 years, respectively (, Mann–Whitney test). There is no such difference between the faculty whose doctorates are from the top 20% of institutions and those whose doctorates are from the remaining 80% (medians of 5.9 years vs. 6.0 years; ).
The trends and diversity observed in distributions remain true even when models are avoided entirely. A direct empirical examination of all DBLP and CV publication time series reveals that a computer science professor’s productivity is also most likely to peak in the fifth year, yet peak productivity can nevertheless occur in any year of a professor’s career (Fig. 6). While the marginal distribution shows that 41.9% of faculty have their peak productivity within the first 6 years, with the modal peak year in year 5, there is substantial variance. Note, for example, that individuals along the bottom of Fig. 6 published the most in their first year as faculty, while individuals along the diagonal published the most in their most recent recorded year as faculty.
Transitions in Authorship Roles.
Finally, other transitions exist that are not quantifiable in publication counts alone, yet these are surprisingly well synchronized with the transitions noted above. As faculty train graduate students, their roles ordinarily shift from lead researcher to senior advisor or principal investigator, and this transition is commonly reflected in a shift from first author to last author. While common, this first/last convention is not universal. For example, papers in theoretical computer science typically order authors alphabetically, so the relative position of these researchers in the author list will not exhibit any consistent pattern over a career. To investigate career-stage transitions in author position, we first identified the set of journals or conferences that list authors alphabetically by computing whether each venue’s authors are alphabetized significantly more often than is expected by chance () and exceeding twice the expected rate (Detection of Alphabetized Publication Venues). These conditions selected 11.2% of publication venues, accounting for 15.4% of all papers in the dataset, which we manually verified includes all top theoretical computer science conferences and excludes all top machine-learning and data-mining conferences. We then discarded these alphabetically biased venues from the following analysis. The remaining data show clear evidence of a progressive shift toward last-authorship position over time, with the relative first/last proportion reaching stability around year 8 (Fig. 7). Interestingly, the onset of this change is earlier among faculty at high-prestige institutions, and their average proportion of last-author papers is significantly higher than those of other faculty, consistent with a hypothesis that faculty at elite institutions tend to begin working with students earlier and have larger or more productive research groups.
As with the aggregate trend in productivity over a faculty career (Fig. 2), the transition from first- to last-author publications (Fig. 7) is based on averaging across many faculty and thus may not reflect the pattern of any particular individual. To characterize individual performances, we compared the fraction of first-authored papers in the first 3 years posthire to the same fraction in the second 3 years, for faculty with careers longer than 6 years (). A substantial drop in this fraction across these two periods would be consistent with the average trend reflecting individual patterns. For this analysis, we treated single-author papers as first-author publications. Overall 70.1% of researchers undergo this transition, publishing a larger fraction of first-author publications in the first 3 years of their faculty career than inthe second 3 years. These fractions are consistent for faculty at top-50 institutions (70.2%) and those at other institutions (70.1%), but individuals at top-ranked institutions appear to make the transition more quickly and completely by the end of the 6-year period (Fig. 8). Despite these trends, there remains substantial diversity among first/last author transitions, reinforcing the notion that averages may be poor descriptors of many individuals.
General Trends in Productivity Data
Meaningful trends in publication rates over a career can be confidently identified from raw publication counts only if two conditions are met. First, raw publication counts must be exhaustive, containing all peer-reviewed publications. Second, field-wide publication rates must be stationary over time. Due to the facts that DBLP data do not satisfy the former, and that computer science as a field does not satisfy the latter, raw publication counts recorded in the DBLP dataset must be adjusted to compensate before they can be analyzed. This section explains the details of two compensatory adjustments to DBLP data. We justify the first adjustment by providing a detailed analysis of the time-varying fraction of publications covered by the DBLP dataset, anchored by a hand-collected benchmark CV dataset (Collection of CV Data). We then justify the second adjustment by identifying a clear and significant overall growth in publication rates over 40 years of computer science publication data. We conclude by discussing several possible explanations for why researcher productivity increases over time.
DBLP has indexed the overwhelming majority of current computer science publications, including peer-reviewed conferences and journals. However, while excellent today, this coverage has increased systematically over time, meaning that DBLP coverage is less complete for older publications and faculty. To quantify trends in the time-varying coverage of our DBLP data, we hand collected the CVs of 109 faculty, representing 10% of individuals whose trajectories are shown in Fig. 4, providing a set of benchmark publication lists (Collection of CV Data).
For each year of our dataset, we compared the number of publications listed on individuals’ CVs to the number of publications listed on their corresponding DBLP profiles, selecting only peer-reviewed conference and journal publications from CVs. Comparing DBLP data with CV benchmarks, two sets of counts reveal that DBLP coverage has increased linearly from around 55% in the 1980s to over 85% in 2011 (Fig. S1). DBLP coverage has grown at a rate of ∼1.06% of additional coverage per year, with a 95% confidence interval indicating that this rate falls between 0.8% and 1.4%. Because the ratio of DBLP publications to CV publications is well described by the line
[S1] |
we use Eq. S1 to convert all nonbenchmarked DBLP publication counts to CV-equivalent publication counts, with estimated parameters of and .
After linearly adjusting all raw publication counts to correct for the expected DBLP coverage in a given year (Eq. S1), we analyzed how individual researcher productivity has changed over the years spanned by our dataset. Due to the fact that we extracted and adjusted DBLP records for only authors in the faculty hiring dataset (1, 26), a straightforward analysis of the number of per-person publications in each calendar year would feature a different mixture of career ages in each year. For example, adjusted publication counts from the 1970s would include only early-career researchers; late-career researchers in the 1970s retired long before our dataset was collected. Because the main text of this paper reveals systematic trends in productivity by career age, this straightforward counting technique would introduce bias.
To quantify the expansion of publication rates over time, without introducing career-age bias, we selected “indicator” career ages at which to measure productivity and compared how productivity at specific points in the academic career has changed over time. The trend in publication growth is consistently positive for all indicator career ages, and, as in the main text, publication rates after year 5 tend to be higher than publication rates in the first 4 years. When these publication rates were normalized by their 2011 values, all indicator sets collapsed onto a common growth line (Fig. S2). Thus, while the indicator sets of , , and revealed that productivity grows at rate between 0.84 and 1.48 additional papers per person per decade, their rates of growth are directly proportional to 2011 productivity, meaning that the shape of the canonical trajectory has not changed over time, but has simply expanded proportionally.
The relative slopes of publication rate expansion are consistent with the relative publication rates in the canonical “average” productivity trajectory (Fig. 2). Indeed, growth in researchers’ first 5 years of productivity is relatively modest, which is expected since researchers, both historically and more recently, spend these years building their research programs by applying for funding and recruiting graduate students and postdoctoral researchers. On the other hand, productivity in year 5—the year of or immediately preceding tenure evaluations at most institutions and, perhaps not coincidentally, the modal year of peak researcher productivity—grows at a faster rate of 1.48 additional papers per person per decade. These rates are both similar comparing years 5, 10, and 15, which describes changes across a larger window of career productivity.
The relationship between 2011-equivalent publications and past publications is consistently linear over the time spanned by our dataset (Fig. S2) and is modeled well by
[S2] |
We therefore used Eq. S2 to convert all CV-equivalent publication counts to 2011-equivalent publication counts, with estimated parameters and . Thus, we applied the two transformations of Eq. S1 (Fig. S1) and Eq. S2 (Fig. S2) in series to publication data to produce the adjusted publication counts used in the main text, unless otherwise specified; benchmark CV data were used for the 109 individuals for whom we collected them, and for those individuals, only Eq. S2 was applied. Note that while model-based correction can adjust for trends in publication rates, alternative model-free approaches have been used by other researchers, which convert publication rates to annual publication rate scores, e.g., ref. 36.
In the main text, we noted that, after applying these two linear adjustments, the median number of early-career publications per person per institution increases over time and correlates with prestige. Fig. S3 illustrates this trend, stratifying individuals into five levels of prestige and revealing that production growth rates between higher- and lower-prestige departments have widened slightly but significantly over time (, two-tailed test).
This trend, observed for early-career publications (i.e., publications within the first 10 years of a career, for individuals with careers of 10 years or longer), is no different from the trend for all posthire publications and all researchers (Fig. S4). Median lifetime career productivity correlates significantly with prestige, and, as in early-career productivity, public and private institutions are similarly affected by this relationship. Using an ordinary least-squares regression of productivity vs. prestige including dummy and interaction terms for public/private status, we found that the relationship between productivity and prestige is not significantly affected by public/private status (, test, for both public/private dummy and public/private–prestige interaction).
While production rates have increased steadily over time and for all levels of prestige, we find that the imbalance in research production has decreased in recent years. Fig. S5 illustrates this inequality across researchers and time by showing the fraction of all publications in our sample that were produced by the most productive fraction of all faculty (a Lorenz curve), for faculty first hired in each of the four decades that our data span and restricting analysis to only publications produced in the first 5 years of an individual’s career. As referenced in the main text, the Gini coefficients for productivity imbalance have declined, from 0.62 in the 1970s to 0.40 in the 2000s. Many factors could potentially drive such a shift toward more balanced research production in science (including, for example, technological advancements and corresponding declines in research costs that may have leveled the playing field for researchers at less-prestigious universities), and we welcome future studies that explore this shift in more detail.
Finally, several possibilities exist that might explain why researchers are becoming more productive. First, the average number of coauthors per publication has steadily increased over time, allowing researchers to work on a larger number of projects. Second, the number of publication venues has also grown, providing more outlets for researchers’ work and potentially facilitating more specialized communities with faster peer review. Third, technological advances including improvements in computer architecture have benefitted researchers universally, increasing the speed at which results are both generated and published. Finally, perhaps the perception of what constitutes the minimum publishable unit of research has changed over time, resulting in a larger number of shorter, more narrowly focused publications in recent years.
Modeling Framework
Eq. 1 is the simple model used in the main text to parameterize the adjusted publication counts over the course of a career. Reproduced below, it consists of two lines with slopes and that intersect at time .
[S3] |
In this article, we fit Eq. S3 to adjusted count data, using least squares. However, there exist other regression frameworks that correspond to generative models for time series data. In this section, we discuss some of these alternatives.
Adjusting Models Instead of Counts.
Although publication time series are naturally count data, a regression framework that is naturally suited for counts, such as Poisson or negative-binomial regression, is not advised. Directly fitting raw counts using a Poisson or negative-binomial model would neglect the adjustments for both the coverage of DBLP and the time-varying changes in publication rates (General Trends in Productivity Data). On the other hand, adjusted publications are nonintegers, rendering them inappropriate for count regressions. However, there are alternatives that would allow for both count regressions and the adjustments in General Trends in Productivity Data. These adjustments come with a price, however, due to the assumptions and free parameters that they introduce.
One alternative solution to fitting a model to adjusted publication data is to fit an adjusted linear model to raw publication data. In other words, adjust the model instead of the data. Due to the fact that this approach would preserve the data as counts, this model would be amenable to Poisson and negative-binomial regression frameworks. Adjusted publications are related to raw publications by the adjustment
[S4] |
where and are slopes and and are intercepts of the linear adjustments for DBLP coverage and publication expansion, respectively, and hats indicate that variables have been estimated from data (General Trends in Productivity Data). Applying this adjustment to the model , which is to hold for adjusted publications, we get
[S5] |
where is the initial year of a particular faculty member’s career, and is the calendar year.
We now turn to details of Poisson and negative-binomial frameworks for fitting the adjusted model Eq. S5 and discuss the assumptions and parameters that they introduce.
Poisson Model.
Consider a Poisson fit of Eq. S5 to a set of data given by . To simplify, let us be explicit about the dependence of on the four parameters, , , , and , which we collectively refer to as .
where we have made clear that does not depend on the model parameters . The likelihood is then
[S6] |
Rather than maximizing , we maximize . Taking the natural log of both sides, we get
[S7] |
Note that the terms and do not depend on the parameters , so they affect the value of the maximum but not its location in parameter space. Dropping them yields a log-likelihood score of
[S8] |
Note that for any trajectory, can be precomputed and does not depend on the parameters . Thus, fitting the 2011-equivalent Poisson model requires that we maximize Eq. S8 with respect to , , , and . This equation must be maximized numerically.
While this adjusted Poisson model is attractive because it naturally fits count data, it imposes assumptions on the data-generating process that are not justified empirically. Namely, the variance and mean of a Poisson distribution are equal, meaning that the Poisson regression expects the same of the data it explains.
Negative-Binomial Model.
The Poisson model above enforces the constraint that the mean is equal to the variance. However, there is no indication that the data support this assumption so we introduce the standard alternative, the negative-binomial model. This model requires both a mean and a heterogeneity parameter , such that the probability of a single observation is
[S9] |
As in the Poisson regression, we once more parameterize the mean, using the piecewise linear model as . However, we must also introduce a model for .
The easiest way forward, mathematically, is to set . This assumes equal heterogeneity around the expected value for all time points in a career . Note that this assumption decouples the heterogeneity from the mean, while under the Poisson model they are directly coupled. One might think of the Poisson model therefore as fitting the parameters to both the trend and the fluctuations together. The fixed- negative-binomial model, on the other hand, fits the parameters to the trend and uses a fixed to accommodate all fluctuations. In this sense this negative-binomial approach is more flexible and uses an additional parameter to gain that flexibility.
The assumption results in a log probability of
[S10] |
and we note that and that both and are constants that do not depend on either or , allowing us to write a log-likelihood score of
[S11] |
Progress here, however, is obstructed by the difficulties of taking derivatives of Gamma functions. Thus, Eq. S11 must be optimized numerically over the parameters and .
While these calculations may be helpful in seeding a way forward in future work, it is important to note that the fixed- negative-binomial model also makes strong assumptions about the generative process that created the data. Indeed, the assumption that fluctuations are uniform over an entire career is strong and is not justified by data. One could also avoid this assumption, but this introduces additional problems, which we now discuss.
The temptation to let each point have a parameterized value of and a free parameter of results in overfitting. Note that this would allow each point in the time series to be fitted by a negative-binomial distribution with a mean given by Eq. S5 and an arbitrarily large or small , resulting in dramatic overfitting. This approach therefore makes few assumptions, but provides little value to the modeler.
A middle ground between fixed and unrestricted would be to parameterize , using a lower-dimensional model. Using the same model for as we used for would be similar, in principle, to the Poisson regression. Using a different model for is an option, but would require, again, a deep focus on the underlying mechanisms hypothesized to explain fluctuations in productivity.
Modeling Outlook.
Generative models for productivity trajectories would be enormously valuable. In this section, we emphasize the assumptions made by the generative models underlying various regression frameworks. In particular, we derive models that are able to be fitted directly to raw count data by including the inverse of the adjustment derived in General Trends in Productivity Data, a quadratic term referred to as .
In terms of impacts, fitting the Poisson model and the fixed- negative-binomial model to the trajectories investigated in this paper does affect the parameters of individuals’ trajectories. However, it does not diminish the diversity of trajectories that we observe. Indeed, the example trajectories shown in Fig. 4 are only subtly affected by the use of one type of generative model or another.
Sensitivity to Timing of Publications
Publication generally signifies the conclusion of a research project but the exact date when an article is published can depend on many factors, including the availability of reviewers, graduation deadlines for graduate students, delays between acceptance and publishing, and synchronization with conference submission deadlines, as well as nonacademic constraints, such as the impending birth of a child. Each of these factors might advance or delay a publication’s appearance in the literature, and furthermore, the effort associated with each publication may span weeks, months, or years. As a result, publication years serve as a noisy indicator of when productivity occurs.
To ensure that our findings are not due to coincidence in the timings of researchers’ publications, we examined the sensitivity of our results to the addition of small amounts of noise. For each researcher and for each of their publications, we added noise drawn from a normal distribution (, ) to the publication year and then rounded to the nearest whole year. In expectation, this process leaves one-half of publication years unaffected and shifts 22.8% by 1 year in either direction, 2.1%by 2 years, and 0.04% by 3 years. We repeated this process 200 times for each researcher (examples shown in Fig. S6) and found that the median number of trials in which the trajectory changed shape compared with the noise-free fit (i.e., or changed sign) was 9 (4.5%; Fig. S7).
While the typical individual’s parameters are robust to noise, those individuals whose trajectories featured few publications or whose noise-free model parameters were near zero were far more likely to change shape. In fact, for 10.5% of individuals, noise led to shape change more often than not, i.e., in greater than 50% of noisy repetitions. Therefore, as an additional check, we asked whether the model parameters inferred for researchers’ noise-free trajectories differ significantly from those inferred for their 200 noise-added trajectories. We used Fisher’s method to combine values for each researcher and found no significant differences in any of the four model parameters (, , , and ). Evaluated separately, fewer than 1% of researchers’ inferred model parameters differ significantly under the noise-free and noise-added fits. We conclude that the general shape of productivity trajectories is robust to small differences in publication year.
Having specified a model for noise around each publication date, an individual’s trajectory naturally becomes a distribution of trajectories, which in turn maps into a corresponding distribution in the parameter space. To analyze this distribution, we inferred model parameters for each of a researcher’s 200 noise-added trajectories, and for those trajectories that were more confidently modeled by Eq. 1 (using the AIC with finite-size correction; Model Selection) vs. a straight line, we compiled the distribution depicted in Fig. S8.
This distribution suggests a complementary approach to investigating the universality of the conventional narrative wherein individuals are considered as distributions rather than point estimates. The latter of these approaches (presented in the main text) requires specifying a threshold for stability that determines whether or not an individual can be confidently mapped into a particular region in the space. The former, on the other hand, requires no such distinction and instead sums the independent distributions of individual faculty to quantify the total probability mass in each region. The point estimate approach makes a statement about individuals, while the distribution approach makes a statement about the population. Importantly, both statements agree: ∼20% of individuals are firmly mapped to the canonical octant, and ∼20% of probability mass maps into the canonical octant.
Model Selection
When the complexity of a model exceeds the complexity of the underlying data, some parameters of the model may no longer be interpreted as meaningful. Although the piecewise linear model of Eq. 1 has only four parameters—two slopes and , an intercept , and a change point —it may nevertheless overfit productivity trajectories that are actually linear. This section provides additional details for our model selection procedures that avoid the overinterpretation of the piecewise change-point parameters .
If a publication trajectory is generated by a straight line with added noise, then fitting a piecewise model will result in approximately equal to , and the location of the change-point will be arbitrary. We apply model selection to identify individuals with career lengths between 10 years and 25 years () whose productivity trajectories are both stable (Sensitivity to Timing of Publications) and consistently better modeled by the piecewise model Eq. 1 than by a straight line [i.e., ordinary least squares (OLS)]. This filter includes only those trajectories whose change-points can be interpreted with confidence.
To perform model selection, we consider three information theoretic model selection techniques: the AIC, the Akaike information criterion with finite-size correction (AICc), and the Bayesian information criterion (BIC), defined as
where is length of the individual’s career, is the number of model parameters ( for OLS, and for Eq. 1), and is the sum of squared errors (differences) between actual and modeled publication counts. The AIC is the least conservative of the three methods, finding that 59.1% () of individuals are better modeled by Eq. 1 than by OLS regression. By contrast, AICc and BIC select only 33.2% () and 44.4% () of individuals, respectively. In the main text, we adopt the most conservative approach, AICc, but the other two methods nevertheless produce qualitatively similar distributions for , with the modal year for remaining at year 5.
Detection of Alphabetized Publication Venues
Conventions of author order vary widely in computer science. In the first/last convention, first authorship is reserved for the lead author or primary contributor to the study, while last authorship indicates the senior author who oversaw or advised the work. In the alphabetical convention, borrowed from mathematics, a paper’s authors are arranged alphabetically by last name. If the trend from first authorship toward last authorship over a career is to be reliably interpreted (Fig. 8), publications with alphabetical author orders must be discarded. This section explains the methods used to statistically identify and remove publication venues that are highly enriched with alphabetical conventions.
Our approach is to count the number of multiauthor papers in each publication venue with alphabetically ordered authors and compare this count to the number expected by chance. (We note that all single-author papers are ignored in this analysis.) A paper with authors will list its authors alphabetically by chance with probability . Noting the number of authors of each multiauthor paper published by a particular venue and assuming independence of ordering decisions, we derive an empirical distribution representing the number of coincidentally alphabetized author lists and ask whether the venue adopts the convention significantly more often than would be expected by chance. Additionally, we require that the number of observed alphabetized lists be at least twice the expected value. These two conditions ensure that the alphabetical convention is both significant and widespread in its adoption in a particular venue. We find that 630 of the 5,622 (11.2%) distinct venues in our dataset alphabetize their author lists. These venues account for 27,237 of the 177,437 (15.4%) multiauthor conference or journal publications for which the publication venue is known.
Manually inspecting the list of alphabetized venues reveals that popular theoretical venues like the Symposium on Theory of Computing (STOC), Foundations of Computer Science (FOCS), the Symposium on Theoretical Aspects of Computer Science (STACS), and the Symposium on Discrete Algorithms (SODA) adhere to the alphabetical convention, while the World-Wide Web Conference (WWW), the Conference on Computer-Supported Cooperative Work and Social Computing (CSCW), the Conference on Knowledge Discovery and Data Mining (KDD), the Conference on Human Factors in Computing Systems (CHI), and the AAAI Conference on Artificial Intelligence (AAAI) do not, matching our expectations.
Least-Squares Fit of f(t)
A least-squares fit of the continuous piecewise-linear equation given in Eq. 1 involves minimizing the sum of squared errors over four parameters, ,, , and . In this section, we provide details that make this fitting process rapid and maximally accurate.
The model fit consists of two steps. First, we assume that is fixed and find the optimal values of , , and . In the second step, we search for the whose corresponding optimal parameters provide the best fit. The sum of squared error is given by
[S12] |
where is the change point, denotes the sum for all , and denotes the sum for all .
In the first step, we imagine to be fixed and simply take partial derivatives with respect to the three parameters, set each equal to zero, and solve. Setting yields three equations,
[S13] |
Thus, for a fixed value of , the above equations provide a linear system of three equations for the three unknowns , , and . The system can be solved numerically, using any linear solver.
In the second step, we embed the optimization above within a search for the optimal value. For each proposed value of , we use Eq. S13 to find the optimal , , and and then compute the associated error using Eq. S12, choosing the that minimizes . Initially, we propose a coarse grid at the level of and then refine the grid by an order of magnitude locally around the best result, repeatedly, until the optimal is known to single precision. This procedure is fast, and due to the result of Eq. S13, limits numerical search to a one-dimensional line.
Discussion
The conventional narrative of faculty productivity over a career is pervasive, with repeated findings reinforcing a canonical trajectory where productivity rises rapidly to a peak early in one’s career and then declines slowly (5–11, 13, 14, 16, 17). This narrative shapes expectations of faculty across career stages, and publication counts have been shown to impact both tenure decisions (2, 3) and the ability to secure funding for future research (4). In this study, we showed that the conventional narrative, while intuitive and certainly applicable to averages of many professors, is a remarkably inaccurate description of most professors’ trajectories. By applying a simple piecewise-linear model to a comprehensive dataset of academic appointment histories and publication records, we found that only about one-fifth of tenured or tenure-track computer science faculty resemble the average, regardless of their department’s prestige.
While diverse, some aspects of a trajectory are nevertheless partially predictable. For example, although the diversity of trajectories remains unaffected, productivity does tend to scale with prestige: Researchers who graduated from or were hired by top-ranked institutions are significantly more productive at the onset of their careers, and, furthermore, productivity of high-prestige faculty tends to grow at faster rates and achieve higher peaks than that of researchers used by other institutions. Together, these results support previously suggested hypotheses that top-ranked universities both select for and facilitate productivity (27). In fact, our results suggest that the early-career transition to leadership roles, a phenomenon also found in other disciplines (36), takes place more quickly at top-ranked institutions, further implicating facilitation effects in addition to selection.
The relationship between productivity trajectories and gender is complicated and requires careful study. Gender has been shown to correlate with differences in productivity across fields (37–39), but these relationships are complicated by prestige (26) and have also changed over time (1). Other work has uncovered differences in collaboration patterns between subfields (40), as well as productivity differences that depend on both student and advisor genders (41). Here, we found that men and women follow the canonical productivity narrative at equal rates. However, among those who do, we found significant differences in initial and peak productivities between men and women. Given the complications revealed in past studies, the extent to which these differences reflect inequalities, past or present, and contribute to women’s underrepresentation in computer science is an important topic of research and warrants future exploration.
Within the space of career trajectories, there is a noticeable tendency toward peak productivities and shifts in publication rates around 5 years after beginning as faculty. This is surely not a coincidence, given the fundamental role of tenure as a change point within the typical academic career, after which the total number of hours worked does not substantially change, but the time devoted to service tends to dramatically increase, with concomitant decreases in research and grant writing (42). However, our data cannot yet say how, from a mechanistic perspective, the existence of tenure requirements drives faculty to change or shape their productivity before or after promotion. If anything, the results in this paper make clear that there are numerous ways in which computer scientists meet promotion requirements, not all of which necessarily involve publishing a large number of papers. Indeed, in parallel with career shapes more broadly, there remains broad diversity in the distributions of productivity peaks and change points. This diversity in overall production, combined with the observation that an individual’s highest-impact work is equally likely to be any of his or her publications (43), implies there are fundamental limits to predicting scientific careers (18).
Computer science is, itself, a multifaceted field, and previous studies of the DBLP dataset revealed that productivity rates differ by subfield (1). This observation, coupled with the menagerie of fluctuating trajectories revealed here, may suggest that year-to-year differences in individual trajectories are related to which subfields a researcher studies. Past work has revealed a first-mover advantage associated with entry into a rapidly growing field (44), so changes to individual research interests may contribute to noisy trajectories, particularly if they coincide with concentrated growth of popular new subfields.
Larger and higher-resolution datasets may improve our ability to identify expanding new subfields and other factors that could explain or predict trajectories. Although DBLP has the advantage of covering computer science journals and peer-reviewed conferences alike, we found that its coverage of those venues was incomplete in predictable ways. By manually collecting CV data for 10% of the scattered trajectories shown in Fig. 4, we adjusted DBLP data for missing publications and established the rate at which publishing rates have grown since 1970. Trajectories derived from DBLP data and benchmark CV data were statistically indistinguishable from each other. Investigations of productivity trajectories outside computer science will lack the field-specific DBLP database and may require additional calibration, name disambiguation, and data deduplication.
The misleading narrative of the canonical productivity trajectory is not likely to be unique to computer science. The rich diversity revealed here demands a reevaluation of the conventional narrative of careers across academia. Other studies that investigate the impact of this pervasive narrative on decisions of promotion, retention, and funding would be particularly valuable. Expectations, whether perceived or enforced through tenure decisions, might give rise to some of our results. If these expectations vary from field to field, it is possible that while diversity remains a feature that spans academia, some types of trajectories may be more common in certain fields. Regardless of whether this is borne out by studies of other fields, models of faculty productivity will need to be revisited and revised.
Acknowledgments
The authors thank Mirta Galesic and Johan Ugander for helpful conversations. All authors were supported by National Science Foundation Award SMA 1633747; D.B.L. was also supported by the Santa Fe Institute Omidyar Fellowship.
Footnotes
The authors declare no conflict of interest.
This article is a PNAS Direct Submission.
This article contains supporting information online at www.pnas.org/lookup/suppl/doi:10.1073/pnas.1702121114/-/DCSupplemental.
References
- 1.Way SF, Larremore DB, Clauset A. Gender, productivity, and prestige in computer science faculty hiring networks. Proc 25th Intl Conf World Wide Web. 2016:1169–1179. [Google Scholar]
- 2.Cole S, Cole JR. Scientific output and recognition: A study in the operation of the reward system in science. Am Socio Rev. 1967;32:377–390. [PubMed] [Google Scholar]
- 3.Massy WF, Wilger AK. Improving productivity: What faculty think about it—and its effect on quality. Change. 1995;27:10–20. [Google Scholar]
- 4.Stephan PE. The economics of science. J Econ Lit. 1996;34:1199–1235. [Google Scholar]
- 5.Lehman HC. Men’s creative production rate at different ages and in different countries. Sci Mon. 1954;78:321–326. [Google Scholar]
- 6.Dennis W. Age and productivity among scientists. Science. 1956;123:724–725. doi: 10.1126/science.123.3200.724. [DOI] [PubMed] [Google Scholar]
- 7.Diamond AM. The life-cycle research productivity of mathematicians and scientists. J Gerontol. 1986;41:520–525. doi: 10.1093/geronj/41.4.520. [DOI] [PubMed] [Google Scholar]
- 8.Horner KL, Rushton JP, Vernon PA. Relation between aging and research productivity of academic psychologists. Psychol Aging. 1986;1:319–324. doi: 10.1037//0882-7974.1.4.319. [DOI] [PubMed] [Google Scholar]
- 9.Bayer AE, Dutton JE. Career age and research-professional activities of academic scientists: Tests of alternative nonlinear models and some implications for higher education faculty policies. J Higher Educ. 1977;48:259–282. [Google Scholar]
- 10.Simonton DK. Creative productivity: A predictive and explanatory model of career trajectories and landmarks. Psychol Rev. 1997;104:66–89. [Google Scholar]
- 11.Levin SG, Stephan PE. Research productivity over the life cycle: Evidence for academic scientists. Am Econ Rev. 1991;81:114–132. [Google Scholar]
- 12.Rinaldi S, Cordone R, Casagrandi R. Instabilities in creative professions: A minimal model. Nonlinear Dyn Psychol Life Sci. 2000;4:255–273. [Google Scholar]
- 13.Cole S. Age and scientific performance. Am J Sociol. 1979;84:958–977. [Google Scholar]
- 14.Symonds MR, Gemmell NJ, Braisher TL, Gorringe KL, Elgar MA. Gender differences in publication output: Towards an unbiased metric of research performance. PLoS One. 2006;1:e127. doi: 10.1371/journal.pone.0000127. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Malmgren R, Ottino J, Nunes AL. The role of mentorship in protégé performance. Nature. 2010;465:622–626. doi: 10.1038/nature09040. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Quetelet LAJ. 1835. Sur l’Homme et le Développement de ses Facultés: Ou, Essai de Physique Sociale; trans Knox R (2013) [A Treatise on Man and the Development of his Faculties], ed Simbert T (Cambridge Univ Press, New York)
- 17.Kaplan H, Hill K, Lancaster J, Hurtado AM. A theory of human life history evolution: Diet, intelligence, and longevity. Evol Anthropol. 2000;9:156–185. [Google Scholar]
- 18.Clauset A, Larremore DB, Sinatra R. Data-driven predictions in the science of science. Science. 2017;355:477–480. doi: 10.1126/science.aal4217. [DOI] [PubMed] [Google Scholar]
- 19.Caplow T, McGee RJ. The Academic Marketplace. Basic Books, Inc.; New York: 1958. [Google Scholar]
- 20.Crane D. Scientists at major and minor universities: A study of productivity and recognition. Am Sociol Rev. 1965;30:699–714. [PubMed] [Google Scholar]
- 21.Reskin BF. Academic sponsorship and scientists’ careers. Sociol Educ. 1979;52:129–146. [Google Scholar]
- 22.Long JS, Allison PD, McGinnis R. Entrance into the academic career. Am Sociol Rev. 1979;44:816–830. [Google Scholar]
- 23.Chubin DE, Porter AL, Boeckmann ME. Career patterns of scientists: A case for complementary data. Am Sociol Rev. 1981;46:488–496. [Google Scholar]
- 24.Zuckerman HA. Stratification in American science. Sociol Inq. 1970;40:235–257. [Google Scholar]
- 25.Long JS. Productivity and academic position in the scientific career. Am Sociol Rev. 1978;43:889–908. [Google Scholar]
- 26.Clauset A, Arbesman S, Larremore DB. Systematic inequality and hierarchy in faculty hiring networks. Sci Adv. 2015;1:e1400005. doi: 10.1126/sciadv.1400005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Allison PD, Long JS. Departmental effects on scientific productivity. Am Sociol Rev. 1990;55:469–478. [Google Scholar]
- 28.Pelz DC, Andrews FM. Scientists in Organizations: Productive Climates for Research and Development. Institute for Social Research; Ann Arbor, MI: 1976. [Google Scholar]
- 29.Dey EL, Milem JF, Berger JB. Changing patterns of publication productivity: Accumulative advantage or institutional isomorphism? Sociol Educ. 1997;70:308–323. [Google Scholar]
- 30.Larsen PO, Von Ins M. The rate of growth in scientific publication and the decline in coverage provided by science citation index. Scientometrics. 2010;84:575–603. doi: 10.1007/s11192-010-0202-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Lotka AJ. The frequency distribution of scientific productivity. J Wash Acad Sci. 1926;16:321–322. [Google Scholar]
- 32.Shockley W. On the statistics of individual variations of productivity in research laboratories. Proc IRE. 1957;45:279–290. [Google Scholar]
- 33.Cole JR. Fair Science: Women in the Scientific Community. Free Press; New York: 1979. [Google Scholar]
- 34.Reskin BF. Scientific productivity, sex, and location in the institution of science. Am J Sociol. 1978;83:1235–1243. doi: 10.1086/226681. [DOI] [PubMed] [Google Scholar]
- 35.Fox MF. Publication productivity among scientists: A critical review. Soc Stud Sci. 1983;13:285–305. [Google Scholar]
- 36.Duch J, et al. The possible role of resource requirements and academic career-choice risk on gender differences in publication rate and impact. PLoS One. 2012;7:e51332. doi: 10.1371/journal.pone.0051332. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Long JS. Measures of sex differences in scientific productivity. Soc Forces. 1992;71:159–178. [Google Scholar]
- 38.Xie Y, Shauman KA. Sex differences in research productivity: New evidence about an old puzzle. Am Sociol Rev. 1998;63:847–870. [Google Scholar]
- 39.Fox MF. Gender, family characteristics, and publication productivity among scientists. Soc Stud Sci. 2005;35:131–150. [Google Scholar]
- 40.Zeng XHT, et al. Differences in collaboration patterns across discipline, career stage, and gender. PLoS Biol. 2016;14:e1002573. doi: 10.1371/journal.pbio.1002573. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Pezzoni M, Mairesse J, Stephan P, Lane J. Gender and the publication output of graduate students: A case study. PLoS One. 2016;11:e0145146. doi: 10.1371/journal.pone.0145146. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Link AN, Swann CA, Bozeman B. A time allocation study of university faculty. Econ Educ Rev. 2008;27:363–374. [Google Scholar]
- 43.Sinatra R, Wang D, Deville P, Song C, Barabási AL. Quantifying the evolution of individual scientific impact. Sci. 2016;354:aaf5239. doi: 10.1126/science.aaf5239. [DOI] [PubMed] [Google Scholar]
- 44.Newman MEJ. The first-mover advantage in scientific publication. Europhys Lett. 2009;86:68001. [Google Scholar]