Skip to main content
PLOS One logoLink to PLOS One
. 2022 Dec 29;17(12):e0278389. doi: 10.1371/journal.pone.0278389

Chaining and the temporal dynamics of scientists’ publishing behaviour

Emmy Liu 1,*, Yang Xu 2
Editor: Qichun Zhang3
PMCID: PMC9799287  PMID: 36580455

Abstract

Scientific progress, or scientific change, has been an important topic in the philosophy and history of science. Previous work has developed quantitative approaches to characterize the progression of science in different fields, but how individual scientists make progress through their careers is not well understood at a comprehensive scale. We characterize the regularity in the temporal dynamics of scientists’ publishing behavior with computational algorithms that predict the historical emerging order of publications from individual scientists. We discover that scientists publish in ways following the processes of chaining that mirror those observed in historical word meaning extension, whereby novel ideas emerge by connecting to existing ideas that are proximal in semantic space. We report findings for predominant exemplar-based chaining that reconstructs the emerging order in the publications of 1,164 award-winning and random-sampled scientists from the fields of Physics, Chemistry, Medicine, Economics, and Computer Science over the past century. Our work provides large-scale evidence that scientists across different fields tend to share similar publishing behavior over time by taking incremental steps that build on their past research outputs.

Introduction

The process of scientific inquiry can follow a winding path that varies across individuals. Isaac Newton explored optics in his earlier career but later studied the laws of motion and gravity [1]. Alan Turing at his earlier age worked on computability theory but later contributed to machine intelligence and the chemical basis of morphogenesis [2]. The problem of scientific change, or the progression of science through generations of scientists, has garnered much attention in the philosophy and history of science. However, whether there is regularity in how individual scientists make progress through time has not been studied comprehensively across scientific fields.

Scientific change has been an important topic in the philosophy of science. Kuhn suggests that science develops in cyclic periods of normal science and revolutionary science [3]. Normal science depends on similarities to exemplars, or canonical instances of scientific work that scientists base their hypotheses on. However, anomalies challenging the existing paradigm may build up and give rise to a period of revolutionary science, also known as a paradigm shift. Kuhn’s view resonates with Popper’s work emphasizing falsifiability as an avenue for the advancement of science [4], and Fayerabend’s incommensurability thesis arguing that successive scientific paradigms may have entirely different conceptual structures and language uses that are incompatible with each other [5]. Other scholars have suggested that paradigm shifts can happen gradually during periods of normal science, and the distinctions between normal science and revolutionary science may be blurry [6]. The philosophical inquiries of scientific change have focused on explaining scientific progress at a macro scale, but they tend to under-explore mechanisms in the scientific trajectories of individual scientists.

Parallel studies of scientific change have been conducted in the history of science and cognitive science. For instance, existing work has investigated the transition from Newtonian classical mechanics to the theory of relativity and quantum mechanics [7], and the effects of new instruments and methods in physics and chemistry [8]. Transition from one conceptual system to another involves an individual scientist noticing contradictions and inconsistencies within the old conceptual system, and constructing new representations of the scientific concepts. This line of work in cognitive science focuses on individuals, but it typically relies on anecdotal case studies.

Recent work has taken more scalable and quantitative approaches to studying scientific progress, but tends to focus on the evolution at the level of a scientific field. Existing work has explored scientific topic shifts, networks of scientific collaboration, as well as the connections among people, topics, methods, and team scale in scientific research [916]. Recent work has also explored the role of atypicality and outliers in scientific publications, such as how “sleeping beauty” papers that became highly influential over the long term at the expense of short-term citations, or understanding the factors underlying “outlier patents” [17, 18]. However, focusing on outliers and exceptional papers might obscure the incremental steps by which such papers are generated, since exceptional or unusual papers from a community perspective could result from the logical continuation of a line of research from an individual perspective. The problem we pursue here is whether simple computational algorithms focused on characterizing fine-grained temporal dynamics of scientists’ publication behavior might reveal general patterns in scientific progress made by individual scientists in different fields.

Although scientific inquiry over time can vary greatly across individuals, we hypothesize that there should be regularity in scientists’ publication behavior through time that reflect common principles of human cognition. Our approach is to uncover this regularity by testing cognitively inspired algorithms against the historical records of scientists’ research outputs as they emerged over time in different fields.

Our proposal is grounded in the cognitive theory of chaining [1922], and we use the existing computational cognitive algorithms that have been shown to capture properties of chaining in the context of word meaning extension [2225]. More specifically, chaining refers to a continuous and incremental process of meaning growth whereby novel ideas connect to existing ideas that are semantically similar, hence forming chain-like structures over time. Existing work has formulated chaining as probabilistic algorithms or models of categorization and applied these formal models to predict the historical growth of word meaning in a variety of domains including container names, numeral classifiers, adjectives, and verbs [2227].

We believe that the general process of scientific progress might mirror the process of word meaning extension, whereby a scientist conceives of a new idea by linking to existing ideas that are closely related. Importantly, previous work has also suggested that certain forms of chaining are preferred or more cognitively natural than other forms, and recent findings suggest that chaining can be best understood as an exemplar-based process [22, 24, 25]. The exemplar model of chaining is rooted in the exemplar theory of categorization, also known as the General Context Model [28]. This model has been demonstrated to predict the historical meaning extensions of numeral classifiers [22], adjectives [24], and verbs [25].

Here we investigate whether similar chaining mechanisms might account for the historical emergence of scientific publications from individual scientists. Independent work in mathematical modeling has proposed that similar processes may underlie the emergence of novelties [29]. However, to our knowledge there has been no formalization or comprehensive evaluation of chaining in the temporal dynamics of scientists’ publishing behavior over time. We focus on investigating three related issues: 1) whether there is systematic predictability in the historical order of emergence of scientists’ papers; 2) how the exemplar-based chaining account fares against alternative competing accounts including the prominent prototype-based view that predicts growth to stem from a central core [30, 31]; 3) whether mechanisms of chaining are domain-general or differ across scientific fields, or between award-winning (or prominent) and randomly-sampled (control) scientists.

To evaluate our proposal at scale, we collected a large set of historical publication data from scientists who vary along two dimensions: 1) scientists in different scientific fields, including Physics (n = 168), Chemistry (n = 120), Medicine (n = 151), Economics (n = 74), and Computer Science (n = 69); 2) award-winning scientists (n = 582) who have made prominent contributions in these fields, including Nobel Laureates and Turing Prize winners, versus randomly sampled scientists (n = 582 in total). Within each field and its corresponding prize category (e.g., Physics and Nobel Prize for Physics), we extracted data exhaustively from all prize-winning scientists since the 20th century, and we constructed the random control group by sampling scientists from the same field and matching the group size to that within the prize category. For each scientist, we collected his or her publications throughout history where the records were available, and we timestamped each publication by the year that the work was published. We describe details of the dataset and its processing in Materials and Methods.

In our analysis, we applied chaining models to the following prediction task: Given a scientist’s publications up to time t, predict which of the yet-to-emerge publications will more likely to appear at the next time step t+ 1. This prediction task is similar to the problem formulated in previous work that examines regularity in the historical emergence of word meanings [23]. Here instead we use this predictive analysis to explore shared patterns that underlie the emergence of scientific publications. If the exemplar-based chaining reflects a general and preferred mechanism in the emergence of scientific papers, we expect this model to best account for data across the 5 scientific fields, and in both the groups of prominent and random-sampled scientists.

Computational framework

We formulate the computational problem by considering the set of papers St a scientist has published by time t, and then applying models of chaining incrementally over time to infer the probability of a yet-to-emerge paper x* from the same scientist actually being published at the next available time t + 1. Our analysis treats each year as an individual timestep, so papers published within the same year are considered to emerge at the same time. We represent the publications of a scientist as a point cloud, where each point denotes a publication. As time unfolds, we ask the question whether there is predictability in how this cloud of points emerges over time. We expect papers that emerge at each next point to be generally related to the previous papers that have emerged, and we formalize the temporal dynamics in the emergence of scientific publications using a suite of chaining models grounded in the literature.

Each model of chaining predicts which papers are most likely to emerge at the next timestep based on a softmax function of the similarities of all papers which are yet to emerge, following Luce’s choice model [32]:

p(x*|St)exp(f(x*,St))ΣxSt*exp(f(x,St)), (1)

Here x* is drawn from the set of papers yet to emerge (or be published) beyond t which we denote as St*, and f() specifies a particular model of chaining. As a model predicts through time, St is updated with the ground truth papers at each timestep in order to minimize error propagation. Whenever “similarity” is mentioned below, we refer to a standard metric used in measuring psychological distance [28, 33, 34], the exponentiated squared Euclidean distance, between paper embeddings x1 and x2 in a semantic space.

sim(x1,x2)=exp(-d(x1,x2)2) (2)

We capture the semantic gist of a scientific paper by its content in the abstract section. We choose to represent a paper by the abstract due to its concise and informative character following a procedure adopted commonly in recent computational studies of scientific publications [35, 36]. Specifically, we represent each paper by a sentence embedding of the abstract generated by a model in natural language processing called SBERT [37], that we used the pretrained version of. This sentence embedding represents the semantic content of each abstract as a point in a high-dimensional vector space. Fig 1 provides a 2-dimensional visualization for the semantic space that includes paper embeddings of two prominent scientists in the fields of computer science and chemistry. data in S1 Appendix includes analyses that validate the measure we used to represent scientific papers. More specifically, we validate that the SBERT embeddings are able to capture semantic differences among abstracts published in different scientific fields, as well as among the abstracts written by individual scientists within fields.

Fig 1. Visualization of scientific publications over semantic space and time.

Fig 1

(A) Semantic representation of Turing Prize Winner Geoffrey Hinton’s published papers throughout his career, with annotations on sampled papers. Each dot represents a paper using BERT-based embedding of its abstract that is projected into 2 dimensions via tSNE dimension reduction [38]. Color gradient indicates year of publication, with more recent papers in warmer colors. Distance between abstract embeddings signifies inverse semantic similarity. (B) Similar information for Nobel Prize Laureate John B. Goodenough. Note that the legend shows the gradation of colours, the latest papers from our dataset were from 2021.

We consider five models of chaining following the existing literature [2224]. Each model postulates a plausible yet different cognitive mechanism for how future scientific outputs might relate to or stem from existing ones and constructs a continuous temporal path among the scientific publications of a scientist. Each chaining process is specified by a likelihood function listed in Table 1.

Table 1. Summary of model specifications.

Each model specifies the likelihood of paper x* to appear at time t, given the existing set of papers St. d denotes Euclidean distance. Ck(x) ⊆ St denotes the k closest points to x.

Model class Likelihood f(x*|St)
kNN ΣxCk(x)exp(-d(x,x)2)
Prototype exp(−sd(x*, prototype(St))2)
Progenitor exp(−sd(x*, prototype(S0))2)
Exemplar ΣxStexp(-sd(x*,x)2)
Local ΣxSt\St-1exp(-sd(x*,x)2)

k-Nearest-neighbor (kNN) models are commonly used for classification and regression, but in this case we use them to predict which paper from St* is likely to emerge next based on that paper’s similarity to the k most similar papers that have already appeared in St. This model captures the intuition that scientists preferentially publish work that is most similar to any one paper they have published before (e.g., 1NN), or most similar to a group of k papers they have published before (e.g., kNN). We tested values of k from 1 to 20 for each scientist.

The prototype model is motivated by Rosch’s work in categorization [30]. This model captures the intuition that a scientist’s research is organized around a central theme, yet scientists tend to publish in relation to that theme by gradually adjusting it and branching out over time. We take the prototype of a scientist’s papers at St to be the average of paper embeddings at St, and define the probability of a paper emerging at time t+ 1 as proportional to its similarity to the prototype at time t. This results in a moving average against which new papers are compared.

The progenitor model is a static variant of the prototype model in which the prototype is fixed at the first paper to emerge. This model captures the intuition that a scientist started with a base theme and then progressed consistently in that direction by incrementally making progress toward the theme. In the case where multiple papers emerge in the first year of publication, we take the progenitor to be the average of these papers’ embeddings. The probability of a paper emerging at any timestep is then proportional to its similarity only to the progenitor.

The exemplar model is based on work on the exemplar theory [28, 39], which can be viewed as a generalization or soft extension of the kNN models where k is equal to the number of total papers at each timestep but each paper is weighted differentially toward prediction. This model captures the intuition that a scientist tends to publish the next paper that is close to many similar papers they have written before, and the influence of each existing paper is proportional to its similarity to the target paper (i.e., a more similar paper exerts a higher influence). In this model, the probability of a paper emerging at time t is proportional to its similarity to all other papers St. We consider the most prominent exemplar model known as the Generalized Context Model [28].

The local model is a variant of the 1NN model wherein candidate papers are only compared against papers that emerged specifically at the previous timestep. This model captures the intuition that scientists branch out directly from the previous papers they wrote at the previous time step and publish papers close to what they last wrote in terms of time.

Table 1 summarizes the specification of the likelihood function f() for each model class. Together with Eq 1, they specify the probability that each model assigns to a paper emerging at time t, given all other papers that have appeared up to that time. Each model class has a single parameter (i.e., model complexity is kept the same for all the chaining models), k in the k-nearest neighbour models and s in the other models. We used sequential model-based optimization implemented in the Python hyperopt library to estimate each parameter s [40], and grid searched for k from 1 to 20. The parameter s controls how sharply similarity tapers with increased (Euclidean) distance from the query point.

Materials and methods

Data and materials availability

Data and code for replicating the analyses and results reported are deposited as follows: https://github.com/nightingal3/metascience/tree/master.

Collection and processing of scientific publication data

To capture a diverse and representative range of scientific outputs, we collected the publication data for both award-winning and randomly sampled scientists from five different fields: Physics, Chemistry, Medicine, Economics, Medicine, and Computer Science. We chose the first 4 fields because each of these fields corresponds to an award category of Nobel Prizes, which is one of the most prestigious and long-standing awards for science. We chose computer science because it is a relatively recent scientific field that is distinguished from those of Nobel Prizes, but it nevertheless features the prestigious Turing Award.

We collected all papers available for scientific scholars from DBLP for the Computer Science field, and from Semantic Scholar for the other fields [41, 42]. For each paper, we collected its title, abstract, year of publication, and coauthor list. We excluded papers from the main analysis if they were missing an abstract or year of publication, and from the authorship analysis if they were missing an abstract, year of publication, or authorship list.

We defined prominent scientists as those who received the highest honour in that scientific field from the inception of the award through to 2020. This was either the Turing Award for the field of Computer Science, or a Nobel Prize for the other fields. For the random group, we randomly sampled scientists from the Semantic Scholar open research corpus labelled with the respective field of study, excluding the set of prominent scientists. We matched the sample size of the random scientists to that of the random scientists. Table 2 summarizes the number of prominent scientists and the range of number of papers published for each field.

Table 2. Sample size of scientists per award category and range of number of publications.

CS Chemistry Economics Medicine Physics
# scientists 69 120 74 151 168
# pub. range 6–373 5–227 5–187 6–217 5–587

Semantic representation of abstracts

We computationally represented the semantic gist of a scientific paper by embedding its abstract via SBERT [37], which is a common method in natural language processing for representing sentences as a high-dimensional vector. We chose to focus on analyzing abstracts because they provide a compact distillation of the paper contributions while also containing some details on the approach and the background that the paper entails. This procedure follows work on summarization of scientific documents, which often uses the abstract as a target for text summarization [35, 36, 4345].

Before constructing the abstract embeddings, we used standard preprocessing in natural language processing while also removing textual features common in scientific abstracts that provide irrelevant information. For the standard preprocessing steps, we removed punctuation and stopwords, lemmatized, and removed bracketed words, as these were most often citations. For abstract-specific features, we removed symbols including LATEX, the prefixes “Abstract” and “Summary”, as well as subheadings within the abstract summarizing the parts of the paper (e.g., “Background”, “Results”). We also excluded any abstracts that were not written in English.

We ran these procedures for each scientist, excluding in turn any scientists that had fewer than 5 papers. For the prominent scientists, to remove unlikely papers, we excluded any papers published after the scientist’s death date or fewer than 15 years after their birth date. We converted the dates of publication to timestamps, so that the first paper published would be given timestamp 0, and the second timestamp 1, regardless of their actual publication dates.

We describe the details of a series of tests that evaluate the validity of our semantic representation for the scientific papers in Supporting Information.

Evaluation of chaining models against historical publication data

For each scientist, we calculated the predictive probability of a model based on the sequence of papers produced by that scientist, where the sequence recapitulates the historical emerging order recorded in the years of publications of the papers. The predictive probability of a sequence of papers is the product of the predictive probabilities of the individual papers as they emerged through time: P = ∏tp(x*|St), where x* denotes the newly emerged paper at the next time step and St denotes the set of papers published up to current time t. The log-likelihood of the sequence of papers is then simply log P. To compare each model against the null model, or a random guesser, we take the log-likelihood-ratio logPP(null) which is a standard statistical measure for model comparison. P(null) is theoretically determinable, because at each time t the probability of randomly guessing the true paper to emerge from a pool of n yet-to-emerge papers is 1n.

Results

We evaluated the chaining models described on the historical publication data against a null baseline model that makes random guesses about the emerging order of the publications. For each scientist, we calculated the predictive likelihood of a model for all the yet-to-emerge papers based on Eq 1 aggregated over time (see Materials and methods for details on model evaluation). Fig 2A and 2B show the raw log-likelihood ratios between each model and the null model for the prominent and random-sampled scientists respectively across the 5 fields. Model comparisons against the null indicate the likelihoods obtained in each of the chaining models are significantly higher than those of the random guesser (Mann-Whitney U test: p < 0.001 for all conditions; see Table in S1 Table for detailed summary statistics). We also examined the relative strength of the models in a winner-take-all analysis, where within each scientist we assessed which model best predicts the emergent pattern of papers over time. We expect the chaining models to perform better than the baseline, and the best model to yield the highest likelihood across scientists. Fig 3A and 3B summarize the results for the groups of prominent and random-sample scientists respectively. Table in S2 Table provides the exact model percentages from this analysis. In all cases, the chaining models dominated the null baseline model, and the exemplar-based chaining model yielded the highest predictive likelihood in scientists across fields (binomial test: p < 0.001; n = 10). Importantly, the same patterns of exemplar advantage were attested in both the group of prominent scientists and the group of random scientists. This initial set of results supports the view that there is temporal continuity in the emergence of scientific ideas with strong statistical tendencies in a shared mechanism of exemplar-based chaining observed robustly across the individuals and scientific fields.

Fig 2. Summary of model predictive performance.

Fig 2

(A) Log-likelihood ratio of model against the null random baseline model based on predicting historical data from the prominent scientists across the 5 fields. 0 on the vertical axis indicates equal performance between a chaining model and the null model; values above 0 indicate superior performance over the null model. (B) Similar information for the random-sampled scientists.

Fig 3. Summary of results from the winner-take-all analysis.

Fig 3

(A) Bar chart indicates the winning percentage of a model in predicting the historical emergence of publications from the prominent (or award-winning) scientists across the 5 fields (“CS” stands for computer science). Different groups of bar charts display winning percentages under different authorship conditions. (B) Similar information for the random-sampled scientists.

To examine whether collaboration might influence the results described, we next performed the same analysis by controlling explicitly for authorship conditions in the publications. A scientist may play a leading or non-leading role in a co-authored publication, and hence collaboration might introduce randomness in the historical emergence of scientific ideas for any given scientist. To assess whether our results are robust to this factor, we repeated likelihood calculation and winner-take-all analysis on each scientist under 4 different authorship conditions: fewer than 3 authors per paper, fewer than 2 authors per paper, single-author papers, and first-author papers. Here, first-authorship means that the scientist in question is either the first or last author. For each scientist, we included only those papers that meet each authorship condition, and re-ran all analyses. Fig 3 shows that across all 4 authorship conditions, 5 scientific fields, and the groups of prominent and random scientists, the chaining models persistently better predicted the historical data than the null model and exemplar-based chaining is the best overall model (binomial test: p < 0.0008; n = 39). These results indicate that the regularity underlying the historical emergence of scientific papers is not affected by the degree of collaboration in a scientist’s work.

To verify that the best-performing model captures genuine temporal effects, we performed a shuffle test where we randomized the emerging order of papers over the timeline such that the same number of papers emerged at each timestep, but papers were assigned with distorted timestamps. We shuffled 100 times and compared the predominance of the best model in the ground truth timeline to the predominance of the best model in the shuffled cases. If the success of exemplar-based chaining reflects the temporal emergence of scientific papers, the predominance of this chaining model in the ground truth condition should be significantly higher than that in the shuffled conditions. Fig 4A and 4B summarize the results from the shuffle tests for prominent and random scientists across the 5 fields. In all but one case, we found that the dominance of the exemplar model over alternative models is significantly higher in the ground truth condition than in the time-shuffled condition (permutation test: p < 0.01 in all cases, except for the prominent computer scientists group p = 0.06). These results indicate that the success of the exemplar chaining model reflects the true historical emergence of scientific publications.

Fig 4. Summary of results from the time-shuffled analysis.

Fig 4

(A) Comparison of the predominance of the exemplar model (model winning percentage indicated by the red dotted line) against the same statistics recorded in the 100 iterations from the time-shuffled tests (winning percentages shown in the boxplot with an accompanying histogram and scattered raw data). (B) Similar information for the random-sampled scientists.

Although the exemplar model is predominant through all the predictive analyses, we observed that there is variation among the chaining models in predicting scientists’ data. In particular, we found the progenitor model to be persistently the second best-performing model (see Fig 3 and Table in S2 Table for the breakdown of model performance). This result shows that despite the fact that exemplar-based chaining may be generally preferred, this preference reflects a statistical tendency and there exist alternative mechanisms in the emergence of scientific publications. For instance, the progenitor model reflects the strategy where a scientist tends to make incremental progress toward a base theme. This model of chaining is statistically less salient than the exemplar-based strategy in our data, but it may be nevertheless adopted quite commonly by scientists. These findings indicate that there are varying strategies in how scientists make progress over time, with exemplar-based strategy being the prevalent kind.

Discussion and conclusion

Our study offers an initial step toward quantifying and evaluating the regularity in the historical emergence of scientists’ outputs across individuals and multiple scientific fields. Our findings reveal a resemblance between the processes of scientific publishing behavior over time and historical word meaning extension: both processes rely on mechanisms of chaining that grow ideas over time based on closely related existing ideas, which suggests that chaining may serve as a general-purpose mechanism in the extension of human knowledge.

We acknowledge that there are aspects that are not captured by our current analysis and we do not suggest our account to be a full picture of scientists’ discovery processes, but rather a high level framework which can be built upon for future work.

One factor influencing scientists’ research output that can be further explored is the role of collaboration. Although we analyzed papers written by different numbers of coauthors, we did not examine in detail how the activities of other researchers might impact scientists. This was partly due to data limitations, since we focused on analyzing the sequences of papers and did not consider the collaboration network of scientists. This analysis can be challenging to carry out with the historical Nobel prize winners, since we do not know of any digitized records that would allow us to build a complete collaborative network of scientists from the earlier periods. However, exploring models that take into account collaboration would be an interesting avenue for future research.

The primary focus of our analysis is on examining which of the cognitively plausible models might best account for the historical emergence of scientists’ publications. There may be more fine-grained patterns which we are unable to explore here, such as the variation in behavior over the course of career and the variation in behavior due to scientists’ differential demographics (e.g., the comparison of senior, more established scientists versus junior scientists).

Throughout our large-scale analysis in multiple scientific fields, we found consistent evidence for exemplar-based chaining in award-winning and random-sampled scientists in their historical emergent patterns of publication. The resolution of our analysis focuses on the level of scientists, so our computational framework does not speak to the emerging properties of individual publications. However, our analysis does reveal that scientists who have done prominent work share similar cognitive strategies in scientific inquiry through time with those who do “normal” science, such that they tend to make progress in continuous ways reflecting a process of chaining that mirrors the processes in historical word meaning extension.

Supporting information

S1 Appendix. Validation of semantic representation.

Details of how validation of semantic representations across fields and scientists was performed is included.

(PDF)

S1 Fig. Validation of semantic representations at field-level.

Validation of semantic representation at the scientific-field level. Comparison of the attested J-value (red dotted line) against the J-values from the shuffled trials across scientific fields (green histogram and boxplot with scattered raw data) in (A) prominent scientists, and (B) random-sampled scientists.

(EPS)

S2 Fig. Validation of semantic representations at scientist-level.

Validation of semantic representation at the individual-scientist level. Comparison of the attested J-value (red dotted line) against the J-values from the shuffled trials across scientists (green histogram and boxplot with scattered raw data) within each of the 5 scientific fields in (A) prominent scientists, and (B) random-sampled scientists. “CS” stands for computer science.

(EPS)

S1 Table. Summary statistics from evaluation against the null model, in prominent and randomly-sampled scientists across fields.

Statistics from model evaluation against the null model in log-likelihood ratio for the award-winning scientists as described in the main text, using Mann-Whitney U test. The Mann-Whitney U test was used in place of a t-test because the data were not normally-distributed (all models outperformed the null, and this is highly significant).

(PDF)

S2 Table. Percentage breakdowns from winner-take-all analyses.

Tthe percentage breakdowns from the winner-take-all analyses described in the main text for the group of award-winning scientists as well as randomly-sampled scientists. The percentage of the model with the highest winning percentage across scientists within each scientific field is marked in bold.

(PDF)

Data Availability

All data and code are available at https://github.com/nightingal3/metascience/tree/master.

Funding Statement

YX was supported by NSERC Discovery Grant RGPIN-2018-05872 (https://www.nserc-crsng.gc.ca/professors-professeurs/grants-subs/dgigp-psigp_eng.asp), SSHRC Insight Grant #435190272 (https://www.sshrc-crsh.gc.ca/funding-financement/programs-programmes/insight_grants-subventions_savoir-eng.aspx), and Ontario Early Researcher Award #ER19-15-050 (https://www.ontario.ca/page/early-researcher-awards) The funders played no role in study design, data collection and analysis, or preparation of the manuscript.

References

  • 1.Westfall RS. Isaac Newton. In: Encyclopedia Britannica; 2021. Available from: https://www.britannica.com/biography/Isaac-Newton.
  • 2.Copeland BJ. Alan Turing. In: Encyclopedia Britannica; 2021. Available from: https://www.britannica.com/biography/Alan-Turing.
  • 3. Kuhn TS. The Structure of Scientific Revolutions. Chicago: University of Chicago Press; 1962. [Google Scholar]
  • 4. Popper K. The Logic of Scientific Discovery. Basic Books; 1959. [Google Scholar]
  • 5. Fayerabend P. Explanation, Reduction, and Empiricism. In: Feigl H, Maxwell G, editors. Scientific Explanation, Space, and Time. University of Minneapolis Press; 1962. [Google Scholar]
  • 6. Toulmin S. Does the Distinction between Normal and Revolutionary Science Hold Water? In: Lakatos I, Musgrave A, editors. Criticism and the Growth of Knowledge: Proceedings of the International Colloquium in the Philosophy of Science, 1965. Cambridge University Press; 1970. p. 39–48. [Google Scholar]
  • 7. Hanson NR. Patterns of Discovery. Cambridge: Cambridge University Press; 1958. [Google Scholar]
  • 8. Chang H. Inventing Temperature: Measurement and Scientific Progress. Oxford: Oxford University Press; 2004. [Google Scholar]
  • 9. Griffiths TL, Steyvers M. Finding scientific topics. Proceedings of the National Academy of Sciences. 2004;101(suppl 1):5228–5235. doi: 10.1073/pnas.0307752101 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Prabhakaran V, Hamilton WL, McFarland D, Jurafsky D. Predicting the Rise and Fall of Scientific Topics from Trends in their Rhetorical Framing. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics; 2016.
  • 11. Zeng A, Shen Z, Zhou J, Fan Y, Di Z, Wang Y, et al. Increasing trend of scientists to switch between topics. Nature Communications. 2019;10:3439. doi: 10.1038/s41467-019-11401-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Zeng A, Shen Z, Z J, Wu J, Fan Y, Wang Y, et al. The science of science: From the perspective of complex systems. Physics Reports. 2017;714-715:1–73. doi: 10.1016/j.physrep.2017.10.001 [DOI] [Google Scholar]
  • 13. Shi F, Foster JG, Evans JA. Weaving the Fabric of Science: Dynamic network models of science’s unfolding structure. Social Networks. 2015;43:73–85. doi: 10.1016/j.socnet.2015.02.006 [DOI] [Google Scholar]
  • 14. Jia T, Wang D, Szymanski BK. Quantifying patterns of research-interest evolution. Nature Human Behaviour. 2017;1(4). doi: 10.1038/s41562-017-0078 [DOI] [Google Scholar]
  • 15. Wu L, Wang D, Evans JA. Large teams develop and small teams disrupt science and technology. Nature. 2019;566(7744):378–382. doi: 10.1038/s41586-019-0941-9 [DOI] [PubMed] [Google Scholar]
  • 16. Battiston F, Musciotto F, Wang D, Barab si AL, Szell M, Sinatra R. Taking census of physics. Nature Reviews Physics. 2019;1(1):89–97. doi: 10.1038/s42254-018-0005-3 [DOI] [Google Scholar]
  • 17. Kneeland MK, Schilling MA, Aharonson BS. Exploring Uncharted Territory: Knowledge Search Processes in the Origination of Outlier Innovation. Organization Science. 2020;31:535–557. doi: 10.1287/orsc.2019.1328 [DOI] [Google Scholar]
  • 18. Lin Y, Evans JA, Wu L. New Directions in Science Emerge from Disconnection and Discord; 2021. Available from: https://arxiv.org/abs/2103.03398. [Google Scholar]
  • 19. Lakoff G. Women, fire, and dangerous things: What Categories Reveal About The Mind. Chicago: University of Chicago Press; 1987. [Google Scholar]
  • 20. Malt BC, Sloman SA, Gennari S, Shi M, Wang Y. Knowing versus naming: Similarity and the linguistic categorization of artifacts. Journal of Memory and Language. 1999;40:230–262. doi: 10.1006/jmla.1998.2593 [DOI] [Google Scholar]
  • 21. Bybee JL, Perkins RD, Pagliuca W. The evolution of grammar: Tense, aspect, and modality in the languages of the world. Chicago: University of Chicago Press; 1994. [Google Scholar]
  • 22. Habibi AA, Kemp C, Xu Y. Chaining and the growth of linguistic categories. Cognition. 2020;202:104323. doi: 10.1016/j.cognition.2020.104323 [DOI] [PubMed] [Google Scholar]
  • 23. Ramiro C, Srinivasan M, Malt B, Xu Y. Algorithms in the historical emergence of word senses. Proceedings of the National Academy of Sciences. 2018;115:2323–2328. doi: 10.1073/pnas.1714730115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Grewal K, Xu Y. Chaining and historical adjective extension. In: Proceedings of the 42nd Annual Meeting of the Cognitive Science Society; 2020.
  • 25.Yu L, Xu Y. Predicting emergent linguistic compositions through time: Syntactic frame extension via multimodal chaining. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing; 2021.
  • 26. Sloman SA, Malt BC, Fridman A. In: Hahn U, Ramscar M, editors. Categorization versus similarity: The case of container names. New York: Oxford University Press; 2001. p. 73–86. [Google Scholar]
  • 27. Xu Y, Regier T, Malt BC. Historical semantic chaining and efficient communication: The case of container names. Cognitive Science. 2016;40:2081–2094. doi: 10.1111/cogs.12312 [DOI] [PubMed] [Google Scholar]
  • 28. Nosofsky RM. Attention, similarity, and the identification–categorization relationship. Journal of Experimental Psychology: General. 1986;115:39–57. doi: 10.1037/0096-3445.115.1.39 [DOI] [PubMed] [Google Scholar]
  • 29. Loreto V, Servedio VDP, Strogatz SH, Tria F. Dynamics on Expanding Spaces: Modeling the Emergence of Novelties. Creativity and Universality in Language. 2016; p. 59–83. doi: 10.1007/978-3-319-24403-7_5 [DOI] [Google Scholar]
  • 30. Rosch E. Cognitive representations of semantic categories. Journal of Experimental Psychology: General. 1975;104:192–233. doi: 10.1037/0096-3445.104.3.192 [DOI] [Google Scholar]
  • 31. Geeraerts D. Diachronic prototype semantics: A contribution to historical lexicology. Oxford: Oxford University Press; 1997. [Google Scholar]
  • 32. Luce RD. Individual Choice Behavior: A Theoretical Analysis. New York: Wiley; 1959. [Google Scholar]
  • 33. Shepard RN. Stimulus and response generalization: Tests of a model relating generalization to distance in psychological space. Journal of Experimental Psychology. 1958;55(6):509–523. doi: 10.1037/h0042354 [DOI] [PubMed] [Google Scholar]
  • 34. Shepard RN. Stimulus and response generalization: Deduction of the generalization gradient from a trace model. Psychological Review. 1958;65(4):242–256. doi: 10.1037/h0043083 [DOI] [PubMed] [Google Scholar]
  • 35. Hofstra B, Kulkarni VV, Galvez SMN, He B, Jurafsky D, McFarland DA. The diversity–innovation paradox in science. Proceedings of the National Academy of Sciences. 2020;117(17):9284–9291. doi: 10.1073/pnas.1915378117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Murphy MC, Mejia AF, Mejia J, Yan X, Cheryan S, Dasgupta N, et al. Open science, communal culture, and women’s participation in the movement to improve science. Proceedings of the National Academy of Sciences. 2020;117(39):24154–24164. doi: 10.1073/pnas.1921320117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Reimers N, Gurevych I. Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing; 2019.
  • 38. Van der Maaten LJP, Hinton G. Visualizing data using t-SNE. Journal of Machine Learning Research. 2008;9:2579–2605. [Google Scholar]
  • 39. Medin DL, Schaffer MM. Context theory of classification learning. Psychological Review. 1978;85:207–238. doi: 10.1037/0033-295X.85.3.207 [DOI] [Google Scholar]
  • 40. Bergstra J, Komer B, Eliasmith C, Yamins D, Cox DD. Hyperopt: a Python library for model selection and hyperparameter optimization. Computational Science and Discovery. 2015;8. doi: 10.1088/1749-4699/8/1/014008 [DOI] [Google Scholar]
  • 41.The DBLP team. DBLP computer science bibliography; 2019. Available from: https://dblp.org.
  • 42.Ammar W, Groeneveld D, Bhagavatula C, Beltagy I, Crawford M, Downey D, et al. Construction of the Literature Graph in Semantic Scholar. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies; 2018.
  • 43.Luu K, Wu X, Koncel-Kedziorski R, Lo K, Cachola I, Smith NA. Explaining Relationships Between Scientific Documents. In: Proceedings of the Joint Conference of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing; 2021.
  • 44.Cachola I, Lo K, Cohan A, Weld DS. Tldr: Extreme summarization of scientific documents. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing; 2020.
  • 45.Cohan A, Dernoncourt F, Kim DS, Bui T, Kim S, Chang W, et al. A discourse-aware attention model for abstractive summarization of long documents. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies; 2018.

Decision Letter 0

Qichun Zhang

5 Sep 2022

PONE-D-22-15379Regularity in the temporal dynamics of publishing behaviour across scientistsPLOS ONE

Dear Dr. Liu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

ACADEMIC EDITOR: Please consider all the comments during the revision and proof read the revised submission.

==============================

Please submit your revised manuscript by Oct 20 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Qichun Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please update your submission to use the PLOS LaTeX template. The template and more information on our requirements for LaTeX submissions can be found at http://journals.plos.org/plosone/s/latex.

Additional Editor Comments :

Some comments have been received from both reviewers and a major revision is helpful to clarify the concerns then enhance the quality of the work.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This is an interesting article highlighting the similarities in the pattern of publishing behaviour of individual scientists over time and the historical word meaning extension. To this end the authors have employed the cognitive theory chaining based on general context model.

The article, though is very well written, it rather very briefly describes the computational framework. It would be better to add more analytical details which would aid the reader to better appreciate the methodology employed.

The article is recommended for publication after the addition of more details concerning model formulation.

Reviewer #2: The paper investigated an important 'social behaviour' using computing technologies. Basically, the topic is interesting for the academics. However, some issues should be taken into account during a major revision.

1) The motivation is not clear, most of the scientific findings are given based on one single amazing idea and the original idea cannot be predicted. So why we need this statistical research work?

2) Various methods have been used in the manuscript such as clustering method KNN, why these methods have been selected? Is the comparative study the main contribution?

3) In figure 1, yellow dot data has been marked as 2025. Where is this data from? I understand that the work is based on historical data.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Dec 29;17(12):e0278389. doi: 10.1371/journal.pone.0278389.r002

Author response to Decision Letter 0


18 Oct 2022

=== Reviewer 1 ===

We appreciate your review and recommendations on our work. We have revised the manuscript to motivate the computational framework better, and to make the formulation of models clearer.

> It would be better to add more analytical details which would aid the reader to better appreciate the methodology employed.

Thank you for your feedback, we have added further description to paragraph 1 of the ``Computational Framework'' section, in which we clarify our methodology and framing. We have also added additional details about SBERT, which is the embedding model we used, in the same section.

> The article is recommended for publication after the addition of more details concerning model formulation.

Thank you, we have added exposition to paragraph 5 of the ``Computational Framework'' section. In particular, each of the models examined is specified completely by Table 1 and Equation 1. We have made this more explicit in the manuscript in paragraph 11 of this section.

=== Reviewer 2 ===

We would like to thank you for your feedback. We clarify our methods and motivation behind the paper: we are not seeking to predict what the amazing ideas that actually emerge will be, but rather trying to characterize the overall process that scientists use to extend their ideas. We have also added more detail to the "Computational Framework" section to make the motivation more clear to readers.

> The motivation is not clear, most of the scientific findings are given based on one single amazing idea and the original idea cannot be predicted. So why we need this statistical research work?

Thank you for this clarification question. We wish to emphasize that the purpose of the paper is not to predict what amazing single scientific ideas will arise, but rather to characterize the overall chaining process in a scientist's publications as they emerge through time. Novel scientific ideas have some relationship to the existing scientific literature and a scientist's past work, and it's this relationship between past work and future work that motivates our analysis with the temporal chaining models. This is similar to the word-sense extension models mentioned in paragraph 6 of the Introduction: we cannot predict exactly what novel sense(s) of a word will arise next, but we can characterize the process of sense extension and hence making predictions about the emerging order of senses over time. Similarly, we can characterize the overall growth of a scientist's publications over time, even if we cannot predict exactly what that scientist will do in the future.

> Various methods have been used in the manuscript such as clustering method KNN, why these methods have been selected? Is the comparative study the main contribution?

We have added exposition and motivation to the ``Computational Framework'' section of the manuscript. These methods were selected based on past work in word-sense extension, and comparative study is the main contribution, as all of these extension mechanisms are cognitively plausible and based on psychological research. Another contribution is that we find the best model to be the exemplar model, which mirrors the findings in word-sense extension, and suggests that this may be a natural and general way in which people extend concepts.

> In figure 1, yellow dot data has been marked as 2025. Where is this data from? I understand that the work is based on historical data.

We have data up until 2020, and the plot has now been fixed, since it was caused by matplotlib's automatic label generation.

Attachment

Submitted filename: Response to Reviewers.pdf

Decision Letter 1

Qichun Zhang

16 Nov 2022

Chaining and the temporal dynamics of scientists' publishing behaviour

PONE-D-22-15379R1

Dear Dr. Liu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Qichun Zhang, PhD

Academic Editor

PLOS ONE

Additional Editor Comments:

All the concerns have been addressed well in the revised version and both reviewers satisfy the revision. Therefore, I recommend accepting this manuscript as it is.

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have carried out the revision by adding a paragraph highlighting the analytical details as required and the paper is accepted.

Reviewer #2: All my concerns have been addressed well in the revised version. Thus I think it is ready for publication.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

Acceptance letter

Qichun Zhang

16 Dec 2022

PONE-D-22-15379R1

Chaining and the temporal dynamics of scientists’ publishing behaviour

Dear Dr. Liu:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Qichun Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Validation of semantic representation.

    Details of how validation of semantic representations across fields and scientists was performed is included.

    (PDF)

    S1 Fig. Validation of semantic representations at field-level.

    Validation of semantic representation at the scientific-field level. Comparison of the attested J-value (red dotted line) against the J-values from the shuffled trials across scientific fields (green histogram and boxplot with scattered raw data) in (A) prominent scientists, and (B) random-sampled scientists.

    (EPS)

    S2 Fig. Validation of semantic representations at scientist-level.

    Validation of semantic representation at the individual-scientist level. Comparison of the attested J-value (red dotted line) against the J-values from the shuffled trials across scientists (green histogram and boxplot with scattered raw data) within each of the 5 scientific fields in (A) prominent scientists, and (B) random-sampled scientists. “CS” stands for computer science.

    (EPS)

    S1 Table. Summary statistics from evaluation against the null model, in prominent and randomly-sampled scientists across fields.

    Statistics from model evaluation against the null model in log-likelihood ratio for the award-winning scientists as described in the main text, using Mann-Whitney U test. The Mann-Whitney U test was used in place of a t-test because the data were not normally-distributed (all models outperformed the null, and this is highly significant).

    (PDF)

    S2 Table. Percentage breakdowns from winner-take-all analyses.

    Tthe percentage breakdowns from the winner-take-all analyses described in the main text for the group of award-winning scientists as well as randomly-sampled scientists. The percentage of the model with the highest winning percentage across scientists within each scientific field is marked in bold.

    (PDF)

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    All data and code are available at https://github.com/nightingal3/metascience/tree/master.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES