Abstract
Currently bibliographic databases have included a large number of Early Access (EA) articles. Taking 47 IEEE journals as examples, this study analyzed and compared the differences in publication stages of EA articles in three typical bibliographic databases, including Web of Science Core Collection, Scopus, and Engineering Village Compendex. Qualitative analysis of data sets that may appear in these three databases and their publication stage modes, and quantitative analysis on the number of records, proportion, and journal distributions of each data set and each publication stage mode were conducted. There were totally 7 sub-data sets and corresponding 26 publication stage modes, with 14 “undifferentiated publication stage modes” and 12 “differentiated publication stage modes”. Although the proportion of EA records from each “differentiated publication stage mode” was mostly below 1.0%, the absolute quantity of EA records with differences in the publication stage was noteworthy reaching 2516. Among the 47 journals, 23 journals have 7–8 publication stage modes, 1 journal having 18 modes, and 40 journals have one or more “differentiated publication stage modes”. Therefore, in IEEE journals, whether for the same EA article or the same journal, the difference in publication stage between these three databases was pervasive and complex.
1 Introduction
Early Access (EA) is a common model in the academic publishing industry, derived from dual format (online and print) journals. This means that the article can be obtained online before it is officially published in a print journal. This model may play an important role in promoting the exchange of academic findings. Studies have shown that the availability of EA articles reduced publication delays [1–3], increased reader access [4–5], and provided researchers with the latest E-publications for systematic reviews [6]. In recent years, several popular bibliographic databases, such as Web of Science, Scopus, and Engineering Village, have collaborated with publishers to index a large number of EA records. According to rough search and statistics, as of January 4, 2024, the Web of Science Core Collection (WoS CC) has included 492,091 EA records (2017–2023), Scopus included 415,508 records (2011–2023), and Engineering Village Compendex (EI) included 164,188 records (2006–2023). It can be seen that in addition to journal websites or publishing platforms, bibliographic databases, as an important tool for literature retrieval and acquisition, have become another way for researchers to access EA articles. In addition, it is worth noting that the increasing number of EA records in bibliographic databases makes it possible to calculate new bibliometric indicators and research evaluation indicators. As announced by the Web of Science, EA content will be introduced in the calculation of journal impact factors starting from 2021 [7]. Therefore, EA articles not only change the basic data of the database, but also have an impact on research achievement acquisition, bibliometric analysis, and research performance evaluation.
In bibliographic databases, the indexing method of EA articles is different from traditional indexing methods. Traditionally, database only involves officially published articles (which have been assigned volume, issue, and page numbers), and does not include EA articles. For EA articles, database indexing is usually carried out in two stages. The first stage involves indexing the article during its early access phase, while the second stage involves cataloging it after its official publication. That is to say, EA articles indexed by the database will have two publication stages in sequence – early access stage and official publication stage. In addition, there are differences in the indexing methods for EA articles among different databases. The WoS CC retains the early access date when indexing the official publication stage of the EA article, while Scopus and EI delete completely the records of the early access stage after indexing the official publication stage. However, the introduction of EA articles and specific indexing methods may cause changes in database functionality and even some errors. In the Web of Science database, the “year published” field in search page can search for the online and final publication years of indexed records, while the “publication year” in the search results page only represents the online publication year [8]. Nowadays, WoS CC has added the function key of final publication year to the search results page. In the Scopus database, the retrieval results for Year of Publication and Date of Publication are inconsistent and one possible reason is that the two fields use different versions of the publication date [9]. In Scopus, after indexing the official version of an article, the link to citation obtained by the Online-First version may be lost [10–11]. Due to the different indexing policies of WoS CC and Scopus for publication dates, there may be large discrepancies in the number of records in the same journal included in these two databases [12]. These previous studies undoubtedly remind us that bibliographic databases may produce some problems in the process of EA article indexing. However, they did not involve a specific analysis the publication stage of EA articles in databases.
As the number of EA articles and their additional data in various databases continues to increase, it is necessary and urgent to conduct an evaluation accordingly. For EA articles, the dual publication stage is a unique mark that distinguishes them from pure electronic journal articles and pure print journal articles. It's closely related to the publication time of the article, which is a finer grained level of information than the article level. For bibliographic databases, the unique method of indexing the two publication stages separately makes the publication stage of EA articles a new feature field in the database, which is one of the parameters that may be involved in literature retrieval, literature analysis, and bibliometric work. Today, EA article indexing has become normal in databases. Just like other field information such as affiliation [13–18], funding [19–22], Digital Object Identifier (DOI) [23–26], author [27–29], document type [30–34], and language [32,35], we have to consider potential impact of publication stage in the evaluation, selection and application of databases. If the publication stage of the same EA article is different in different databases, that is, the publication stage information of the same EA article obtained through different databases is different, this will undoubtedly cause confusion, and it is impossible to determine which database is correct. Further, if different EA articles in the same journal have different publication stage modes, which will undoubtedly add to the confusion. To this end, this paper intends to conduct an exploratory study on the publication stage of EA articles and their differences in different bibliographic databases, in order to reveal the current situation and problems of EA article indexing in databases, especially the publication stage indexing. At present, there is little research specifically on the practice of indexing EA articles in bibliographic databases, and there is limited information to refer to. This paper intends to make a comprehensive and systematic analysis of the indexing characteristics and differences of EA article publication stage in common and typical bibliographic databases (including WoS CC, Scopus and EI) by combining qualitative and quantitative methods and examples. It is hoped that the research results can truly reflect the status and problems of publication stage indexing of EA articles in different databases, and provide valuable reference information for the development and use of databases.
2 Literature review
2.1 Research on the publication stage of EA articles
The publication stage of EA articles, especially the role of online publication and the delay between print publication and online publication, has sparked a series of studies. A survey on over 200 journals from different publishers showed that printed versions of journals were typically published an average of 3 months later than online versions [2]. The “advanced online publication” (AOP) mechanism in biomedical journals and food research journals reduced publication delays for papers [1,3]. The delay between print and online publication of neuroscience journals can artificially raise a journal's impact factor [36]. A 2010 survey found that advance online publication increased the impact factor of ophthalmic journals [37]. Online publication of in press articles affected severely the calculation of immediacy index of journals [38]. The publishing time (the period between submission and online availability) was negatively and significantly related to journal impact factor [39]. The impact factor and the immediacy index values of information science journals with early view mechanism were higher than the others [40]. Mendeley readers accrued and continued to steadily increase when articles were first available online [4]. Journals with “articles in press” mechanism had the citation advantage and higher CiteCore value [41]. The articles-in-press mechanism in energy and fuels journals was not significantly related to publication lag, but affected citation speed [42]. The electronic access increased after the articles’ online release and print publication [5].
Other research issues were the ways publishers and databases handle EA articles and their subsequent impact. The treatment of online publishing time by different publishers varied greatly [43–44]. The existence of different dates associated with the publication, such as online date, journal publication month, database index date, first tweet mentioning date, and Altmetric publication and first-seen dates, have caused confusion in the publication date of an article [43]. There were black box problems in the search of publication time in Web of Science and Scopus [8–9]. The indexing method of Scopus database for online-first article led to “citation infanticide” [10–11]. Different or incorrect indexing policies between the Web of Science and Scopus databases may lead to differences in the number of articles in the same journal [12].
Faced with the characteristics and influence of EA articles, some scholars have also discussed the appropriate time parameters for database indexing and bibliometric indication calculation. Tort et al. [36] suggested that databases use online publication dates for indexing article and calculating citation metrics. Heneberg [38] proposed that the immediacy index should be calculated based on the date of first publication (online or print, whichever comes first). Echeverría et al. [44] pointed out that when calculating bibliometric indicators of journals, the Advance Online Publication (AOP) date of articles needs to be transparent and standardized. Hu et al. [45] suggested that using online years may be more appropriate than publication years for the high citation papers (HCP) in ESI. Liu et al. [12] recommended that both Scopus and WoS CC use official publication dates consistently for indexing.
2.2 Comparative Studies of different bibliographic databases
Due to the continuous development and changes in the types and functions of bibliographic databases, there have been a large number of studies on database evaluation with a wide range of topics. Since the 21st century, this research field has received a lot of attention. Related studies have not only assessed the journal coverage [51,53,57,72,74–77,79,80,82,83,85], subject classification [63,64,66], retrieval functions [60,63,78,83], number of publications [49,59,63,65,66,70,84,86–88], citation counts, and associated metrics of databases [46–48,50,52–56,58–63,66–69,71–73,80,81,87], but also evaluated the comprehensiveness, availability, accuracy (or error) of various fields, which are particularly country/institution/address [13–18], funding information [19–22], DOI [23–26], author ID [27–29], document type [30–34], and language [32,35].
Many of database evaluation studies were about comparative analysis between different databases, especially WoS and Scopus, the two major bibliographic databases. In order to show the latest progress of the comparative study of these two databases, the relevant literatures in the past three years are listed in Table 1.
Table 1. Comparative studies of WoS and Scopus in the past three years.
Reference | Database | Comparison content |
---|---|---|
Gusenbauer [75] | 56 bibliographic databases including Google Scholar, Scopus, Web of Science, etc. | Disciplinary coverage |
Purnell [18] | Scopus, Web of Science, Dimensions, and Microsoft Academic | Author affiliation |
Asubiaro [76] | Web of Science, Scopus, EMBASE, MEDLINE, African Index Medicus, and African Journals Online | Journal coverage |
Asubiaro and Onaolapo [77] | Web of Science, Scopus, and CrossRef | Journal coverage |
Kokol [22] | Scopus and Web of Science | Funding information coverage |
Singh et al. [78] | Web of Science、Scopus and Dimensions | Topic retrieval quality |
Asubiaro et al. [79] | Web of Science and Scopus | Journal coverage |
Chinchilla- Rodríguez et al. [29] | Web of Science and Scopus | Corresponding authorship field |
Gerasimov et al. [80] | Google Scholar, Web of Science, Scopus, Crossref, and DataCite | Citation coverage of datasets |
Moustafa [81] | CrossRef, Web of Science, Publisher records, Google Scholar, PubMed, and Scopus | Citations |
Ortega and Delgado-Quirós [82] | Dimensions, OpenAlex, PubMed, Scilit, Scopus, The Lens, and Web of Science | Coverage and overlap of retracted literature |
Zhu et al. [34] | Web of Science, Scopus and journals’ websites | Document type |
Ei Compendex is widely regarded as the top-notch platform for engineering literature. The comparative on it mainly involved coverage, research areas, and retrieval function (Relevant studies are shown in Table 2).
Table 2. Comparative studies related to Ei Compendex.
Reference | Database | Comparison content |
---|---|---|
Bartol and Hocevar [83] | Web of Science, Biosis Previews, CAB Abstracts, Chemical Abstracts, Compendex/Inspec, Francis, Medline, Pascal, and Sociological Abstracts | City-data collection and retrieval accuracy |
Kostoff [84] | Science Citation Index (SCI), INSPEC, and Ei Compendex | Scientific and technological outputs |
Meier and Conkling [85] | Google Scholar and Compendex | Coverage of engineering literature |
Cusker [86] | Google Scholar and Elsevier Compendex | Retrieval capabilities |
Cole et al. [87] | Compendex, Scopus, and Google Scholar | Citation discoverability |
McAllister and Torres [88] | Web of Science and Ei Compendex | OA publications |
3 Research question
In view of the different indexing methods of EA articles in different literature databases and their potential impact on the publication stage field, this study will take IEEE journals as an example, analyze and compare the differences of publication stage of EA articles in three typical bibliographic databases, including Web of Science Core Collection (WoS CC), Scopus and Engineering Village compendex (EI). The specific research questions include:
Differences in publication stages of the same EA article in WoS CC, Scopus and EI.
Differences in publication stage modes of different EA articles from the same journal in WoS CC, Scopus and EI.
4 Methodology
4.1 Selection of journals
IEEE (Institute of Electrical and Electronics Engineers) is one of the world's famous professional societies. Combined with the author's previous research on EA articles and their publishers (journals) in WoS CC [89], although IEEE journals do not contain a large number of EA articles, they have a certain scale, ranking 9th in 2018–2022. Therefore, considering the academic status of IEEE journals and their number of EA articles (appropriate amount of data analysis work), the author decided to choose IEEE journals as case for this study. Regarding the database, the author mainly considers whether it includes IEEE journals and its international popularity. In addition to the two comprehensive bibliographic databases (WoS CC and Scopus), the most famous engineering database (EI) in the world is also used in this study.
In order to meet the research purposes of this paper, the selected journal must meet three conditions: (1) It must be included in WoS CC, Scopus, and EI databases simultaneously; (2) It must provide an EA publication policy; (3) It must be indexed all by the EA indexing method in the above three databases. The selection process was as follows: First, in WoS CC, the search strategy of [Publisher = IEEE and Document Type = (Article OR Review)] was adopted to obtain 301 publication titles. Next, these publication titles were searched one by one on the IEEE Xplore platform. Publication titles that served as conference proceedings, did not have EA publication policies, or could not be found in IEEE were delete. The WoS database started indexing EA content in December 2017, so publication titles without publication records from 2017 and onwards were also deleted. The number of eligible journals was reduced to 186. To meet condition (3), these 186 journals were further screened to ensure that each journal contained at least one “EA record,” resulting in a list of 49 journals. “EA record” here refer to the record indexed by the WoS CC database using EA indexing method, identified by the “Early Access” tag in the Document Type field or the “Early Access Date” field [90]. These 49 journals were then checked in the Scopus and EI, and 2 of them were not found. Finally, 47 journals were identified. Of these 47 journals, only one had a unique situation, having only EA records in WoS CC and only official publication records in Scopus and EI during the period of data collection for this study. The author made a special check on this journal. According to the search results before and after the data collection time in this paper, this journal had EA records in Scopus and EI, which met the condition (3). Therefore, through the above screening process, this paper has obtained the most IEEE journals that meet the research requirements.
Here, this study used WoS CC for journal selection due to the different EA indexing rules of the three databases. Specifically, the WoS CC database retains indexing information for both early access records and official publication records, while Scopus and EI delete early access records after indexing official publication records. If a journal's early access records in Scopus and EI are deleted during data retrieval, it will not be possible to determine whether this journal meet the requirements for this study. As with the special case above, this paper has avoided this problem by using WoS CC. If a journal has early access records in Scopus and EI, but only official publication records in WoS CC, this means that WoS CC's indexing rules for this journal are still traditional rules, rather than EA indexing rules [does not meet condition (3)], so using WoS CC has avoided some unnecessary workload.
4.2 Creation of data sets
Records of the above 47 journals published in 2021–2023 were searched in the three databases. For the same journal, the search was completed on the same day. The total records of 43,011 (WoS CC), 45,212 (Scopus), and 44,937 (EI) were collected from 20 June to 2 July 2023. The records for individual journals ranged from 87 to 4340 (WoS CC), 83–4531 (Scopus), and 83 to 4489 (EI).
In this study, all records from three databases were pair-to-pair compared. The DOI served as the criterion for determining whether a record was the same. This paper classified records into three types of data sets based on the number of databases in which they were located: records with unique DOI in one database (Type 1), records with the same DOI in two databases (Type 2), and records with the same DOI in three databases (Type 3). The meanings of these three types of records were as follows: records that appeared in only one database, records that appeared in two databases, and records that appeared in three databases. According to the different databases, these three types of data sets were further divided into 7 sub-data sets (see Table 3).
Table 3. Information of data sets and publication stage modes.
Type of data set | DOI matching | Database | Sub-data set | Publication stage mode | Label of publication stage mode | Whether there is a difference in publication stages |
---|---|---|---|---|---|---|
Type 1: Records with a unique DOI in one database | DOIW ≠ DOIS ≠ DOIE | WoS CC | D1 | PSW = EA | PS1 | No |
PSW = OP | PS2 | No | ||||
Scopus | D2 | PSS = EA | PS3 | No | ||
PSS = OP | PS4 | No | ||||
EI | D3 | PSE = EA | PS5 | No | ||
PSE = OP | PS6 | No | ||||
Type 2: Records with the same DOI in two databases | DOIW = DOIS ≠ DOIE | WoS CC and Scopus | D4 | PSW = PSS = EA | PS7 | No |
PSW = PSS = OP | PS8 | No | ||||
PSW = EA, PSS = OP | PS9 | Yes | ||||
PSW = OP, PSS = EA | PS10 | Yes | ||||
DOIW = DOIE ≠ DOIS | WoS CC and EI | D5 | PSW = PSE = EA | PS11 | No | |
PSW = PSE = OP | PS12 | No | ||||
PSW = EA, PSE = OP | PS13 | Yes | ||||
PSW = OP, PSE = EA | PS14 | Yes | ||||
DOIS = DOIE ≠ DOIW | Scopus and EI | D6 | PSS = PSE = EA | PS15 | No | |
PSS = PSE = OP | PS16 | No | ||||
PSS = EA, PSE = OP | PS17 | Yes | ||||
PSS = OP, PSE = EA | PS18 | Yes | ||||
Type 3: Records with the same DOI in three databases | DOIW = DOIS = DOIE | WoS CC, Scopus and EI | D7 | PSW = PSS = PSE = EA | PS19 | No |
PSW = PSS = PSE = OP | PS20 | No | ||||
PSW = PSS = EA, PSE = OP | PS21 | Yes | ||||
PSW = PSE = EA, PSS = OP | PS22 | Yes | ||||
PSS = PSE = EA, PSW = OP | PS23 | Yes | ||||
PSW = PSS = OP, PSE = EA | PS24 | Yes | ||||
PSW = PSE = OP, PSS = EA | PS25 | Yes | ||||
PSS = PSE = OP, PSW = EA | PS26 | Yes |
Note: W, S, and E in the DOI subscript in the table represented WoS CC, Scopus, and EI, respectively.
According to some previous studies, there may be omissions and errors in DOI in bibliographic databases [23–26]. In this study, the problem of DOI missing was also found in the above comparison. There were 48 DOI missing records in WoS CC (0.11% of the total records 43011), 7 in EI (0.02% of the total records 44937), and none in Scopus. There were several possible cases of missing DOI. (1) The record existed in only one database, and the DOI was missing; (2) The record was missing the DOI in one of the databases, but had the same DOI in the other two databases; (3) The same record had two indexed records in a database, one of which was an EA record without a DOI (belonging to the DOI missing record), and the other was an official publication record with a DOI. Due to the low proportion of missing DOI records, this paper directly classified them into the Type 1 data set for analysis, and did not do further processing (such as using the title to check, etc.). For possible DOI errors, such as the DOI of a particular article in different databases may be different (not a problem of missing DOI), this paper did not make special comparison and confirmation.
4.3 Qualitative analysis of publication stage modes
Although EA records in the three databases have different field markers, they have the same two basic Publication Stages (PS), namely Early Access (EA) and Official Publication (OP). Combining the three databases in this study, there were six publication stages: PSW = EA, PSS = EA, PSE = EA, PSW = OP, PSS = OP, PSE = OP. Here, W, S, and E in the PS subscript represented WoS CC, Scopus, and EI, respectively.
The seven sub-data sets above had a total of 26 possible combinations of publication stages, which were referred to as “publication stage modes” in this paper. These modes were labeled separately in Table 3. Among these 26 modes, 14 were defined as the “undifferentiated publication stage modes”, including publication stages of unique records and those of the same record that did not differ across different databases. This type of modes covered 6 modes (PS1-PS6) in D1-D3 and 8 modes (PS7, PS8, PS11, PS12, PS15, PS16, PS19, PS20) in D4-D7. The remaining 12 modes (PS9, PS10, PS13, PS14, PS17, PS18, PS21-PS26) were defined as “differentiated publication stage modes”, meaning that publication stages of the same record varied across different databases.
4.4 Quantitative analysis of publication stage modes
Based on the above data sets and publication stage modes, this paper quantitatively analyzed the Record Numbers (RN), Record Proportion (RP), and their average values of various data sets and publication stage modes. To analyze the distribution of journals, this study identified journals whose RN ≠ 0 and counts the corresponding Journal Numbers (JN). The specific process and calculation equation was as follows.
a. The RN of each journal in the 7 sub-data sets (RNi,j) was counted. Equation 1 was used to calculate the average value of RNi,j of a certain sub-data set (), and equations 2 and 3 were used to calculate the RP of each journal (RPi,j) and its average value ().
(1) |
(2) |
(3) |
b. The RN of the 26 publication stage modes of each journal (RNi,k) was counted. Equation 4 was used to calculate the average value of RNi,k of a certain publication stage mode (). Equation 5 was used to calculate the RP of a certain publication stage mode in its sub-data set (RPi,k), and equation 6 was used to calculate the average value of RPi,k ().
(4) |
(5) |
(6) |
In equations 1–6, i was the i journal (a total of 47 journals), j was the j sub-data set (a total of 7 sub-data sets), and k was the k publication stage mode (a total of 26 publication stage modes). The value range of k in equation 5 was divided into 1–2, 3–4, 5–6, 7–10, 11–14, 15–18, and 19–26 according to the sub-data set and its corresponding publication stage modes.
-
c
To analyze the distribution of journals in each sub-data set, journals with RNi,j ≠ 0 were identified, and the Journal Numbers (JNj) were counted. To investigate the distribution of journals in each publication stage mode, journals with RNi,k ≠ 0 were identified, and the Journal Numbers (JNk) were counted.
-
d
To analyze and visually display the overall situation of the RPi,j and the distribution of journals in each sub-data set, this paper drew figures of the RPi,j in three types of data sets (Type 1, Type 2, and Type 3) respectively, and made a comparison.
-
e
To analyze and visually display the RPi,k and the journals distribution of publication stage modes contained in each sub-data set, this paper drew figures of the RPi,k of publication stage modes contained in each sub-data set, respectively, and made a comparison.
-
f
Based on Table 3, the calculation results of equation 2 and equation 5, and the statistical results of step c, this paper counted the number of sub-data sets, the number of publication stage modes, and the number of differentiated publication stage modes contained in each journal, and displayed with figures.
4.5 Analysis of research questions
For research question 1, the main contents of the analysis include: Is there a difference in the publication stage of the same EA article in different databases? If there are differences, what is the publication stage mode? What is the quantity and proportion? The general analysis idea is: First determine what type of data set that the record belongs to. If it is Type 1, it means that there is no so-called difference in the publication stage, as shown in the six publication stage modes listed in Table 3 (PS1-PS6). If it is Type 2 and Type 3, then further analyze the proportion of PS7-PS26 listed in Table 3. There are 8 “undifferentiated publication stage modes”, including PS7, PS8, PS11, PS12, PS15, PS16, PS19, PS20, and 12 “differentiated publication stage modes”, including PS9, PS10, PS13, PS14, PS17, PS18, PS21-PS26. If the number or proportion of these 12 differentiated modes is not 0, it means that the publication stage of the same record is different in different databases. Through the specific number and proportion information, we can know whether these 12 modes exist, which modes are the main ones. And through the number of journal distributions, we can also know whether such differences are common and whether it has certain rules.
For research question 2, the main contents of the analysis include: Is there a difference in the publication stage of different EA articles from the same journal in different databases? If there are differences, how many publishing stage modes are there? What is it exactly? What is the quantity and proportion? The analysis can be divided into three steps. The first step is to analyze the record number, proportion and average value of each data set in each journal. The second step is to analyze the record number, proportion and average value of each publication stage mode in each journal. The third step, based on the results of the first and second steps, analyzes how many types of data set that each journal includes, how many publication stage modes involved and how many “differentiated publication stage modes” involved. If more than one data set type or publication stage mode is involved in a journal, it can indicate that the publication stage mode of different EA articles in the journal is different. Based on the above results, we can also obtain the journal distribution number of each data set, the journal distribution number of each publication stage mode, and the relationship between the journal distribution number and the number of data sets, the number of publication stage modes, and the number of “differentiated publication stage modes”. Through the above journal distribution information, we can explain whether the difference is universal. At the same time, we can analyze whether the difference among different journals has certain regularity by combining the specific information of quantity and proportion.
5 Results and discussions
5.1 Research question 1- Differences in the publication stage of the same EA article
By analyzing the database where records were located, it was found that the records from the 47 IEEE journals involve all seven sub-data sets of three types (D1-D7). The number of records in descending order was D7 (WoS CC, Scopus, and EI) ˃ D6 (Scopus and EI) ˃ D1 (WoS CC) ˃ D2 (Scopus) ˃ D3 (EI) ˃ D5 (WoS CC and EI) ˃ D4 (WoS CC and Scopus). Among them, D7 accounted for the largest proportion, reaching 89.96% (see Table 4).
Table 4. The RN, RP, JN of D1-D7 and PS1-PS26.
Sub-data set | (%) | JN j | Publication stage mode | (%) | JN k | ||||
---|---|---|---|---|---|---|---|---|---|
D1 | 743 | 16 | 3.06 | 42 | PS1 | 157 | 3 | 18.44 | 20 |
PS2 | 586 | 12 | 70.92 | 39 | |||||
D2 | 349 | 7 | 0.71 | 34 | PS3 | 229 | 5 | 62.27 | 32 |
PS4 | 120 | 3 | 10.07 | 7 | |||||
D3 | 67 | 1 | 0.14 | 5 | PS5 | 13 | 0 | 7.56 | 5 |
PS6 | 54 | 1 | 3.08 | 2 | |||||
D4 | 41 | 1 | 0.19 | 13 | PS7 | 1 | 0 | 0.71 | 1 |
PS8 | 38 | 1 | 24.11 | 12 | |||||
PS9 | 2 | 0 | 2.84 | 2 | |||||
PS10 | 0 | 0 | 0.00 | 0 | |||||
D5 | 51 | 1 | 0.21 | 9 | PS11 | 16 | 0 | 7.45 | 4 |
PS12 | 33 | 1 | 8.87 | 5 | |||||
PS13 | 0 | 0 | 0.00 | 0 | |||||
PS14 | 2 | 0 | 2.84 | 2 | |||||
D6 | 2646 | 56 | 5.73 | 47 | PS15 | 1906 | 41 | 63.35 | 44 |
PS16 | 691 | 15 | 35.65 | 39 | |||||
PS17 | 20 | 0 | 0.41 | 1 | |||||
PS18 | 29 | 1 | 0.59 | 1 | |||||
D7 | 42176 | 897 | 89.96 | 47 | PS19 | 4564 | 97 | 10.04 | 40 |
PS20 | 35149 | 748 | 85.10 | 47 | |||||
PS21 | 241 | 5 | 0.15 | 1 | |||||
PS22 | 69 | 1 | 0.04 | 1 | |||||
PS23 | 224 | 5 | 0.68 | 16 | |||||
PS24 | 187 | 4 | 0.12 | 2 | |||||
PS25 | 210 | 4 | 0.31 | 20 | |||||
PS26 | 1532 | 33 | 3.55 | 33 |
Note: RN, RP, JN in the table represented Record Numbers, Record Proportion, and Journal Numbers, respectively. In the subscript, i was the i journal (a total of 47 journals), j was the j sub-data set (a total of 7 sub-data sets), and k was the k publication stage mode (a total of 26 publication stage modes).
In the Type 1 data set (D1, D2, and D3), records existed in only one database and there was no difference in their publication stages. Both the number (RN) and proportion (RP) of such records were low (as shown in Table 4 and Fig 1). The average proportion of D1 (WoS CC) records () was 3.06%, and of D2 (Scopus) and D3 (EI) records were both less than 1.00%.
Fig 1. The record proportion of data sets (D1-D7).
In the Type 2 data set (D4, D5, and D6), records existed in two databases simultaneously. As could be seen from Table 4, the absolute number of D4 (WoS CC and Scopus) and D5 (WoS CC and EI) that differed in the publication stage of PS9 (PSW = EA, PSS = OP) and PS14 (PSW = OP, PSE = EA) was relatively small (only 4 records), while the absolute number of D6 (Scopus and EI) that differed in the publication stage of PS17 (PSS = EA, PSE = OP) and PS18 (PSS = OP, PSE = EA) was relatively large (49 records in total). In D4 (WoS CC and Scopus), PS8 (PSW = PSS = OP) was the main publication stage mode, PS10 (PSW = OP, PSS = EA) had no records, while PS7 (PSW = PSS = EA) and PS9 (PSW = EA, PSS = OP) only had 1 and 2 records, respectively. In D5 (WoS CC and EI), PS11 (PSW = PSE = EA) and PS12 (PSW = PSE = OP) were the main publication stage modes. PS13 (PSW = EA, PSE = OP) had no records, while PS14 (PSW = OP, PSE = EA) only had 2 records. In D6 (Scopus and EI), PS15 (PSS = PSE = EA) and PS16 (PSS = PSE = OP) were the main publication stage modes. PS17 (PSS = EA, PSE = OP) and PS18 (PSS = OP, PSE = EA) had 20 and 29 records respectively, but they all came from only one journal.
In the Type 3 data set (D7), records existed in three databases simultaneously. Table 4 showed that the total number of records () of PS20 (PSW = PSS = PSE = OP) was 35149, ranking first among all publication stage modes, with an average proportion of 85.10%, significantly higher than other models. The PS19 (PSW = PSS = PSE = EA) ranked second with the record number of 4564. The absolute number of records with “differentiated publication stage models” (PS21-PS26) is large (2463 in total), and the proportion of PS26 (PSS = PSE = OP, PSW = EA) was the highest (3.55%), while the proportion of the other 5 modes was less than 1.00%.
In summary, the results showed that the publication stage of the same EA article in IEEE journals is different in different databases. In the above 12 “differentiated publication stage modes”, in addition to PS10 (PSW = OP, PSS = EA) and PS13 (PSW = EA, PSE = OP), the remaining 10 modes all appeared, of which PS26 (PSS = PSE = OP, PSW = EA) was the most common. The relative number of records (proportion) of each “differentiated publication stage model” was mostly below 1.00%, which is “uncommon”. But judging by the absolute number of records (2516 in total), the discrepancy is “common”, which is remarkable.
If the publication stage of an article is inconsistent in different databases, it may bring about multiple problems, including research performance standards, calculation of measurement indicators, and setting of retrieval strategies, etc.
Different research institutions and evaluation departments may have different standards for the certification of the publication time of articles (online access time or official publication time). Then which database to take as the criterion will become a problem, because different databases are likely to lead to different results. When an article is indexed by multiple databases, there is a certain risk of relying solely on the search results from a single database. Therefore, this point needs to be taken into consideration in the formulation of performance evaluation policies. It is recommended to verify multiple databases or refer to the journal website (if the journal website is updated in a timely manner), which may further increase time and price costs.
When counting the number of papers and their corresponding citations and other related indicates, the time data associated with the publication stage may become a cause of result deviation, especially when the online access time and official publication time are not in the same year. Correspondingly, this will also affect the setting of retrieval strategies, and the impact of different time search fields on results should be considered.
5.2 Research question 2 – Differences in publication stage modes of different EA articles from the same journal
Firstly, from the perspective of sub-data sets. For the same journal, it is also necessary to start from the sub-data set to analyze whether it contains different publication stage modes. According to the journal distribution results of seven data sets (see Table 4), the ranking of the number of journals was as follows: D7 (WoS CC, Scopus, and EI) =D6 (Scopus and EI) ˃ D1 (WoS CC) ˃ D2 (Scopus) ˃ D4 (WoS CC and Scopus) ˃ D5 (WoS CC and EI) ˃ D3 (EI).
Secondly, from the perspective of publication stage modes. According to Table 4, regarding the “undifferentiated publication stage modes” in a single database, the journal distribution numbers of PS1 (PSW = EA), PS2 (PSW = OP), and PS3 (PSS = EA) is relatively large, which are 20, 39 and 32 respectively. In the two databases, the journal distribution numbers of PS15 (PSS = PSE = EA), PS16 (PSS = PSE = OP) and PS8 (PSW = PSS = OP) reached 44, 39 and 12 respectively, and the journal numbers of other modes were less than or equal to 5. In the three databases, the number of journals in PS19 (PSW = PSS = PSE = EA) and PS20 (PSW = PSS = PSE = OP) was 40 and 47 respectively. Regarding the “differentiated publication stage modes”, in the two databases, PS9 (PSW = EA, PSS = OP) and PS14 (PSW = OP, PSE = EA) existed in only two journals, with only one record per journal. P17 (PSS = EA, PSE = OP) and P18 (PSS = OP, PSE = EA) appeared in only one journal, with 20 and 29 records respectively. PS24 (PSW = PSS = OP, PSE = EA) existed in only 2 journals, with 186 and 1 record, respectively. In the three databases, PS21 (PSW = PSS = EA,PSE = OP), PS22 (PSW = PSE = EA, PSS = OP), and PS23 (PSS = PSE = EA, PSW = OP) had the EA stage in two databases and the OP stage in one database. PS21 and PS22 appeared in only one journal, while PS23 appeared in 16 journals. PS24 (PSW = PSS = OP, PSE = EA), PS25 (PSW = PSE = OP, PSS = EA), PS26 (PSS = PSE = OP, PSW = EA) had the OP stage in two databases and the EA stage in one database. PS24 appeared in only 2 journals, while P25 and P26 involved a larger number of journals, 20 and 33 respectively. Looking further at Fig 2, different publication stage modes showed their own characteristics in the record proportion and journal distribution, and it seemed to be no rule to follow.
Fig 2. The record proportion of publication stage modes (PS1-PS26).
Finally, according to the summary results in Fig 3, among 47 IEEE journals, each journal had 2 or more different sub-data sets, of which 24 journals (accounting for 51.06%) had 4 ones at the same time, 11 journals had 5 ones, 8 journals had 3 ones, 2 journals had 2 ones, and one journal even had 7 ones at the same time. In addition, each of the 47 journals contained at least 4 publication stage modes simultaneously, of which 13 journals contained 8 modes, 10 journals contained 7 modes, 7 journals contained 5 modes, 4 journals contained 9 modes, 4 journals contained 10 modes, and 3 journals contained 6 modes, 3 journals contained 11 modes, and one journal even contained 18 modes simultaneously. Among the 47 journals, except for 7 journals that did not had “undifferentiated publication stage mode”, the remaining 40 journals (85.11%) all had one or more differentiated modes, among which 19 journals had one differentiated mode, 10 journals had 2 differentiated modes, and 9 journals had 3 differentiated modes, one journal had 4 differentiated models, and one journal had 7 differentiated models.
Fig 3. Relationship between numbers of journal with numbers of sub-data sets, publication stage modes and differentiated publication stage modes.
Based on the above results, it could be concluded that the publication stage modes of different EA articles in the same IEEE journal were different, and the publication stage modes contained in the same IEEE journal were diverse and complex. At the same time, this phenomenon not only appeared in a single journal, with a certain universality, but also seemed to have no special rules to follow, with uncertainty.
This result undoubtedly indicates that it is difficult to examine a certain journal using a fixed publication stage model. To comprehensively understand the inclusion and publication status of articles of a certain journal in different databases, specific analysis is required on a case-by-case basis. Generally, databases index bases on the original data provided by journals. However, the diversity presented in results of this study would remind databases and publishers to think about and improve the indexing process and rules, because both may lead to such results due to their own reasons. Specifically, possible reasons include different original data provided by publishers to databases, different indexing methods and efficiencies of databases, errors in the original data, and errors of indexing, etc. In addition, it was found in the results that some differentiated modes only occurred in one or two journals. If this is an accidental situation, perhaps it can be effectively avoided through regular checks by the database and journal publishers.
6 Conclusion and limitation
This exploratory study revealed that in the three typical bibliographic databases (WoS CC, Scopus and EI), the publication stage of the same EA article in IEEE journals was different, and the publication stage mode of different EA articles in one IEEE journal was also different. In IEEE journals, these differences were universal, complex and uncertain. It is hoped that the results of this study will draw people's attention to the differences in the EA articles indexing of different bibliographic databases. Therefore, it is unreliable to use a single database to determine the publication stage and the corresponding publication time information of a certain article or a certain journal. It is hoped that results of this study can draw the attention of various stakeholders, including researchers, research institutions and evaluation departments, as well as databases and journal publishers, to the indexing differences of EA articles across different bibliographic databases. This paper had some limitations in the scope of the study. If the future study is carried out with a wider range of journals and more EA article samples, the comprehensiveness of the study results will be further improved.
Supporting information
(XLSX)
(XLSX)
(XLSX)
(XLSX)
(XLSX)
(XLSX)
Data Availability
All relevant data are within the paper and its Supporting Information files.
Funding Statement
General scientific research project of Zhejiang Education Department (Y202455786). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Dong P, Loh M, Mondry A. Publication lag in biomedical journals varies due to the periodical’s publishing model. Scientometrics. 2006;69(2):271–86. doi: 10.1007/s11192-006-0148-3 [DOI] [Google Scholar]
- 2.Das A, Das P. Delay between online and offline issue of journals: A critical analysis. Library & Information Science Research. 2006;28(3):453–9. doi: 10.1016/j.lisr.2006.03.019 [DOI] [Google Scholar]
- 3.Amat CB. Editorial and publication delay of papers submitted to 14 selected Food Research journals. Influence of online posting. Scientometrics. 2007;74(3):379–89. doi: 10.1007/s11192-007-1823-8 [DOI] [Google Scholar]
- 4.Maflahi N, Thelwall M. How quickly do publications get read? The evolution of mendeley reader counts for new articles. Asso for Info Science & Tech. 2017;69(1):158–67. doi: 10.1002/asi.23909 [DOI] [Google Scholar]
- 5.Asai S. Access Patterns of Electronic Articles: The Case Study of Scientometrics. Pub Res Q. 2022;38(2):295–307. doi: 10.1007/s12109-022-09887-0 [DOI] [Google Scholar]
- 6.Thompson JC, Quigley JM, Halfpenny NJA, Scott DA, Hawkins NS. Importance and methods of searching for E-publications ahead of print in systematic reviews. Evid Based Med. 2016;21(2):55–9. doi: 10.1136/ebmed-2015-110374 [DOI] [PubMed] [Google Scholar]
- 7.Nandita Q. The JCR Reload and a look ahead to the introduction of early access content in 2021. https://clarivate.com/blog/the-jcr-reload-and-a-look-ahead-to-the-introduction-of-early-access-content-in-2021/ [Google Scholar]
- 8.Liu W. A matter of time: publication dates in Web of Science Core Collection. Scientometrics. 2020;126(1):849–57. doi: 10.1007/s11192-020-03697-x [DOI] [Google Scholar]
- 9.Liu W, Wang P, Wang H. A matter of time: Publication dates in Scopus. Journal of Information Science. 2024. doi: 10.1177/01655515241293753 [DOI] [Google Scholar]
- 10.Franceschini F, Maisano D, Mastrogiacomo L. Empirical analysis and classification of database errors in Scopus and Web of Science. Journal of Informetrics. 2016;10(4):933–53. doi: 10.1016/j.joi.2016.07.003 [DOI] [Google Scholar]
- 11.Franceschini F, Maisano D, Mastrogiacomo L. The museum of errors/horrors in Scopus. Journal of Informetrics. 2016;10(1):174–82. doi: 10.1016/j.joi.2015.11.006 [DOI] [Google Scholar]
- 12.Liu W, Huang M, Wang H. Same journal but different numbers of published records indexed in Scopus and Web of Science Core Collection: causes, consequences, and solutions. Scientometrics. 2021;126(5):4541–50. doi: 10.1007/s11192-021-03934-x [DOI] [Google Scholar]
- 13.García-Zorita C, Martín-Moreno C, Lascurain-Sánchez ML, Sanz-Casado E. Institutional addresses in the Web of Science: the effects on scientific evaluation. Journal of Information Science. 2006;32(4):378–83. doi: 10.1177/0165551506065813 [DOI] [Google Scholar]
- 14.Falahati Qadimi Fumani MR, Goltaji M, Parto P. Inconsistent transliteration of Iranian university names: a hazard to Iran’s ranking in ISI Web of Science. Scientometrics. 2012;95(1):371–84. doi: 10.1007/s11192-012-0818-2 [DOI] [Google Scholar]
- 15.Schmidt CM, Cox R, Fial AV, Hartman TL, Magee ML. Gaps in affiliation indexing in Scopus and PubMed. J Med Libr Assoc. 2016;104(2):138–42. doi: 10.3163/1536-5050.104.2.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Liu W, Hu G, Tang L. Missing author address information in Web of Science—An explorative study. Journal of Informetrics. 2018;12(3):985–97. doi: 10.1016/j.joi.2018.07.008 [DOI] [Google Scholar]
- 17.Savchenko I, Kosyakov D. Lost in affiliation: apatride publications in international databases. Scientometrics. 2022;127(6):3471–87. doi: 10.1007/s11192-022-04392-9 [DOI] [Google Scholar]
- 18.Purnell PJ. The prevalence and impact of university affiliation discrepancies between four bibliographic databases—Scopus, Web of Science, Dimensions, and Microsoft Academic. Quantitative Science Studies. 2022;3(1):99–121. doi: 10.1162/qss_a_00175 [DOI] [Google Scholar]
- 19.Sirtes D. Funding acknowledgements for the German Research Foundation (DFG). The dirty data of the web of science database and how to clean it up[C]. 14th international society of scientometrics and informetrics conference (ISSI). 2013. p. 784–95. [Google Scholar]
- 20.Paul-Hus A, Desrochers N, Costas R. Characterization, description, and considerations for the use of funding acknowledgement data in Web of Science. Scientometrics. 2016;108(1):167–82. doi: 10.1007/s11192-016-1953-y [DOI] [Google Scholar]
- 21.Liu W, Tang L, Hu G. Funding information in Web of Science: an updated overview. Scientometrics. 2020;122(3):1509–24. doi: 10.1007/s11192-020-03362-3 [DOI] [Google Scholar]
- 22.Kokol P. Discrepancies among Scopus and Web of Science, coverage of funding information in medical journal articles: a follow-up study. J Med Libr Assoc. 2023;111(3):703–8. doi: 10.5195/jmla.2023.1513 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Franceschini F, Maisano D, Mastrogiacomo L. Errors in DOI indexing by bibliometric databases. Scientometrics. 2014;102(3):2181–6. doi: 10.1007/s11192-014-1503-4 [DOI] [Google Scholar]
- 24.Gorraiz J, Melero-Fuentes D, Gumpenberger C, Valderrama-Zurián J-C. Availability of digital object identifiers (DOIs) in Web of Science and Scopus. Journal of Informetrics. 2016;10(1):98–109. doi: 10.1016/j.joi.2015.11.008 [DOI] [Google Scholar]
- 25.Zhu J, Hu G, Liu W. DOI errors and possible solutions for Web of Science. Scientometrics. 2018;118(2):709–18. doi: 10.1007/s11192-018-2980-7 [DOI] [Google Scholar]
- 26.Krauskopf E, Salgado M. Inconsistency in the registration of the Digital Object Identifier (DOI) of articles on Web of Science and Scopus. IB. 2023;37(96):129–44. doi: 10.22201/iibi.24488321xe.2023.96.58784 [DOI] [Google Scholar]
- 27.Kawashima H, Tomizawa H. Accuracy evaluation of Scopus Author ID based on the largest funding database in Japan. Scientometrics. 2015;103(3):1061–71. doi: 10.1007/s11192-015-1580-z [DOI] [Google Scholar]
- 28.Aman V. Does the Scopus author ID suffice to track scientific international mobility? A case study based on Leibniz laureates. Scientometrics. 2018;117(2):705–20. doi: 10.1007/s11192-018-2895-3 [DOI] [Google Scholar]
- 29.Chinchilla-Rodríguez Z, Costas R, Robinson-García N, Larivière V. Examining the quality of the corresponding authorship field in Web of Science and Scopus. Quantitative Science Studies. 2024;5(1):76–97. doi: 10.1162/qss_a_00288 [DOI] [Google Scholar]
- 30.Donner P. Document type assignment accuracy in the journal citation index data of Web of Science. Scientometrics. 2017;113(1):219–36. doi: 10.1007/s11192-017-2483-y [DOI] [Google Scholar]
- 31.Kan Yeung AW. Document type assignment by Web of Science, Scopus, PubMed, and publishers to “Top 100” papers. MJLIS. 2021;26(3):97–103. doi: 10.22452/mjlis.vol26no3.5 [DOI] [Google Scholar]
- 32.Visser M, van Eck NJ, Waltman L. Large-scale comparison of bibliographic data sources: Scopus, Web of Science, Dimensions, Crossref, and Microsoft Academic. Quantitative Science Studies. 2021;2(1):20–41. doi: 10.1162/qss_a_00112 [DOI] [Google Scholar]
- 33.Mokhnacheva YuV. Document Types Indexed in WoS and Scopus: Similarities, Differences, and Their Significance in the Analysis of Publication Activity. Sci Tech Inf Proc. 2023;50(1):40–6. doi: 10.3103/s0147688223010033 [DOI] [Google Scholar]
- 34.Zhu M, Lu X, Chen F, Yang L, Shen Z. An explorative study on document type assignment of review articles in Web of Science, Scopus and journals’ websites. Journal of Data and Information Science. 2024;9(1):11–36. doi: 10.2478/jdis-2024-0003 [DOI] [Google Scholar]
- 35.Vera-Baceta M-A, Thelwall M, Kousha K. Web of Science and Scopus language coverage. Scientometrics. 2019;121(3):1803–13. doi: 10.1007/s11192-019-03264-z [DOI] [Google Scholar]
- 36.Tort ABL, Targino ZH, Amaral OB. Rising publication delays inflate journal impact factors. PLoS One. 2012;7(12):e53374. doi: 10.1371/journal.pone.0053374 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Chen H, Chen CH, Jhanji V. Publication times, impact factors, and advance online publication in ophthalmology journals. Ophthalmology. 2013;120(8):1697–701. doi: 10.1016/j.ophtha.2013.01.044 [DOI] [PubMed] [Google Scholar]
- 38.Heneberg P. Effects of print publication lag in dual format journals on scientometric indicators. PLoS One. 2013;8(4):e59877. doi: 10.1371/journal.pone.0059877 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Alves-Silva E, Porto ACF, Firmino C, Silva HV, Becker I, Resende L, et al. Are the impact factor and other variables related to publishing time in ecology journals?. Scientometrics. 2016;108(3):1445–53. doi: 10.1007/s11192-016-2040-0 [DOI] [Google Scholar]
- 40.Al U, Soydal I. Publication lag and early view effects in information science journals. AJIM. 2017;69(2):118–30. doi: 10.1108/ajim-12-2016-0200 [DOI] [Google Scholar]
- 41.González-Betancor SM, Dorta-González P. Publication modalities ‘article in press’ and ‘open access’ in relation to journal average citation. Scientometrics. 2019;120(3):1209–23. doi: 10.1007/s11192-019-03156-2 [DOI] [Google Scholar]
- 42.Lin WYC. Effects of open access and articles-in-press mechanisms on publishing lag and first-citation speed: a case on energy and fuels journals. Scientometrics. 2021;126(6):4841–69. [Google Scholar]
- 43.Haustein S, Bowman TD, Costas R. When is an article actually published? An analysis of online availability, publication, and indexation dates. arXiv (Cornell University); 2015. doi: 10.48550/arXiv.1505.00796 [DOI] [Google Scholar]
- 44.Echeverría M, Stuart D, Cordón-García JA. Influencia de la fecha de publicación online de los artículos científicos en los indicadores bibliométricos. Rev esp doc Cient. 2017;40(3):e183. doi: 10.3989/redc.2017.3.1422 [DOI] [Google Scholar]
- 45.Hu Z, Tian W, Xu S, Zhang C, Wang X. Four pitfalls in normalizing citation indicators: An investigation of ESI’s selection of highly cited papers. Journal of Informetrics. 2018;12(4):1133–45. doi: 10.1016/j.joi.2018.09.006 [DOI] [Google Scholar]
- 46.Jacso P. As we may search - Comparison of major features of the Web of Science, Scopus, and Google Scholar citation-based and citation-enhanced databases. Current Science. 2005;89(9):1537–47. [Google Scholar]
- 47.Bakkalbasi N, Bauer K, Glover J, Wang L. Three options for citation tracking: Google Scholar, Scopus and Web of Science. Biomed Digit Libr. 2006;3:7. doi: 10.1186/1742-5581-3-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Meho LI, Yang K. Impact of data sources on citation counts and rankings of LIS faculty: Web of science versus scopus and google scholar. J Am Soc Inf Sci. 2007;58(13):2105–25. doi: 10.1002/asi.20677 [DOI] [Google Scholar]
- 49.Norris M, Oppenheim C. Comparing alternatives to the Web of Science for coverage of the social sciences’ literature. Journal of Informetrics. 2007;1(2):161–9. doi: 10.1016/j.joi.2006.12.001 [DOI] [Google Scholar]
- 50.Falagas ME, Pitsouni EI, Malietzis GA, Pappas G. Comparison of PubMed, Scopus, Web of Science, and Google Scholar: strengths and weaknesses. FASEB J. 2008;22(2):338–42. doi: 10.1096/fj.07-9492LSF [DOI] [PubMed] [Google Scholar]
- 51.Gavel Y, Iselid L. Web of Science and Scopus: a journal title overlap study. Online Information Review. 2008;32(1):8–21. doi: 10.1108/14684520810865958 [DOI] [Google Scholar]
- 52.Gorraiz J, Schloegl C. A bibliometric analysis of pharmacology and pharmacy journals: Scopus versus Web of Science. Journal of Information Science. 2008;34(5):715–25. doi: 10.1177/0165551507086991 [DOI] [Google Scholar]
- 53.López-Illescas C, de Moya-Anegón F, Moed HF. Coverage and citation impact of oncological journals in the Web of Science and Scopus. Journal of Informetrics. 2008;2(4):304–16. doi: 10.1016/j.joi.2008.08.001 [DOI] [Google Scholar]
- 54.Meho LI, Rogers Y. Citation counting, citation ranking, and h‐index of human‐computer interaction researchers: A comparison of Scopus and Web of Science. J Am Soc Inf Sci. 2008;59(11):1711–26. doi: 10.1002/asi.20874 [DOI] [Google Scholar]
- 55.Archambault É, Campbell D, Gingras Y, Larivière V. Comparing bibliometric statistics obtained from the Web of Science and Scopus. J Am Soc Inf Sci. 2009;60(7):1320–6. doi: 10.1002/asi.21062 [DOI] [Google Scholar]
- 56.Jacsó P. The h‐index for countries in Web of Science and Scopus. Online Information Review. 2009;33(4):831–7. doi: 10.1108/14684520910985756 [DOI] [Google Scholar]
- 57.Jacsó P. Database source coverage: hypes, vital signs and reality checks. Online Information Review. 2009;33(5):997–1007. doi: 10.1108/14684520911001963 [DOI] [Google Scholar]
- 58.Levine‐Clark M, Gil E. A comparative analysis of social sciences citation tools. Online Information Review. 2009;33(5):986–96. doi: 10.1108/14684520911001954 [DOI] [Google Scholar]
- 59.Vieira ES, Gomes JANF. A comparison of Scopus and Web of Science for a typical university. Scientometrics. 2009;81(2):587–600. doi: 10.1007/s11192-009-2178-0 [DOI] [Google Scholar]
- 60.Li J, Burnham JF, Lemley T, Britton RM. Citation Analysis: Comparison of Web of Science®, Scopus™, SciFinder®, and Google Scholar. Journal of Electronic Resources in Medical Libraries. 2010;7(3):196–217. doi: 10.1080/15424065.2010.505518 [DOI] [Google Scholar]
- 61.Adriaanse LS, Rensleigh C. Comparing Web of Science, Scopus and Google Scholar from an Environmental Sciences perspective. South African Journal of Libraries and Information Science. 2011;77(2). doi: 10.7553/77-2-58 [DOI] [Google Scholar]
- 62.Lasda Bergman EM. Finding Citations to Social Work Literature: The Relative Benefits of Using Web of Science, Scopus, or Google Scholar. The Journal of Academic Librarianship. 2012;38(6):370–9. doi: 10.1016/j.acalib.2012.08.002 [DOI] [Google Scholar]
- 63.Markpin T, Premkamolnetr N, Ittiritmeechai S, Wongkaew C, Yochai W, Ratchatahirun P, et al. The effects of choice of database and data retrieval methods on research performance evaluations of Asian universities. Online Information Review. 2013;37(4):538–63. [Google Scholar]
- 64.Bartol T, Budimir G, Dekleva-Smrekar D, Pusnik M, Juznic P. Assessment of research fields in Scopus and Web of Science in the view of national research evaluation in Slovenia. Scientometrics. 2013;98(2):1491–504. doi: 10.1007/s11192-013-1148-8 [DOI] [Google Scholar]
- 65.Hilbert F, Barth J, Gremm J, Gros D, Haiter J, Henkel M, et al. Coverage of academic citation databases compared with coverage of scientific social media. Online Information Review. 2015;39(2):255–64. doi: 10.1108/oir-07-2014-0159 [DOI] [Google Scholar]
- 66.Harzing A-W, Alakangas S. Google Scholar, Scopus and the Web of Science: a longitudinal and cross-disciplinary comparison. Scientometrics. 2015;106(2):787–804. doi: 10.1007/s11192-015-1798-9 [DOI] [Google Scholar]
- 67.Dorsch I. Relative visibility of authors’ publications in different information services. Scientometrics. 2017;112(2):917–25. doi: 10.1007/s11192-017-2416-9 [DOI] [Google Scholar]
- 68.Martín-Martín A, Orduna-Malea E, Thelwall M, Delgado López-Cózar E. Google Scholar, Web of Science, and Scopus: A systematic comparison of citations in 252 subject categories. Journal of Informetrics. 2018;12(4):1160–77. doi: 10.1016/j.joi.2018.09.002 [DOI] [Google Scholar]
- 69.Pech G, Delgado C. Percentile and stochastic-based approach to the comparison of the number of citations of articles indexed in different bibliographic databases. Scientometrics. 2020;123(1):223–52. doi: 10.1007/s11192-020-03386-9 [DOI] [Google Scholar]
- 70.Kumar A. Comparing scientific productivity using Scopus and Web of Science (WoS): a case of Indian R&D laboratories. Asian Journal of Technology Innovation. 2020;29(3):414–26. doi: 10.1080/19761597.2020.1816837 [DOI] [Google Scholar]
- 71.Levine-Clark M, Gil EL. A new comparative citation analysis: Google Scholar, Microsoft Academic, Scopus, and Web of Science. Journal of Business & Finance Librarianship. 2021;26(1–2):145–63. doi: 10.1080/08963568.2021.1916724 [DOI] [Google Scholar]
- 72.Martín-Martín A, Thelwall M, Orduna-Malea E, Delgado López-Cózar E. Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations’ COCI: a multidisciplinary comparison of coverage via citations. Scientometrics. 2021;126(1):871–906. doi: 10.1007/s11192-020-03690-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Pranckutė R. Web of Science (WoS) and Scopus: The Titans of Bibliographic Information in Today’s Academic World. Publications. 2021;9(1):12. doi: 10.3390/publications9010012 [DOI] [Google Scholar]
- 74.Singh VK, Singh P, Karmakar M, Leta J, Mayr P. The journal coverage of Web of Science, Scopus and Dimensions: A comparative analysis. Scientometrics. 2021;126(6):5113–42. doi: 10.1007/s11192-021-03948-5 [DOI] [Google Scholar]
- 75.Gusenbauer M. Search where you will find most: Comparing the disciplinary coverage of 56 bibliographic databases. Scientometrics. 2022;127(5):2683–745. doi: 10.1007/s11192-022-04289-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 76.Asubiaro TV. Sub-Saharan Africa’s biomedical journal coverage in scholarly databases: a comparison of Web of Science, Scopus, EMBASE, MEDLINE, African Index Medicus, and African Journals Online. J Med Libr Assoc. 2023;111(3):696–702. doi: 10.5195/jmla.2023.1448 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 77.Asubiaro TV, Onaolapo S. A comparative study of the coverage of African journals in Web of Science, Scopus, and CrossRef. Asso for Info Science & Tech. 2023;74(7):745–58. doi: 10.1002/asi.24758 [DOI] [Google Scholar]
- 78.Singh P, Singh VK, Piryani R. Scholarly article retrieval from Web of Science, Scopus and Dimensions: A comparative analysis of retrieval quality. Journal of Information Science. 2023. doi: 10.1177/01655515231191351 [DOI] [Google Scholar]
- 79.Asubiaro T, Onaolapo S, Mills D. Regional disparities in Web of Science and Scopus journal coverage. Scientometrics. 2024;129(3):1469–91. doi: 10.1007/s11192-024-04948-x [DOI] [Google Scholar]
- 80.Gerasimov I, KC B, Mehrabian A, Acker J, McGuire MP. Comparison of datasets citation coverage in Google Scholar, Web of Science, Scopus, Crossref, and DataCite. Scientometrics. 2024;129(7):3681–704. doi: 10.1007/s11192-024-05073-5 [DOI] [Google Scholar]
- 81.Moustafa K. Variations in Citations Across Databases: Implications for Journal Impact Factors. Semin Ophthalmol. 2024;39(5):400–3. doi: 10.1080/08820538.2024.2322428 [DOI] [PubMed] [Google Scholar]
- 82.Ortega JL, Delgado-Quirós L. The indexation of retracted literature in seven principal scholarly databases: a coverage comparison of dimensions, OpenAlex, PubMed, Scilit, Scopus, The Lens and Web of Science. Scientometrics. 2024;129(7):3769–85. doi: 10.1007/s11192-024-05034-y [DOI] [Google Scholar]
- 83.Bartol T, Hocevar M. The capital cities of the ten new European Union countries in selected bibliographic databases. Scientometrics. 2005;65(2):173–87. doi: 10.1007/s11192-005-0266-3 [DOI] [Google Scholar]
- 84.Kostoff RN. Comparison of China/USA science and technology performance. Journal of Informetrics. 2008;2(4):354–63. doi: 10.1016/j.joi.2008.06.004 [DOI] [Google Scholar]
- 85.Meier JJ, Conkling TW. Google Scholar’s Coverage of the Engineering Literature: An Empirical Study. The Journal of Academic Librarianship. 2008;34(3):196–201. doi: 10.1016/j.acalib.2008.03.002 [DOI] [Google Scholar]
- 86.Cusker J. Elsevier Compendex and Google Scholar: A Quantitative Comparison of Two Resources for Engineering Research and an Update to Prior Comparisons. The Journal of Academic Librarianship. 2013;39(3):241–3. doi: 10.1016/j.acalib.2013.02.001 [DOI] [Google Scholar]
- 87.Cole C, Davis AR, Eyer V, Meier JJ. Google Scholar’s Coverage of the Engineering Literature 10 years Later. The Journal of Academic Librarianship. 2018;44(3):419–25. doi: 10.1016/j.acalib.2018.02.013 [DOI] [Google Scholar]
- 88.McAllister J, Torres J. Discovering Open Access Trends in Engineering: A Bibliometric Analysis of Open Access Publications at a Large Research University. Science & Technology Libraries. 2022;42(1):50–67. doi: 10.1080/0194262x.2022.2032535 [DOI] [Google Scholar]
- 89.Zhu Y. A study of the index time of early access articles. Scientometrics. 2024;130(1):159–86. doi: 10.1007/s11192-024-05188-9 [DOI] [Google Scholar]
- 90.Web of Science Core Collection: Early Access articles 2022. [https://support.clarivate.com/ScientificandAcademicResearch/s/article/Web-of-Science-Core-Collection-Early-Access-articles?language=en_US] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
(XLSX)
(XLSX)
(XLSX)
(XLSX)
(XLSX)
(XLSX)
Data Availability Statement
All relevant data are within the paper and its Supporting Information files.