Contributorship statements were introduced by scholarly journals in the late 1990s to provide more details on the specific contributions made by authors to research papers. After more than a decade of idiosyncratic taxonomies by journals, a partnership between medical journals and standards organizations has led to the establishment, in 2015, of the Contributor Roles Taxonomy (CRediT), which provides a standardized set of 14 research contributions. Using the data from Public Library of Science (PLOS) journals over the 2017–2018 period (N = 30,054 papers), this paper analyzes how research contributions are divided across research teams, focusing on the association between division of labor and number of authors, and authors’ position and specific contributions. It also assesses whether some contributions are more likely to be performed in conjunction with others and examines how the new taxonomy provides greater insight into the gendered nature of labor division. The paper concludes with a discussion of results with respect to current issues in research evaluation, science policy, and responsible research practices.

Scientific authorship is regularly considered as the primary currency in academia, whether for hiring, promotions or priority disputes (Biagioli & Galison, 2003; Cronin, 2001; Pontille, 2004). Yet, from the 1950 onwards, issues have been progressively raised about the use of authorship for attributing scientific capital (Bourdieu, 2001). These issues can be grouped into three categories. The first one relates to the increasing number of authors per article (Larivière, Sugimoto et al., 2015; Zuckerman, 1968). In some domains, such as clinical research, genomics, and high-energy physics—where articles often bear several hundreds or thousands of names in the byline—identifying respective contributions and, thus, assessing individual researchers’ contributions, is increasingly difficult. Second, with the rise of multidisciplinary projects, the meanings attributed to authorship—and to name ordering—have multiplied, with unintended consequences for authorship (Paul-Hus, Mongeon et al., 2017; Smith, Williams-Jones et al., 2020a, 2020b). The frictions of conventions have sown discord among the participants in research projects (Wilcox, 1998) and greater confusion has also prevailed among gatekeepers (Bhandari, Guyatt et al., 2014). Third, scientific research has regularly—and some may argue increasingly (Azoulay, Furman et al., 2015)—been shaken by cases of fraud. In some alleged cases, all authors on a work under investigation have asked journals to remove their names from publications. Such a systematic denial of responsibility has gone to the point that certain articles have found themselves “orphaned” (Rennie & Flanagin, 1994).

Considering these different aspects as undermining factors in the fair attribution of scientific contributions, researchers, journal editors, research administrators, and members of funding bodies have been looking for alternative ways to assign authorship. This issue has been of particular concern in the biomedical sciences, given the immediate public concerns that occur when research lacks transparency. Over the last few decades, discussions and debates have taken place in major journals, several workshops have been organized, and an “authorship task force” group was formed to imagine better ways of attributing credit for scholarly publications (Davidoff, 2000). This collective exploration has resulted in at least two concomitant phenomena: the development of more precise vocabulary for authorship malpractice and the development of new authorship attribution devices.

A new vocabulary has progressively emerged to characterize controversial authorship practices (Pontille, 2016; Sismondo, 2009). Omission of a researcher who contributed significantly to the project—ghost authorship—is one of the most frequent of transgressions and also one of the most difficult to count (given that omission is often of junior scholars or more technical contributors who may lack capital in science). This is particularly problematic in some disciplines; for example, a survey of biomedicine suggests that about one-fifth of all papers exhibit ghost authorship (Wislar, Flanagin et al., 2011). Ghost authorship is also fairly common in industry-initiated trials, where most ghost authors are statisticians (Gøtzsche, Hróbjartsson et al., 2007). This may be less malicious than other forms of authorship misconduct and more of a reflection of differing forms of capital exchange between industry and academe. However, the more pernicious relative of ghost authorship is ghost management of research by pharmaceutical companies (Sismondo & Doucet, 2010). These practices demonstrate the flip side of authorship: Where ghost authorship calls attention to the lack of rewards for the author, ghost management highlights the issues that arise when there is no transparency in accountability. Honorary authorship falls on the other side of the coin: providing reward where there was no labor. Two forms of this have been identified: guest and gift authorship. Guest authorship designates already recognized names that stand as a sign of quality and potentially increase the chances for the article to be published (Haeussler & Sauermann, 2013). Gift authorship sets up a principle of reciprocal exchange between colleagues, resulting in the inclusion of people as authors regardless of their actual contribution (Smith, 1994; Street, Rogers et al., 2010). Levels of honorary authorship on scholarly papers have been reported as between 20% and 40% (Flanagin, Carey et al., 1998; Hardjosantoso, Dahi et al., 2020; Mowatt, Shirran et al., 2002).

To mitigate instances of misconduct, new attribution devices have been proposed. For instance, Richard Horton, editor in chief of the Lancet, suggested that the relationship between journal editors and researchers be conceived as a legal contract, each of the parties being held up to mutual engagements (Horton, 1997). The proposal that received the most attention, however, was the systematic description, in scholarly articles, of each author’s contribution (Rennie, Yank, & Emanuel, 1997). This approach allows both readers and editors to identify precisely which work was done by individual researchers. Explicitly based on suggestions made during the previous decade (Moulopoulos, Sideris, & Georgilis, 1983; Saffran, 1989), the concept of “contributorship” was aimed at better distinguishing credit and responsibility, two interrelated features of authorship (Birnholtz, 2006). Such contributorship statements were the focus of experiments before they were finally introduced in the “instructions to authors” of several biomedical journals (Northridge, 1998; Rennie, Flanagin, & Yank, 2000; Smith, 1997) and the recommendations of regulatory authorities, such as the International Committee of Medical Journal Editors (ICMJE) and Committee on Publication Ethics (COPE).

Linked to a conception of research activity heavily influenced by accountability, these contributorship statements allowed for both finer recognition of and responsibility for the specific tasks performed, but also assumed that the research process can be segmented into different acts that can be properly ascribed to individual contributors. The segmentation of scientific contributions was not introduced by contributorship, but rather emerged from researchers who have proposed taxonomies in response to Moulopoulos et al.’s (1983) work. These idiosyncratic taxonomies differed in the number of contributions listed (from six to 15) and their degree of accuracy. For example, “writing up the paper” was sometimes considered as one contribution, while in other taxonomies it was supplemented with “critical revision of manuscript,” or even split into “writing the first draft of the paper,” “writing later draft(s),” and “approving final draft” (Goodman, 1994).

Biomedical journals were the main drivers of new taxonomies. Two peculiarities have resulted from this. First, these taxonomies are characterized by research task contributions clearly specific both to the biomedical sciences (“collecting samples or specimens,” “providing DNA probes”) and clinical research (“referred patients to study,” “provision of study materials or patients”). Second, there are significant differences in not only the number of contributions from one journal to another but also the variations in contribution taxonomies and their organization (Bates, Anić et al., 2004; Baerlocher, Gautam et al., 2009; McDonald, Neff et al., 2010). Journals request contributions in free-text form, organized as a predefined list of research tasks to choose from, or even as hierarchical items that make some contribution roles a prerequisite for others. As these taxonomies evolve, studies have investigated the relationship between the structure of these forms, the number of contributions described, and the differences in perception among coauthors of the same article (Ilakovac, Fister et al., 2007; Ivaniš, Hren et al., 2008, 2011; Marušić, Bates et al., 2006).

Early taxonomies paved the way for large-scale empirical studies of authorship practices in science. For instance, Larivière, Desrochers et al. (2016) analyzed contributorship statements—divided into five contributions—for 87,002 papers published in all Public Library of Science (PLOS) journals, focusing on labor distributions across disciplines, authors’ order, and seniority. They showed that the division of scientific labor is higher in medical research than in natural sciences, and that in all domains but medicine, the most common task among authors was drafting and editing of the manuscript. Results of this and subsequent analyses (Macaluso, Larivière et al., 2016) also showed strong distinction between tasks performed and author characteristics: Younger researchers and women were more likely to perform technical contributions, whereas older, male researchers were more often associated with conceptual contributions. Authors’ order was also strongly associated with number of contributions: First authors were generally associated with the vast majority of contributions, followed by last authors—who generally were not involved in technical work—and then by middle authors, whose contributions were fewer and more likely to be technical (Larivière et al., 2016).

These findings were confirmed by Sauermann and Haeussler (2017), who analyzed more than 12,000 articles published between 2007 and 2011 in PLOS ONE. As with Larivière et al. (2016), they found that first and last authors were associated with more contributions than middle authors. In an examination of team size, they demonstrated that the number of contributions per author decreases with the number of authors, but remains stable for last authors. They complemented this analysis with a survey of 6,000 corresponding authors from these papers. Their findings suggest that a majority of corresponding authors believe that contributorship statements provided more information about the contribution, but only a minority think that contributorship provides more information on the importance of contributions. Furthermore, they found that in one-fifth of papers, contributorship statements were determined by the corresponding authors alone.

Sauermann and Haeussler (2017) suggested that it was difficult to predict the contribution based on author order alone. Corrêa, Silva et al. (2017)—also using the PLOS ONE data set—confirmed this uncertainty between authors’ order and contributions made. Using a network-based approach, they found that the relationship becomes increasingly random as the number of authors per paper increases. They also provided evidence of how division of labor increases as the number of authors increases and showed that contributions can be grouped into three categories: those who write, those who perform data analysis, and those who conduct experiments.

These studies provided novel insight on the relationship between authorship and one coarse-grained contributorship taxonomy. However, the previously used five contributorship categories fail to account for the complexity of contemporary science. To address the need for a more refined taxonomy, an International Workshop on Contributorship and Scholarly Attribution was organized at Harvard in May 2012 at the initiative of the Wellcome Trust (IWCSA, 2012). One outcome was a pilot project involving publishers, funders, and scientists to design a cross-disciplinary standardized taxonomy for contributor roles and contribution types, which would be practicable for all scientific fields. The goal was to be interoperable with different databases and to reduce the many ambiguities that remain with earlier contributorship typologies. In the eyes of its promoters, this standardized taxonomy would not only codify the contributions of each researcher with fine granularity, allowing for specific skills to be easily identified, but also rely on an infrastructure to manage the complex relationships between the information, its archiving, and its consultation in real time.

An initial prototype comprised of 14 types of contribution roles was designed and tested among corresponding authors of work published in various (mostly biomedical) journals (Allen, Scott et al., 2014). Based on the positive result of this experiment, a partnership with two information industry standards organizations (Consortia Advancing Standards in Research Administration Information (CASRAI) and the U.S.-based National Information Standards Organization (NISO)) was established to achieve broader consultation and to refine the preliminary taxonomy. An updated version of the taxonomy was made public in 2015 under the name CRediT (Contributor Roles Taxonomy) to provide “a controlled vocabulary of contributor roles” (Brand, Allen et al., 2015) for published research outputs.

The introduction of CRediT provides more details on the division of scientific labor than was given with previous contributorship taxonomies. First, not only may a given role be assigned to multiple contributors, but when this is the case, a degree of contribution may optionally be specified as “lead,” “equal,” or “supporting.”1 The granularity of contribution roles is thus more precise and the same contribution role can be prioritized among contributors. Second, the 14 contribution roles go beyond the commonly identified research tasks in traditional authorship. They notably include various roles related to research data, such as “resources” (provision of study materials, reagents, materials, patients, laboratory samples, animals, etc.), “data curation” (annotation, scrubbing, and maintenance), “software” (programming, software development, designing computer programs, etc.), or “visualization” (preparation, creation and/or presentation of the published work, specifically visualization/ data presentation). Third, the writing process is divided into two main roles, “original draft” and “review and editing,” introducing nuance in this primary contributorship role. With these improvements, CRediT is suited to account for both the division of scientific labor and the allocation of individual contributions.

PLOS adopted CRediT in 2016 (Atkins, 2016). By the end of 2018, more than 30,000 articles had employed this new taxonomy. In this paper, we provide an examination of these articles to investigate whether the more fine-grained analysis provides a more nuanced portrait of division of labor than was possible with previous taxonomies. More specifically, we examine how research contributions are divided across research teams, focusing on the association between number of authors and division of labor, and on the relationship between authors’ position and specific tasks performed. We also consider the association between each of the 14 contributions, to assess whether some contributions are more likely to be performed in conjunction with others.

In their review of the taxonomy, Allen, O’Connell, and Kiermer (2019) identify how CRediT can be a useful tool in science of science. As they state: “If we can understand how collaborations work and when, or how to optimize the best team mix, then we may be able to incentivize the sorts of behaviours and activities that can bring about and accelerate discovery” (p. 74). They particularly draw attention to the issues of diversity in team composition and how contributorship studies can provide insights into how to best support women and early career researchers as they progress in science. Therefore, we also explore how the new taxonomy provides greater insight into the gendered nature of science, comparing this with the earlier PLOS typology (Macaluso et al., 2016).

Launched in 2014, CRediT categorizes contributions made to scholarly papers into 14 categories (Table 1). Several journals—such as eLife, Cell, F1000—and publishers—PLOS, Elsevier, Springer, BMJ—have adopted it or, in the case of major publishers, have seen some of their journals adopt it. By early 2019, more than 120 journals had implemented the taxonomy (Allen et al., 2019), a number that increased substantially at the end of 2019 with the adoption of the typology by 1,200 journals from Elsevier (Elsevier, 2019). Our analysis is based on one of these publishers—PLOS—which provided us with all of its contributorship information for papers published between June 15, 2017 and December 31, 2018 (N = 30,770). The data covered all PLOS journals and included publication date, Digital Object Identifier (DOI), journal name, author name as it appears on the paper, and associated CRediT contributions for each author3.

Table 1.

Definition of each contribution found in the Contributor Roles Taxonomy (CRediT)2

ContributionDefinition
Conceptualization Ideas; formulation or evolution of overarching research goals and aims. 
Data curation Management activities to annotate (produce metadata), scrub data and maintain research data (including software code, where it is necessary for interpreting the data itself) for initial use and later reuse. 
Formal analysis Application of statistical, mathematical, computational, or other formal techniques to analyze or synthesize study data. 
Funding acquisition Acquisition of the financial support for the project leading to this publication. 
Investigation Conducting a research and investigation process, specifically performing the experiments, or data/evidence collection. 
Methodology Development or design of methodology; creation of models. 
Project administration Management and coordination responsibility for the research activity planning and execution. 
Resources Provision of study materials, reagents, materials, patients, laboratory samples, animals, instrumentation, computing resources, or other analysis tools. 
Software Programming, software development; designing computer programs; implementation of the computer code and supporting algorithms; testing of existing code components. 
Supervision Oversight and leadership responsibility for the research activity planning and execution, including mentorship external to the core team. 
Validation Verification, whether as a part of the activity or separate, of the overall replication/reproducibility of results/experiments and other research outputs. 
Visualization Preparation, creation and/or presentation of the published work, specifically visualization/data presentation. 
Writing—original draft Preparation, creation and/or presentation of the published work, specifically writing the initial draft (including substantive translation). 
Writing—review & editing Preparation, creation and/or presentation of the published work by those from the original research group, specifically critical review, commentary or revision—including pre- or postpublication stages. 
ContributionDefinition
Conceptualization Ideas; formulation or evolution of overarching research goals and aims. 
Data curation Management activities to annotate (produce metadata), scrub data and maintain research data (including software code, where it is necessary for interpreting the data itself) for initial use and later reuse. 
Formal analysis Application of statistical, mathematical, computational, or other formal techniques to analyze or synthesize study data. 
Funding acquisition Acquisition of the financial support for the project leading to this publication. 
Investigation Conducting a research and investigation process, specifically performing the experiments, or data/evidence collection. 
Methodology Development or design of methodology; creation of models. 
Project administration Management and coordination responsibility for the research activity planning and execution. 
Resources Provision of study materials, reagents, materials, patients, laboratory samples, animals, instrumentation, computing resources, or other analysis tools. 
Software Programming, software development; designing computer programs; implementation of the computer code and supporting algorithms; testing of existing code components. 
Supervision Oversight and leadership responsibility for the research activity planning and execution, including mentorship external to the core team. 
Validation Verification, whether as a part of the activity or separate, of the overall replication/reproducibility of results/experiments and other research outputs. 
Visualization Preparation, creation and/or presentation of the published work, specifically visualization/data presentation. 
Writing—original draft Preparation, creation and/or presentation of the published work, specifically writing the initial draft (including substantive translation). 
Writing—review & editing Preparation, creation and/or presentation of the published work by those from the original research group, specifically critical review, commentary or revision—including pre- or postpublication stages. 

Table 2 presents the characteristics of the data set. The bulk of the papers were published in the megajournal PLOS ONE (87.9%), which is the second largest megajournal (Siler, Larivière, & Sugimoto, 2020). Our data set contains comprehensive data for all journals with the exception of PLOS Biology, for which contributorship information could only be obtained for 13 papers4. Important differences are observed in terms of mean number of authors per paper, with PLOS Computational Biology having, on average, slightly less than five authors per paper, while PLOS Medicine has almost three times the rates of PLOS Computational Biology. However, the mean numbers of contributions per paper are quite constant across journals, with a maximum of 11.8 in PLOS Biology and a minimum of 10.6 in PLOS ONE. Given the strong focus on medical sciences of the multidisciplinary journal PLOS ONE (Siler et al., 2020) and of other PLOS journals, the results need to be interpreted as illustrative of the use of the CRediT taxonomy in those disciplines.

Table 2.

Number of papers published with CRediT contributions, mean number of authors and mean number of CRediT contributions per paper, by PLOS journal

JournalNo. papersNo. papers in WoS% papers in WoSMean No. authorsMean No. contributions
PLOS Biology 13 13 100 7.2 11.8 
PLOS Computational Biology 763 754 98.8 4.9 11.1 
PLOS Genetics 786 778 99.0 8.5 11.1 
PLOS Medicine 250 249 99.6 14.2 10.8 
PLOS Neglected Tropical Diseases 1,144 1,115 97.5 9.1 11.1 
PLOS ONE 27,057 26,398 97.6 6.8 10.6 
PLOS Pathogens 757 747 98.7 9.4 11.0 
All journals 30,770 30,054 97.7 7.0 10.6 
JournalNo. papersNo. papers in WoS% papers in WoSMean No. authorsMean No. contributions
PLOS Biology 13 13 100 7.2 11.8 
PLOS Computational Biology 763 754 98.8 4.9 11.1 
PLOS Genetics 786 778 99.0 8.5 11.1 
PLOS Medicine 250 249 99.6 14.2 10.8 
PLOS Neglected Tropical Diseases 1,144 1,115 97.5 9.1 11.1 
PLOS ONE 27,057 26,398 97.6 6.8 10.6 
PLOS Pathogens 757 747 98.7 9.4 11.0 
All journals 30,770 30,054 97.7 7.0 10.6 

Contribution information provided by PLOS did not, however, contain author order; to obtain this information we had to match each PLOS paper with its record in our in-house version of Clarivate Analytics’ Web of Science based on the DOI; this was feasible for 30,054 papers (97.7% of the PLOS data set; see Table 2 for percentages by journal), which included 222,938 authorships. Once the papers were matched with the WoS, we matched each author in both data sources to obtain their individual order in the authors’ list. This was first based on a perfect match of the full name string (e.g., Derek John de Solla Price = Derek John de Solla Price). However, as several names could not be matched because they were written in different manners in both databases (e.g., Derek de Solla Price, Derek J. Price, Derek Price), we performed additional matching focusing on specific parts of the name string. More specifically, we iteratively focused on the first and last 2–5 characters of the names; this allowed us to match 221,637 authorships (99.4% of the sample).

For this subset of authors who could be attributed an author order, we assigned a gender based on their given names. Such gender assignation of researchers has become a relatively standard practice and was shown to obtain relatively high precision and recall (Karimi, Wagner et al., 2016; Santamaría & Mihaljević, 2018). In this paper, we used the algorithm developed in Larivière, Ni, et al. (2013), which was created using several country-level lists of given names along with their gender. The algorithm has been tested for precision, and was found to be 98.3% precise for men and 86.7% for women (see the supplementary material in Larivière et al. (2013) for more details). The algorithm assigned a gender to 82.2% of the authorships covered in this analysis (Table 3). This percentage varies by author order, however, with a higher proportion of last authors assigned a gender, and a lower proportion of first authors. The percentage of female authorships in the PLOS data set represents 39.9% of authorships to which a gender could be assigned, which is slightly greater than the percentage of female authorships found in the WoS for disciplines of the medical sciences (about 35%).

Table 3.

Number of authorships with gender assigned, by author order

GenderFirstMiddleLastAny order
N%N%N%N%
Gender assigned 26,005 79.9 129,198 82.3 27,064 84.4 182,267 82.2 
 Female 12,094 37.2 52,106 33.2 8,600 26.8 72,800 32.8 
 Male 13,911 42.7 77,092 49.1 18,464 57.6 109,467 49.4 
Initials 382 1.2 2,085 1.3 447 1.4 2,914 1.3 
Unisex 704 2.2 3,899 2.5 911 2.8 5,514 2.5 
Unknown 5,462 16.8 21,847 13.9 3,633 11.3 30,942 14.0 
Total 32,553 100.0 157,029 100.0 32,055 100.0 221,637 100.0 
GenderFirstMiddleLastAny order
N%N%N%N%
Gender assigned 26,005 79.9 129,198 82.3 27,064 84.4 182,267 82.2 
 Female 12,094 37.2 52,106 33.2 8,600 26.8 72,800 32.8 
 Male 13,911 42.7 77,092 49.1 18,464 57.6 109,467 49.4 
Initials 382 1.2 2,085 1.3 447 1.4 2,914 1.3 
Unisex 704 2.2 3,899 2.5 911 2.8 5,514 2.5 
Unknown 5,462 16.8 21,847 13.9 3,633 11.3 30,942 14.0 
Total 32,553 100.0 157,029 100.0 32,055 100.0 221,637 100.0 

Figure 1 presents, for each PLOS journal, the percentage of papers on which each contribution appears. This provides an indication of importance of each task across the spectrum of PLOS journals and, conversely, of the tasks that are not performed by any of the authors on a given paper. Nearly all papers had an author writing the original draft (99%), as well as authors reviewing and editing (96%) and conceptualizing (95%) them. This suggests that these remain essential research acts—all papers are conceptualized and written. The percentages of papers with at least one author contributing to formal analysis (91%), methodology (90%), and investigation (86%) are also very high, suggesting that empirical papers are the bulk of those published in these journals. The supervision task is contained in 84% of papers; the 16% of papers without such a task likely do not include trainees as coauthors. Data curation is present in 79% of papers—although this percentage is higher in journals like PLOS Medicine—and 70% of papers contain project administration and funding acquisition, with the latter task accounting for a higher percentage in PLOS Pathogens and PLOS Genetics. Resources, validation, and visualization are present in about half of all papers. Software contribution appears in less than 40% of papers, except in PLOS Computational Biology, where it is found in almost three-quarters of papers.

Figure 1.

Percentage of papers with specific CRediT contribution, by journal (30,054 papers published in 2017 and 2018).

Figure 1.

Percentage of papers with specific CRediT contribution, by journal (30,054 papers published in 2017 and 2018).

Close modal

To assess division of labor across authors, we compiled, for each journal, the percentage of authors who performed a given contribution. As shown in Figure 2, the majority of authors contribute to writing—review and editing (68%), as well as methodology (55%), investigation (53%), and conceptualization (51%). Worth mentioning is the fact that 95% of authors from PLOS Medicine have contributed to the review and editing of the manuscript; this is likely due to the second criterion of the ICMJE which states that all authors should have “[drafted] the work or [revised] it critically for important intellectual content” (International Committee of Medical Journal Editors, 2019, p. 2). All other CRediT contributions were, on average, performed by a minority of authors. Formal analysis, data curation, and validation were, on average, performed by 42–45% of authors across all PLOS journals, with higher percentages of authors contributing to formal analysis at PLOS Computational Biology and PLOS Genetics, as well as a higher share of authors contributing to validation at PLOS Computational Biology. Contrary to what was observed in the previous typology used by PLOS (Larivière et al., 2016), where more than half of authors (and as much as 80% in social sciences and physics, among others) had “written the paper,” the writing of the original draft is a contribution done by a much narrower percentage of authors (39% across all PLOS journals). Tasks typically performed by principal investigators (resources, supervision, project administration, and funding acquisition), as well as contributions than can be considered to be more specialized (visualization and software) are performed by a minority of authors (between 31% and 38%), with higher percentages of visualization and software for PLOS Computational Biology.

Figure 2.

Percentage of authors who performed a given CRediT contribution (when contribution appears on the paper), by journal (30,054 papers published in 2017 and 2018).

Figure 2.

Percentage of authors who performed a given CRediT contribution (when contribution appears on the paper), by journal (30,054 papers published in 2017 and 2018).

Close modal

Figure 3 shows the percentage of men and women, respectively, who have performed a specific CRediT contribution. The newly adopted taxonomy reinforces some of the initial findings for gender, particularly the gendered divide between conceptual and empirical work: Although 57% of women contributed to the investigation, this percentage is of 49% for men. A similar gap is also observed for data curation. Men, on the other hand, are more likely to conduct tasks associated with seniority, such as funding acquisition and supervision (30% more likely than women), contributing resources, software, conceptualization, and project administration. Although such differences are likely influenced by the fact that women academics are on average younger than men (McChesney & Bichsel, 2020), other studies have shown that gender differences in contributions remained constant with age as well as with the number of authors per paper (Macaluso et al., 2016).

Figure 3.

Percentage of male and female authors who have performed a specific CRediT contribution (30,054 papers published in 2017 and 2018).

Figure 3.

Percentage of male and female authors who have performed a specific CRediT contribution (30,054 papers published in 2017 and 2018).

Close modal

A striking feature of CRediT compared to previous studies based on the PLOS typology (Macaluso et al., 2016) is in the writing of the manuscript. Using the previous PLOS typology, it appeared that men dominated in the writing of the manuscript. However, the nuanced division between writing the original draft and doing reviewing and editing demonstrated a delineation between labor roles for men and women: Women are 6% more likely to have written the original draft, whereas men are 8% more likely to review and edit the manuscript. While those differences are not necessarily sizeable, the fact that we observe a clear inversion of leading genders in the two contributions associated with writing is quite striking. This also demonstrates that the original finding obtained in Macaluso et al. (2016) was skewed by the ubiquity of the “review” portion of writing. Once the taxonomy isolated original drafting of the text, the contribution of women as more likely to write the original draft emerges. This suggests that the more nuanced taxonomy lends greater insight into contrasted divisions of labor.

Division of labor, furthermore, varies as a function of numbers of authors. Figure 4 presents the percentage of authors who have performed a given task, for papers between 1 and 20 authors (N = 29,689 papers, 96.5% of the data set). Obviously, for single-authored papers, 100% of tasks are performed by a single author. As the number of authors increases, tasks are increasingly divided—although the extent to which they are varies as a function of the tasks involved. In other words, while some tasks are performed by a smaller proportion of authors as the number of authors increases, other tasks remain relative stable once a certain threshold is met. For instance, the writing—review and editing task remains performed by a high percentage of authors (i.e., more than half of authors), even when there are 20 authors on a paper. In a similar manner, the proportion of authors who contribute to investigation stabilizes once 10 authors are reached, with, again, about half of authors contributing to the task. Other tasks, however, are increasingly divided as the number of authors increases. For instance, the proportion of authors who perform supervision and writing of the original draft—among others—decreases steadily as the number of authors increases, which suggests, as shown in the inset, that these tasks remain performed by a few authors. More specifically, even in papers by 20 authors, between three and four authors have been involved in those two tasks.

Figure 4.

Percentage of authors who performed a given CRediT contribution, by number of authors, for papers between 1 and 20 authors (N = 29,689 papers). Inset: mean number of authors who performed a subset of CRediT contributions (writing—review and editing, investigation, writing—original draft, and project administration).

Figure 4.

Percentage of authors who performed a given CRediT contribution, by number of authors, for papers between 1 and 20 authors (N = 29,689 papers). Inset: mean number of authors who performed a subset of CRediT contributions (writing—review and editing, investigation, writing—original draft, and project administration).

Close modal

As shown with the previous PLOS typology, there is a strong relationship between authors’ order and tasks performed (Larivière et al., 2016; Sauermann & Haeussler, 2017). Figure 5 presents the percentage of authors who have performed a given CRediT contribution, as a function of their order on the byline of the article (first, middle, last). Taken globally, the figure shows an inverse relationship between the tasks performed by first authors and the tasks performed by last authors. More specifically, first authors are much more likely to write the original draft of the manuscript, curate the data, and perform the formal analysis, visualization, and investigation, as well as contribute to the methodology. Globally, the mean number of tasks to which first authors contribute is higher for first authors, followed by last authors, and then by middle authors. Last authors, on the other hand, are much more likely to have contributed to supervision, funding, resources, and project administration. Conceptualization, and reviewing and editing of the manuscript, are performed by both first and last authors in relatively similar proportions, although last authors are slightly more likely to have performed the tasks. There are no tasks that middle authors are more likely to perform than first and last authors. However, there are a few tasks where their participation is relatively more important: They are more likely to contribute to supervision and to resources than first authors, and more likely to contribute to data curation, investigation, and software than last authors.

Figure 5.

Percentage of authors who performed CRediT contribution, by authors’ order.

Figure 5.

Percentage of authors who performed CRediT contribution, by authors’ order.

Close modal

Figure 6 presents the contributions that are the most likely associated with each other (i.e., performed by the same authors), as well as the asymmetry of these relationships. More specifically, it shows the percentage of authors who have performed contribution A who have also performed contribution B. For example, the figure shows that, although 93% of authors who have contributed to the funding acquisition have reviewed and edited the manuscript, only 46% of authors who reviewed and edited the manuscript have acquired funding. This relationship is among the most asymmetrical, along with software, project administration, visualization, resources, and supervision, on the one hand, and their relationship with reviewing and editing the manuscript. That is not surprising: Writing and editing the manuscript is a task that most authors perform, irrespective of their other contributions to the manuscript. At the other end of the spectrum, funding acquisition is the contribution that has the lowest relationship with other tasks, except with supervision and project administration. A similar phenomenon is observed for supervision, project administration, and resources. Software also has little relation with other tasks, except for visualization.

Figure 6.

Percentage of authors who have performed contribution A who also have performed contribution B.

Figure 6.

Percentage of authors who have performed contribution A who also have performed contribution B.

Close modal

Our analysis has delved into the ways in which scientific labor is accounted using a more refined contributorship taxonomy than was previously available. While confirming several previous findings (Corrêa et al., 2017; Larivière et al., 2016; Sauermann & Haeussler, 2017), the research has provided novel information on the composition and distribution of labor across teams. For example, contributorship information reveals the types of labor that are critical for producing scientific research: Almost all research articles include conceptualization, operationalization, and communication through writing. Deviations by discipline, however, reveal the importance of other more niche tasks, such as visualization and software, acknowledged in certain domains. These findings suggest greater heterogeneity in evaluation processes to attend to the importance of tasks by discipline. Privileging one type of labor will inevitably lead to inequities across disciplines, where specific tasks performed remain either nonperformed or unacknowledged through authorship and contributorship. Furthermore, both the heterogeneity of labor types and the number of contributions per paper suggests that mentoring and doctoral education may need to be reconfigured to address the changing composition of team science (Sugimoto, 2016).

The bureaucratization of science can be considered as an inevitable consequence of the ubiquity of collaborative science (Larivière et al., 2015). As team size increases, the mean number of authors contributing to investigation, for instance, also increases, which suggest that the expansion of teams is largely a function of the increasing number of researchers who contribute to technical tasks, and of the acknowledgment that this contribution warrants authorship (Shapin, 1989). This is not associated with a concomitant rise in those who have written papers’ first drafts or supervisors: There can only be a few supervisors and original authors, but there is a constant expansion in other forms of labor, recognized through authorship (Pontille, 2016). As Shapin (1989) observed: “Scientists’ authority over technicians typically means that it is the former who decide how officially to arrange the relationship, whether to ‘make them’ authors or coauthors, what counts as genuine knowledge as opposed to mere skill, and what technicians’ work signifies in scientific terms” (p. 562). Our research suggests that, despite the steep increase in number of authors, the number of scientific leaders remains small (Robinson-Garcia, Costas et al., 2020). Such division of labor and capital reinforces scientific hierarchies and cumulative advantages (Merton, 1968). Our investigation of the current multiple authorship practices and contributorship distributions illuminates the selective attribution process among coauthors, wherein having one’s name in an article byline does not equate to or result in leadership positions. Consequently, the growing proportion of “supporting authors” (Milojevic, Radicchi, & Walsh, 2018) has strong implications for the composition of the scientific workforce.

The high proportion of data curation—present in 79% of papers—draws attention to a heavily overlooked labor role in science. The majority of articles involve this task, but there is relatively little training provided to doctoral students, nor are many scientists prepared to engage in this. With the increasing prevalence of calls for open science (e.g., McKiernan, Bourne et al., 2016), it is essential that data be properly curated for better sharing and transparency. For example, several countries have established policies requiring the sharing of data created through funded research. Interviews with scientists, however, have revealed strong social and technical challenges to fulfilling these mandates (e.g., Borgerud & Borglund, 2020). Data curation work continue to be widely underresourced, despite increasing calls for data transparency (Leonelli, 2016) and the overwhelming importance of this work, as demonstrated by our analysis. Future work should ensure that data curation is both valued and supported in research environments.

Women are more likely to be associated with this data curation, as well as other technical work, such as investigation, which confirmed results obtained in previous analyses (Macaluso et al., 2016). However, CRediT provided a much more nuanced way to evaluate the conceptual vs. technical divisions identified in earlier research (Macaluso et al., 2016). Furthermore, and perhaps more importantly, the taxonomy elucidated a key difference in one of the main contribution types: writing. Whereas the original five categories contained a single writing category, where men dominated, the new classification distinguished between the editing and reviewing and the much more labor-intensive writing of the first draft. In this distinction, the role of women emerged starkly. Given that they are underrepresented in first and last authorships, this is particularly striking and speaks to some of the underlying injustices in the division of labor and calculation of production (Penders & Shaw, 2020; Rossiter, 1993). This can be critical for the career of women and other underrepresented minorities. As sociologist Mary Frank Fox (2005) observed: “…until we understand factors that are associated with productivity, and variation in productivity by gender, we can neither assess nor correct inequities in rewards, including rank, promotion, and salary […] because publication productivity operates as both cause and effect of status in science […] productivity reflects women’s depressed rank and status, and partially accounts for it.” It is no surprise, therefore, that junior scholars were the most concerned about their representation in contributorship statements and expressed the greatest desire for broad participation in these discussions (Sauermann & Haeussler, 2017). There is a considerable need for greater transparency about the career lifecycles and interoperability between systems (Cañibano, Woolley et al., 2019). The integration of CRediT and ORCID is a useful start to this.

It is clear from the data that contributorship provides a lens to add greater transparency in the capital exchange for authorship. In addition to providing greater accountability for research, contributorship also sheds greater light on the flaws in the current system. Our work demonstrates a clear division of labor as team size increases and the corresponding isolation of certain contribution types. While this facilitates efficiency and may be necessary for certain types of research, it inevitably increases the chances of potential misconduct, mistake, or fraud, given that several team members provide their contributions without direct oversight5. One critical role, therefore, may be validation. However, this was present in only 55% of papers (performed by 42% of authors). One may argue that this is merely idiosyncratic interpretations of the contributorship roles, where some authors may consider validation a part of the “investigation” or “formal analysis.” However, the task definition is clear: “verification, whether as a part of the activity or separate, of the overall replication/reproducibility of results/experiments and other research outputs.” The lack of validation in the PLOS papers reinforces the concerns of the “reproducibility crisis” (Baker, 2016). To address this, journals could require validation as a mandatory contribution type for empirical work. Contributorship statements are not without limitation. One strong concern at present is the assumed relationship between the actual labor and the indicator of this labor in contributorship statements. Undoubtedly, when scholars mutually ascribe the different tasks of CRediT to themselves, they maintain the opacity necessary to favor good working relationships between colleagues and teams. As criteria for authorship vary considerably across disciplines (Paul-Hus et al., 2017; Pontille, 2004, 2016), so too might the interpretation of contribution roles. More research is necessary to understand whether CRediT provides a valid representation of the work.

Another related general concern has simultaneously been raised by some clinical researchers and regulatory bodies regarding these expansive categories: if contributorship removes “much of the ambiguity surrounding contributions, it leaves unresolved the question of the quantity and quality of contribution that qualify for authorship” (International Committee of Medical Journal Editors, 2019). As with any system tied to capital, there is likely to be goal displacement as the taxonomy gains wider acceptance and use. For example, the disproportionately high degree of PLOS Medicine authors associated with writing and editing may be less a disciplinary difference and more an adherence to the ICMJE criteria. And, as some critically emphasized, the contributorship procedure favors pharmaceutical firms that, without having to pretend to intervene intellectually by figuring in an article byline, could now become “contributors” and thus avoid allegations of conflicts of interest (Matheson, 2011). This suggests that authors may modify their behavior in order to meet certain requirements, norms, or incentives. Further investigations are thus needed to explore such issues.

Despite the accountability it aspires to, any description of scientific contributions, even the fine-grained provided by CRediT, can never be complete. As Sauermann and Haeussler (2017) noted, contributorship statements may reduce misconduct while simultaneously leading scientists to avoid association with those tasks with a greater potential for risk. Scientists may also begin to adopt similar practices of ghost, guest, and gift authorship to contributorship. The systematic description of work does not, therefore, preclude invisibility, but only displaces it elsewhere. As a consequence, it leaves ghostwriting of articles and potential honorary contributorship in the backrooms of scientific research. Contributorship statements are not a panacea for the problems of authorship misconduct; however, they do contribute to clarifying the contributions that are sufficiently important to warrant authorship from those that are not. Issues with authorship are not an indication of problems inherent with the contributorship model, but symptomatic of a larger structural problem in the contemporary scientific community, which is the demand, by both policymakers and researchers themselves, for procedural ways of assessing excellence and scientific performance.

Over the last few decades, transparency in authorship and scholarly publishing have become increasingly discussed in academe. This is due to several interrelated phenomena. First, bibliometric evaluations have become widespread across all countries, and have been applied to the promotion of individual researchers (Quan, Chen, & Shu, 2017) and to institutions, mostly through the ever-expanding university rankings (Debackere & Glänzel, 2004). Secondy, the rise in the number of PhD graduates, linked with the relative stability of faculty positions, is increasing the competition among new graduates, who are ever more aware—as this is often made explicit—that publications are the currency that will allow them to find a position. The pressures wrought by this system have led to several authorship malpractices. There are flagrant acts of “civil disobedience” in authorship, such as adding humorous fictional coauthors, pets, or celebrities to a paper (Penders & Shaw, 2020). However, some new authorship issues are more pernicious, such as adding children as coauthors so that they can begin to build their publication record (Zastrow, 2019), and the growth of predatory publishing (Grudniewicz, Moher et al., 2019) and publication bazaars (Hvistendahl, 2013). These latter actions demonstrate how critical authorship is for the reward structure of science and the misconduct that can arise as a result of these pressures to publish.

By fragmenting scientific production process into clearly distinct tasks, CRediT was designed to transcend the customary rules specific to name orderings in scientific publications. Information about the conditions of production of research is being made available in each scientific article, and the systematic description of contributions according to CRediT is not limited to the authorship practices of a particular discipline. On the contrary, it can easily be adjusted to various kinds of division of scientific labor and their specific hierarchical principles across research teams (e.g., a team led by a leader, a project carried out among peers, a multicenter research project). In other words, CRediT is not at odds with the distinct authorship practices in place across disciplines. Rather, based on the traceability of individual performance, it provides additional information on the attribution process. Simultaneously, as other accounting devices (Strathern, 2000), the systematic description of contributions, especially through CRediT, comes with ambivalence. While it undoubtedly introduces greater transparency in both reward and accountability related to the division of labor involved in a published article, it simultaneously fuels a regression of trust at the root of scientific relations (Pontille, 2015). Put differently, the beneficiaries of the information made available—especially women and junior scholars—may become the potential victims of devices that facilitate monitoring and surveillance at the heart of scientific activity.

All of these elements have one point in common: the (sole) emphasis on scholarly publications as the criterion for research excellence. It seems that, along the way, we have forgotten what drives researchers to do what we do, and why our societies have made the choice to support us in this endeavor, which is to discover new things. We have replaced a “taste for science” by a “taste for publication” (Osterloh & Frey, 2014). As per Gingras (2018), the meaning of scholarly publications has changed from a unit of (new) knowledge produced, to an accounting—or accountability—unit. Directly related to CRediT, “the systematic description of contributions leads toward accounting management for scientific activity. […] As a divisible, accounting unit, each scientific act may even be associated with a specific amount” (Pontille, 2016: 122). In this way, contributorship does not dismantle performance-based rewards (Debackere & Glänzel, 2004; Sivertsen, 2010), but rather serves to bring greater precision in accounting.

The authors have no competing interests.

This research was funded by the Canada Research Chairs program (950-231768).

The PLOS data set can be downloaded from https://doi.org/10.6084/m9.figshare.13277168.v1.

1

Although this is included in CRediT, these distinctions were not given in the data provided by PLOS for our analysis.

3

This made the processing of contributorships much more straightforward than what is provided through the bulk download of the full text of papers in XML format (http://api.plos.org/text-and-data-mining/). See, for instance, Larivière et al. (2016). In this case, the full names of authors were provided, along with each contribution role, thereby facilitating the author-matching process.

4

A different editorial system for PLOS Biology made it difficult for PLOS to provide us with the data for this journal. Therefore, while the PLOS Biology contributorship data is included in the global analysis, individual data for the journal is not provided (i.e., Figures 1 and 2).

5

Some journals (e.g., BMJ) identify a role for a “guarantor,” who will take responsibility for the entire manuscript. This is also the implied role taken by many corresponding authors. It is not, however, made explicit and is not easily defensible in misconduct cases.

Allen
,
L.
,
O’Connell
,
A.
, &
Kiermer
,
V.
(
2019
).
How can we ensure visibility and diversity in research contributions? How the Contributor Role Taxonomy (CRediT) is helping the shift from authorship to contributorship
.
Learned Publishing
,
32
(
1
),
71
74
.
Allen
,
L.
,
Scott
,
J.
,
Brand
,
A.
,
Hlava
,
M.
, &
Altman
,
M.
(
2014
).
Publishing: Credit where credit is due
.
Nature
,
508
(
7496
),
312
313
.
Atkins
,
H.
(
2016
).
Author credit: PLOS and CRediT update. July 8, 2016
. .
Azoulay
,
P.
,
Furman
,
J. L.
,
Krieger
,
J. L.
, &
Murray
,
F.
(
2015
).
Retractions
.
Review of Economics and Statistics
,
97
(
5
),
1118
1136
.
Baerlocher
,
M. O.
,
Gautam
T.
,
Newton
M.
, &
Tomlinson
G.
(
2009
).
Changing author counts in five major general medicine journals: Effect of author contribution forms
.
Journal of Clinical Epidemiology
,
62
(
8
),
875
877
.
Baker
,
M.
(
2016
).
1500 scientists lift the lid on reproducibility
.
Nature
,
553
(
7604
),
452
454
.
Bates
,
T.
,
Anić
,
A.
,
Marušić
,
M.
, &
Marušić
,
A.
(
2004
).
Authorship criteria and disclosure of contributions: Comparison of 3 general medical journals with different author contribution forms
.
Journal of the American Medical Association
,
292
(
1
),
86
88
.
Bhandari
,
M.
,
Guyatt
,
G. H.
,
Kulkarni
,
A. V.
,
Devereaux
,
P. J.
,
Leece
,
P.
, …
Busse
,
J. W.
(
2014
).
Perceptions of authors’ contributions are influenced by both byline order and designation of corresponding author
.
Journal of Clinical Epidemiology
,
67
(
9
),
1049
1054
.
Biagioli
,
M.
, &
Galison
,
P.
(Eds.) (
2003
).
Scientific authorship: Credit and intellectual property in science
.
New York and London
:
Routledge
.
Birnholtz
,
J. P.
(
2006
).
What does it mean to be an author? The intersection of credit, contribution, and collaboration in science
.
Journal of the American Society for Information Science and Technology
,
57
(
13
),
1758
1770
.
Borgerud
,
C.
, &
Borglund
,
E.
(
2020
).
Open research data, an archival challenge?
Archival Science
,
20
,
279
302
.
Bourdieu
,
P.
(
2001
).
Science de la science et réflexivité
.
Paris
:
Raisons d’Agir
.
Brand
,
A.
,
Allen
,
L.
,
Altman
,
M.
,
Hlava
,
M.
, &
Scott
,
J.
(
2015
).
Beyond authorship: Attribution, contribution, collaboration, and credit
.
Learned Publishing
,
28
(
2
),
151
155
.
Cañibano
,
C.
,
Woolley
,
R.
,
Iversen
,
E. J.
,
Hinze
,
S.
,
Hornbostel
,
S.
, &
Tesch
,
J.
(
2019
).
A conceptual framework for studying science research careers
.
Journal of Technology Transfer
,
44
(
6
),
1964
1992
.
Corrêa
Jr.,
E. A.
,
Silva
,
F. N.
,
Costa
,
L. D. F.
, &
Amancio
,
D. R.
(
2017
).
Patterns of authors contribution in scientific manuscripts
.
Journal of Informetrics
,
11
(
2
),
498
510
.
Cronin
,
B.
(
2001
).
Hyperauthorship: A postmodern perversion or evidence of a structural shift in scholarly communication practices?
Journal of the American Society for Information Science and Technology
,
52
(
7
),
558
569
.
Davidoff
,
F.
(
2000
).
Who’s the author? Problems with biomedical authorship, and some possible solutions
.
Science Editor
,
23
(
4
),
111
119
.
Debackere
,
K.
, &
Glänzel
,
W.
(
2004
).
Using a bibliometric approach to support research policy making: The case of the Flemish BOF-key
.
Scientometrics
,
59
(
2
),
253
276
.
Elsevier
. (
2019
).
Elsevier expands CRediT approach to authorship
.
Flanagin
,
A.
,
Carey
,
L. A.
,
Fontanarosa
,
P. B.
,
Philips
,
S. G.
,
Pace
,
B. P.
, …
Rennie
,
D.
(
1998
).
Prevalence of articles with honorary authors and ghost authors in peer-reviewed medical journals
.
Journal of the American Medical Association
,
280
(
3
),
222
224
.
Frank Fox
,
M.
(
2005
).
Gender, family characteristics, and publication productivity among scientists
.
Social Studies of Science
,
35
(
1
),
131
150
.
Gingras
,
Y.
(
2018
).
Les transformations de la production du savoir: De l’unité de connaissance à l’unité comptable
.
Zilsel
,
4
(
2
),
139
152
.
Goodman
,
N. W.
(
1994
).
Survey of fulfilment of criteria for authorship in published medical research
.
BMJ
,
309
(
6967
),
1482
.
Gøtzsche
,
P. C.
,
Hróbjartsson
,
A.
,
Johansen
,
H. K.
,
Haahr
,
M. T.
,
Altman
,
D. G.
, &
Chan
,
A.-W.
(
2007
).
Ghost authorship in industry-initiated randomised trials
.
PLOS Medicine
,
4
(
1
),
e19
.
Grudniewicz
,
A.
,
Moher
,
D.
,
Cobey
,
K. D.
,
Bryson
,
G. L.
,
Cukier
,
S.
, …
Ciro
,
J. B.
(
2019
).
Predatory journals: No definition, no defence
.
Nature
,
576
,
210
212
.
Haeussler
,
C.
, &
Sauermann
,
H.
(
2013
).
Credit where credit is due? The impact of project contributions and social factors on authorship and inventorship
.
Research Policy
,
42
(
3
),
688
703
.
Hardjosantoso
,
H. C.
,
Dahi
,
Y.
,
Verhemel
,
A.
,
Dahi
,
I.
, &
Gadiradi
,
P. S.
(
2020
).
Honorary authorships in the opthalmological literature
.
Journal of Current Opthalmology
,
32
(
2
),
199
202
.
Horton
,
R.
(
1997
).
The signature of responsibility
.
Lancet
,
350
(
9070
),
5
6
.
Hvistendahl
,
M.
(
2013
).
China’s publication bazaar
.
Science
,
342
,
1035
1039
.
Ilakovac
,
V.
,
Fister
,
K.
,
Marusic
,
M.
, &
Marusic
,
A.
(
2007
).
Reliability of disclosure forms of authors’ contributions
.
Canadian Medical Association Journal
,
176
(
1
),
41
46
.
International Committee of Medical Journal Editors
. (
2019
).
Recommendations for the conduct, reporting, editing, and publication of scholarly work in medical journals
.
Ivaniš
,
A.
,
Hren
,
D.
,
Marušić
,
M.
, &
Marušić
,
A.
(
2011
).
Less work, less respect: Authors’ perceived importance of research contributions and their declared contributions to research articles
.
PLOS ONE
,
6
(
6
),
e20206
.
Ivaniš
,
A.
,
Hren
,
D.
,
Sambunjak
,
D.
,
Marušić
,
M.
, &
Marušić
,
A.
(
2008
).
Quantification of authors’ contributions and eligibility for authorship: Randomized study in a general medical journal
.
Journal of General Internal Medicine
,
23
(
9
),
1303
1310
.
IWCSA
. (
2012
).
Report on the International Workshop on Contributorship and Scholarly Attribution
,
May 16, 2012
.
Harvard University and the Wellcome Trust
. .
Karimi
,
F.
,
Wagner
,
C.
,
Lemmerich
,
F.
,
Jadidi
,
M.
, &
Strohmaier
,
M.
(
2016
).
Inferring gender from names on the web: A comparative evaluation of gender detection methods
.
Proceedings of the 25th International Conference Companion on World Wide Web
(pp.
53
54
).
New York
:
ACM
.
Larivière
,
V.
,
Desrochers
,
N.
,
Macaluso
,
B.
,
Mongeon
,
P.
,
Paul-Hus
,
A.
, &
Sugimoto
,
C. R.
(
2016
).
Contributorship and division of labor in knowledge production
.
Social Studies of Science
,
46
(
3
),
417
435
.
Larivière
,
V.
,
Ni
,
C.
,
Gingras
,
Y.
,
Cronin
,
B.
, &
Sugimoto
,
C. R.
(
2013
).
Bibliometrics: Global gender disparities in science
.
Nature
,
504
(
7479
),
211
.
Larivière
,
V.
,
Sugimoto
,
C.R.
,
Tsou
,
A.
, &
Gingras
,
Y.
(
2015
).
Team size matters: Collaboration and scientific impact since 1900
.
Journal of the Association for Information Science and Technology
,
66
(
7
),
1323
1332
.
Leonelli
,
S.
(
2016
).
Open data: Curation is under-resourced
.
Nature
,
538
,
41
.
Macaluso
,
B.
,
Larivière
,
V.
,
Sugimoto
,
T.
, &
Sugimoto
,
C. R.
(
2016
).
Is science built on the shoulders of women? A study of gender differences in contributorship
.
Academic Medicine
,
91
(
8
),
1136
1142
.
Matheson
,
A.
(
2011
).
How industry uses the ICMJE guidelines to manipulate authorship—and how they should be revised
.
PLOS Medicine
,
8
(
8
),
e1001072
.
Marušić
,
A.
,
Bates
,
T.
,
Anić
,
A.
, &
Marušić
,
M.
(
2006
).
How the structure of contribution disclosure statements affects validity of authorship: A randomized study in a general medical journal
.
Current Medical Research and Opinion
,
22
(
6
),
1035
1044
.
McChesney
,
J.
, &
Bichsel
,
J.
(
2020
).
The aging of tenure-track faculty in higher education: Implications for succession and diversity
.
Knoxville, TN
:
College and University Professional Association for Human Resources
.
McDonald
,
R. J.
,
Neff
,
K. L.
,
Rethlefsen
,
M. L.
, &
Kallmes
,
D. F.
(
2010
).
Effects of author contribution disclosures and numeric limitations on authorship trends
.
Mayo Clinic Proceedings
,
85
(
10
),
920
927
.
McKiernan
,
E. C.
,
Bourne
,
P. E.
,
Brown
,
C. R.
,
Buck
,
S.
,
Kenall
,
A.
, …
Yarkoni
,
T.
(
2016
).
How open science helps researchers succeed
.
eLife
,
5
,
e16800
.
Merton
,
R. K.
(
1968
).
The Matthew Effect in science
.
Science
,
159
(
3810
),
56
63
.
Milojevic
,
S.
,
Radicchi
,
F.
, &
Walsh
,
J. P.
(
2018
).
Changing demographics of scientific careers: The rise of the temporary workforce
.
Proceedings of the National Academy of Sciences
,
115
(
50
),
12616
12623
.
Moulopoulos
,
S. D.
,
Sideris
,
D. A.
, &
Georgilis
,
K. A.
(
1983
).
For debate…Individual contributions to multiauthor papers
.
British Medical Journal
,
287
(
6405
),
1608
1610
.
Mowatt
,
G.
,
Shirran
,
L.
,
Grimshaw
,
J. J. M.
,
Rennie
,
D.
,
Flanagin
,
A.
, …
Bero
,
L. A.
(
2002
).
Prevalence of honorary and ghost authorship in Cochrane reviews
.
Journal of the American Medical Association
,
287
(
21
),
2769
2771
.
Northridge
,
M.
(
1998
).
Annotation: New rules for authorship in the journal: Your contributions are recognized—and published!
American Journal of Public Health
,
88
(
5
),
733
734
.
Osterloh
,
M.
, &
Frey
,
B. S.
(
2014
).
Ranking games
.
Evaluation Review
,
39
(
1
),
102
129
.
Paul-Hus
,
A.
,
Mongeon
,
P.
,
Sainte-Marie
,
M.
, &
Larivière
,
V.
(
2017
).
The sum of it all: Revealing collaboration patterns by combining authorship and acknowledgements
.
Journal of Informetrics
,
11
(
1
),
80
87
.
Penders
,
B.
, &
Shaw
,
D.
(
2020
).
Civil disobedience in scientific authorship: Resistance and insubordination in science
.
Accountability in Research
,
27
(
6
),
347
371
.
Pontille
,
D.
(
2004
).
La signature scientifique. Une sociologie pragmatique de l’attribution
.
Paris
:
CNRS Éditions
.
Pontille
,
D.
(
2015
).
Les transformations de la contribution scientifique
.
Histoire de la Recherche Contemporaine
,
4
(
2
),
152
162
.
Pontille
,
D.
(
2016
).
Signer ensemble. Contribution et évaluation en sciences
.
Paris
:
Economica
.
Quan
,
W.
,
Chen
,
B.
, &
Shu
,
F.
(
2017
).
Publish or impoverish
.
Aslib Journal of Information Management
,
69
(
5
),
486
502
.
Rennie
,
D.
, &
Flanagin
,
A.
(
1994
).
Authorship! Authorship!: Guests, ghosts, grafters, and the two-sided coin
.
Journal of the American Medical Association
,
271
(
6
),
469
471
.
Rennie
,
D.
,
Flanagin
,
A.
, &
Yank
,
V.
(
2000
).
The contributions of authors
.
Journal of the American Medical Association
,
284
(
1
),
89
91
.
Rennie
,
D.
,
Yank
,
V.
, &
Emanuel
,
L.
(
1997
).
When authorship fails: A proposal to make contributors accountable
.
Journal of the American Medical Association
,
278
(
7
),
579
585
.
Robinson-Garcia
,
N.
,
Costas
,
R.
,
Sugimoto
,
C. R.
,
Larivière
,
V.
, &
Nane
,
T.
(
2020
).
Task specialization and its effects on research careers
.
eLife
,
9
,
e60586
.
Rossiter
,
M. W.
(
1993
).
The Matthew Matilda effect in science
.
Social Studies of Science
,
23
(
2
),
325
341
.
Saffran
,
M.
(
1989
).
On multiple authorship: Describe the contribution
.
The Scientist
,
3
(
6
),
9
11
.
Santamaría
,
L.
, &
Mihaljević
,
H.
(
2018
).
Comparison and benchmark of name-to-gender inference services
.
PeerJ Computer Science
,
4
,
e156
.
Sauermann
,
H.
, &
Haeussler
,
C.
(
2017
).
Authorship and contribution disclosures
.
Science Advances
,
3
(
11
),
e1700404
.
Shapin
,
S.
(
1989
).
The invisible technician
.
American Scientist
,
77
(
6
),
554
563
.
Siler
,
K.
,
Larivière
,
V.
, &
Sugimoto
,
C. R.
(
2020
).
The diverse niches of megajournals: Specialism within generalism
.
Journal of the Association for Information Science and Technology
,
71
(
7
),
800
816
.
Sismondo
,
S.
(
2009
).
Ghosts in the machine: Publication planning in the medical sciences
.
Social Studies of Science
,
39
(
2
),
171
198
.
Sismondo
,
S.
, &
Doucet
,
M.
(
2010
).
Publication ethics and the ghost management of medical publication
.
Bioethics
,
24
(
6
).
Sivertsen
,
G.
(
2010
).
A performance indicator based on complete data for the scientific publication output at research institutions
.
ISSI Newsletter
,
6
(
1
),
22
28
.
Smith
,
E.
,
Williams-Jones
,
B.
,
Master
,
Z.
,
Larivière
,
V.
,
Sugimoto
,
C. R.
, …
Resnik
,
D. B.
(
2020a
).
Misconduct and misbehavior related to authorship disagreements in collaborative science
.
Science and Engineering Ethics
,
26
,
1967
1993
.
Smith
,
E.
,
Williams-Jones
,
B.
,
Master
,
Z.
,
Larivière
,
V.
,
Sugimoto
,
C. R.
, …
Resnik
,
D. B.
(
2020b
).
Researchers’ perceptions of ethical authorship distribution in collaborative research teams
.
Science and Engineering Ethics
,
26
,
1995
2022
.
Smith
,
J.
(
1994
).
Gift authorship – a poisoned chalice
.
British Medical Journal
,
309
(
6967
),
1456
1457
.
Smith
,
R.
(
1997
).
Authorship: Time for a paradigm shift?
British Medical Journal
,
314
(
7086
),
992
.
Strathern
,
M.
(
2000
).
The tyranny of transparency
.
British Educational Research Journal
,
26
(
3
),
309
321
.
Street
,
J. M.
,
Rogers
,
W. A.
,
Israel
,
M.
, &
Braunack-Mayer
,
A. J.
(
2010
).
Credit where credit is due? Regulation, research integrity and the attribution of authorship in the health sciences
.
Social Science & Medicine
,
70
(
9
),
1458
1465
.
Sugimoto
,
C. R.
(
2016
).
Toward a twenty-first century dissertation. Future of the Dissertation Workshop
.
Council of Graduate Schools
.
Wilcox
,
L. J.
(
1998
).
Authorship: The coin of the realm, the source of complaints
.
Journal of the American Medical Association
,
280
(
3
),
216
217
.
Wislar
,
J. S.
,
Flanagin
,
A.
,
Fontanarosa
,
P. B.
, &
DeAngelis
,
C. D.
(
2011
).
Honorary and ghost authorship in high impact biomedical journals: A cross sectional survey
.
British Medical Journal
,
343
,
d6128
.
Zastrow
,
M.
(
2019
).
More South Korean academics caught naming kids as co-authors
.
Nature
,
575
,
267
268
.
Zuckerman
,
H. A.
(
1968
).
Patterns of name ordering among authors of scientific papers: A study of social symbolism and its ambiguity
.
American Journal of Sociology
,
74
(
3
),
276
291
.

Author notes

Handling Editor: Ludo Waltman

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/legalcode.