Abstract
This study quantifies and analyzes the individual-level abilities of scientists utilizing either an exploration or an exploitation strategy. Specifically, we present a Research Strategy Q model, which untangles the coupling effect of scientists’ research ability (Qα) and research strategy ability () on research performance. Qα indicates scientists’ fundamental ability to publish high-quality papers, while indicates scientists’ proficiency in terms of exploration and exploitation strategies. Five research strategies proposed by our previous study are employed. We generate synthetic data and collect empirical data as our experimental data set. We show that these research strategies present different benefit and risk characteristics. Adopting some exploitation strategies tends to stifle research performance, while exploration strategies are high risk and high yield. Qα and have predictive power for research performance. Moreover, we find that, first, scholars who prefer to execute a research strategy, π, may not necessarily be better at executing π. Second, some scholars have differences in their abilities towards different strategies, while other scholars have differences in their abilities towards the same strategy. Third, exploration and exploitation are not contradictory but complementary from the perspective of proficiency, while they are mutually exclusive from the perspective of selection preference.
PEER REVIEW
1. INTRODUCTION
Nowadays, an increasing number of scholars are devoting themselves to scientific research and publish a huge number of papers yearly. Evaluating and understanding scholars’ academic performance is a crucial research direction in Science of Science (Zhou, Wang et al., 2020), which serves academic contexts, such as grant applications, performance evaluation, and personnel promotion (Wang, Zhou, & Zeng, 2023). Fair and diversified scientific evaluation is necessary for promoting scientific progress (Ain, Riaz, & Afzal, 2019; Sinatra, Wang et al., 2016). The exploration and exploitation dilemma inevitably faced by scholars makes exploration-exploitation the crucial dimension by which to evaluate individual scholars (Yang, Hu et al., 2023).
With the development of big data analysis techniques, data-driven bibliometrics-based evaluation methods have become a powerful supplement to peer review. Bibliometrics-based methods mainly analyze scientists’ research performance from number of citations and papers, citation and coauthorship network, and academic texts. However, traditional author-level evaluation metrics generally evaluate scholars from a single-dimensional view of their impact (Zhang, Wang et al., 2023), cannot measure the innovation level of scientists (Petersen, Riccaboni et al., 2012), and neglect the research focus of scholars. Recent studies show that the exploration-exploitation research strategies present significantly different return and risk characteristics (Jia, Wang, & Szymanski, 2017; Yin, Wang et al., 2019), and scholars show distinct preferences for them (Foster, Rzhetsky, & Evans, 2015; Huang, Lu et al., 2022). Inspired by these studies, this study aims to address the following research issues:
propose an explanatory model to untangle the effect of individual-level ability and five research strategies proposed by Huang et al. (2022) on research performance;
quantitatively examine the return and risk characteristics of these research strategies; and
comparatively analyze scientists’ abilities towards different strategies.
In this paper, we propose a Research Strategy Q model (RSQ) that untangles and quantifies the effect of scientists’ research ability (Qα) and research strategy ability () on their research performance. Qα is defined as the underlying fundamental ability of a scientist to publish high-quality papers (Huang, Lu et al., 2023b; Sinatra et al., 2016), while is defined as the scientist’s proficiency in exploration and exploitation strategies. The five research strategies proposed by Huang et al. (2022) are adopted. We generate synthetic data and separately select about 20,000 scholars in the physics, chemistry, and computer science fields as our data sets. Our results show that our model can effectively uncouple the distinct abilities of a scientist and disclose that different research strategies have obviously distinct benefits and risk features. Moreover, a scholar may have separate proficiency levels for different research strategies (e.g., ≠ ), while a number of scholars may have separate proficiency levels for the same specific research strategy (e.g., ≠ ). Importantly, exploration and exploitation are not contradictory but complementary from the perspective of proficiency, while they are mutually exclusive from the perspective of selection preference. Our code is open and accessible1.
This study has the following theoretical and practical implications. The RSQ model is presented to explain how individual research ability and proficiency in five research strategies affect scientists’ research performance. On the one hand, the RSQ model discloses the unique characteristics of different research strategies, and sheds light on the topic selection behavior from the perspective of exploration and exploitation. On the other hand, the RSQ model can be utilized as a novel tool to evaluate scientists. Governments, academic institutions, and funding agencies may employ our model to evaluate scholars, by which they can improve their talent assessment mechanism.
The rest of this paper is organized as follows. In Section 2, we review related studies. In Section 3, we present the problem definition and our model. In Section 4, we clarify our data set. In Section 5, we provide the analysis of experimental results. In Section 6, we conclude with theoretical and practical implications, limitations, and points for future research.
2. BACKGROUND
2.1. Reviews on Scholar Evaluation
Peer review and bibliometrics-based methods are two mainstream methods for evaluating scholars’ research performance. In peer review, scholars and their publications are manually evaluated by domain experts or peers based on their knowledge and experience. This subjective method is effective in evaluating small numbers of scientists. However, with more and more scholars devoting themselves to scientific research and the increasing volume of publications, evaluation purely based on peer review is not feasible because it is both time and effort consuming (Bornmann & Daniel, 2009a). In addition, peer review is inevitably influenced by subjective factors, such as cognitive distortions (Protasiewicz, Pedrycz et al., 2016). Hence, peer review is often combined with bibliometrics-based methods.
Bibliometrics-based methods generally evaluate the productivity and impact of scholars based on their bibliometric information, such as citation count, number of papers, h-index, and variants of the h-index (Ain et al., 2019). The h-index, proposed by Hirsch (2005), simultaneously measures the productivity and impact of scholars, and shows a certain degree of consistency with peer review (Johnson & Lovegrove, 2008). However, the h-index suffers from some obvious problems. For example, it cannot completely reflect the citation distribution of a scientist’s papers (Bornmann & Daniel, 2009b); it is affected by self-citation (Bartneck & Kokkelmans, 2011); it puts new scholars at a disadvantage; and it is insensitive to changes in research performance (Rousseau & Leuven, 2008). To solve these issues, several variants of the h-index have been proposed. For instance, Egghe (2006) proposed the g-index, defined as the largest number such that the top g papers received at least g2 citations. The g-index is more able to take into account highly cited papers. Alonso, Cabrerizo et al. (2010) computed the hg-index by taking the geometric mean of the h-index and g-index. Schreiber (2018) developed a timely h-index, ht, that only considers a scientist’s publication and citation records of the previous t years. ht alleviates the time bias caused by the dominance of rather old papers in the h-index. Khurana and Sharma (2022) presented hc by adding the weight of the highest cited paper to the h-index, which improves the evaluation performance for low-ranked scholars. Ain et al. (2019) examined the differences between the h-index and its citation intensity-based variants, and found that a high correlation exists among most of them. However, these metrics overlook the impact of research topics and strategies on scientists’ research performance.
PageRank-based algorithms have been utilized to rank scientists, in which citation networks and coauthorship networks are constructed. For example, Sidiropoulos and Manolopoulos (2006) ranked scholars based on the average PageRank score of their papers in the citation network. Gao, Wang et al. (2016) combined the h-index and PageRank algorithm to present the PR-index, which provides a more balanced impact measure than existing indices. Yan and Ding (2011) introduced a weighted PageRank algorithm based on a coauthorship network to measure author impact. Zhang et al. (2023) proposed a RelRank algorithm based on the PageRank algorithm, in which a coauthorship network consisting of the author–venue relationship is developed. RelRank can identify high-impact scholars who are devoted to specific publication outlets. However, the above author-level evaluation indicators still evaluate scientists from a single-dimensional view of their impact (Zhang et al., 2023) and cannot evaluate exploration-exploitation performance of scientists.
Sinatra et al. (2016) proposed a novel Q model, which uncouples the effects of productivity, research ability, and luck on scholars’ research performance. The Q parameter estimated by the probabilistic model is a time-invariant indicator that can truly account for a scientist’s individual ability. However, the Q model neglects the effect of research topics. Yang et al. (2023) presented the disruptive h-index and consolidating h-index based on disruption metrics (Funk & Owen-Smith, 2017) to comprehensively evaluate the disruptive and consolidating impact of scientists. However, the disruptive/consolidating h-index is hard to interpret, because the h-index simply combines two numbers of two distinct units. Inspired by their studies, I employ an interpretable probabilistic graphical model to evaluate the exploration-exploitation performance of scholars.
2.2. Reviews on Exploration and Exploitation
The conflicting demands between productive tradition and risky innovation bring an “essential tension” to scientists (Kuhn, 1977), and scientists inevitably face a dilemma between exploration and exploitation (Cohen, McClure, & Yu, 2007). Consolidating studies (i.e., exploitation) tend to be more easily accepted by the current scientist evaluation system, and therefore have a greater probability of being published (Jia et al., 2017). Innovating studies (i.e., exploration) generally have extremely high time cost and sunk cost, but are more likely to have greater impact in the long run (Yin et al., 2019). Many studies have studied the choice of exploration and exploitation in terms of scientists’ topic selection behavior.
Jia et al. (2017) disclosed three fundamental properties dominating changes in research interests (i.e., heterogeneity, recency, and subject proximity), and also proposed a random-walk-based to reproduce the empirical observations, by which they shed light on the interplay between exploitation and exploration. Huang, Huang et al. (2023a) developed a random walk-based model based on reinforcement learning theory to examine the role of productivity and impact in the evolution of scientists’ research interests, and found that scientists prefer to select topics that help them improve their productivity and impact. Franco, Malhotra, and Simonovits (2014) showed that scholars prefer to submit papers that reject the null hypothesis, to ensure publication. Liu, Wang et al. (2018) and Liu, Dehmamy et al. (2021) examined the career histories of artists, film directors, and scientists, and showed that successful individuals generally undergo extensive exploration before engaging in sufficient exploitation. Zeng, Shen et al. (2019) analyzed the topic-switching dynamics of physicists, and showed that they shift more frequently among topics than they did in the past. Chen and Ding (2023) analyzed the publication records of 117 Nobel laureates in physics, and found that they generally explore about three topics alternately in different periods and focus on exploiting their core topics throughout their careers. Foster et al. (2015) introduced five research strategies (i.e., jump, new consolidation, repeat consolidation, new bridge, repeat bridge) based on chemical entities networks. They demonstrated that the aggregate distribution of research strategies remains extremely stable, and risky innovation strategies are especially rare. Recently, Huang et al. (2022) proposed five novel research strategies under exploration and exploitation behavior; that is, mature topic selection strategy (πM), popular topic selection strategy (πP), diverse topic selection strategy (πD), emerging topic selection strategy (πE), and combinatorial innovation topic selection strategy (πC). They also presented a series of metrics to identify the research strategies adopted by a paper. They found that impactful scientists frequently execute πD, πP, πE, πC but execute πM less often. The five strategies and corresponding metrics are summarized in Table 1.
The five research strategies proposed by Huang et al. (2022)
Type . | Strategy . | Definition . | Metrics . |
---|---|---|---|
Exploitation | Mature topic selection strategy (πM) | Scientists consolidate mature topics that have been extensively studied | −logP1(Xα,i; tα,i) and −logP2(Xα,i; tα,i) |
Exploitation | Popular topic selection strategy (πP) | Scientists follow hot topics that have recently received widespread attention | Productivity ratio (Xα,i, tα,i) and Impact ratio (Xα,i, tα,i) |
Exploitation | Diverse topic selection strategy (πD) | Scientists study topics that cover diverse knowledge in semantic space | Average distance (Xα,i) and Average volume (Xα,i) |
Exploration | Emerging topic selection strategy (πE) | Scientists explore emerging topics identified by novelty index | μ(Novelty index, Xα,i, tα,i) and σ(Novelty index, Xα,i, tα,i) |
Exploration | Combinatorial innovation topic selection strategy (πC) | Scientists creatively combine dissimilar topics | μ(Innovation index, Xα,i, tα,i) and σ(Innovation index, Xα,i, tα,i) |
Type . | Strategy . | Definition . | Metrics . |
---|---|---|---|
Exploitation | Mature topic selection strategy (πM) | Scientists consolidate mature topics that have been extensively studied | −logP1(Xα,i; tα,i) and −logP2(Xα,i; tα,i) |
Exploitation | Popular topic selection strategy (πP) | Scientists follow hot topics that have recently received widespread attention | Productivity ratio (Xα,i, tα,i) and Impact ratio (Xα,i, tα,i) |
Exploitation | Diverse topic selection strategy (πD) | Scientists study topics that cover diverse knowledge in semantic space | Average distance (Xα,i) and Average volume (Xα,i) |
Exploration | Emerging topic selection strategy (πE) | Scientists explore emerging topics identified by novelty index | μ(Novelty index, Xα,i, tα,i) and σ(Novelty index, Xα,i, tα,i) |
Exploration | Combinatorial innovation topic selection strategy (πC) | Scientists creatively combine dissimilar topics | μ(Innovation index, Xα,i, tα,i) and σ(Innovation index, Xα,i, tα,i) |
Note. Xα,i and tα,i denote the topics and publication time of a paper nα,i.
Although related studies have achieved fruitful results in understanding exploration and exploitation, to the best of our knowledge, no relevant studies have tried to quantify the exploration and exploitation strategy abilities of scholars. As mentioned, exploration and exploitation strategies have significantly different return and risk characteristics (Jia et al., 2017; Yin et al., 2019), and scholars show distinct preferences for them (Foster et al., 2015; Huang et al., 2022). This study employs five strategies presented by Huang et al. (2022), and proposes an explanatory model by which scholars’ proficiency in terms of different strategies can be decoupled and analyzed.
3. METHODOLOGY
3.1. Problem Definition
Previous studies demonstrate that the research ability of a scientist, α, directly or indirectly affects α’s research performance (Li, Zhang et al., 2022; Sinatra et al., 2016), and the research strategy, π, adopted by α also has an impact on his or her research performance (Foster et al., 2015; Huang et al., 2022). To our minds, a scholar may present different proficiency levels for distinct research strategies. For example, some scholars may be better at exploring new topics, while others may be better at consolidating mature topics. In this paper, the research ability of a scientist, α, denoted as Qα, is defined as a relatively stable underlying ability of α to publish high-quality papers by using their available knowledge (Huang et al., 2023b; Sinatra et al., 2016). High-quality papers are those that made contributions in a certain field and inspired subsequent research, thus usually with high scientific impact. The proficiency level of a research strategy, π, of α is defined as the research strategy ability, denoted as , which represents the ability of α to make good use of π. Hence, Qα is mainly related to individual ability, while is related to individual ability and the characteristics of the research strategy. Qα and jointly affect scholars’ research performance. Decoupling the relationship among them and quantifying them are interesting and meaningful research questions.
Specifically, the current study aims to untangle and analyze the impact of research ability (Qα) and research strategy ability () on α’s research performance, by which we provide a novel approach for evaluating scholars from exploration and exploitation and shed light on the benefits and risk features of different research strategies. Notably, Qα and are both seen as unobserved inherent attributes of α, and therefore the probabilistic graphical model is a suitable tool to analyze them.
3.2. Research Strategy Q Model
We propose the Research Strategy Q (RSQ) model based on the Q model proposed by Sinatra et al. (2016) and five research strategies (πM, πP, πD, πE, πC) proposed by Huang et al. (2022). The generative process of the RSQ model is to model the process of scholars publishing scientific papers through adopting different research strategies. Specifically, for each scientist, α, individual publication records, citation records, and research strategy execution records are generated in the following steps.
Step 1. For each scientist (α), α’s number of publications (Nα) is sampled from a Poisson distribution, Nα ∼ Poisson(λ), in which λ represents the average productivity of scholars in a specific field.
Step 2. For each scientist (α), α’s research ability (Qα) is sampled from a lognormal distribution, Qα ∼ lognormal(μQ, ), in which μQ represents the average research ability of scholars in a specific field.
Step 3. For each scientist (α), α’s research strategy ability on a research strategy, π (), is also sampled from a lognormal distribution, ∼ lognormal(μπ, ), in which μπ and σπ indicate the benefit and risk features of π. Five research strategies are studied, and therefore there are five , π ∈ π, π = {πM, πP, πD, πE, πC}.
Step 4. For each paper (i) authored by α, whether a research strategy, π, is executed or not is sampled from a Bernoulli distribution, ∼ Bernoulli(), in which indicates the probability of executing π. The Bernoulli distributions of different research strategies are independent of each other.
- Step 5. For each paper (i) authored by α, the quality of the paper (Cα,i) is simultaneously determined by Qα, , , and pi based on a multiplicative process, as shown in Eq. 1. pi indicates the random factor in i’s quality (i.e., luck), and is sampled from a lognormal distribution, pi ∼ lognormal(μP, ). represents the quality of the paper and is indirectly measured by its citations.(1)
The above generative process is shown in Figure 1. We assume that the productivity (Nα) follows a Poisson distribution, which is suitable for count data. The citations () follows a lognormal distribution, which has been proven to be reasonable in related studies (Huang et al., 2023b; Sinatra et al., 2016). The research strategy execution variable () is a binary variable, and therefore the Bernoulli distribution is employed. We follow the hypothesis of the Q model that the research ability (Qα) and the luck (pi) follow a lognormal distribution. To make Eq. 1 reasonable, research strategy ability () is also assumed to follow a lognormal distribution. Notably, five research strategies—mature topic selection strategy (πM), popular topic selection strategy (πP), diverse topic selection strategy (πD), emerging topic selection strategy (πE), and combinatorial innovation topic selection strategy (πC)—are adopted in the RSQ model. The definitions and quantitative indicators of these strategies have been clearly introduced in Table 1. In a nutshell, πE and πC respectively represent studying emerging topics and combinatorial innovation of topics, and are exploration strategies. πM, πP, and πD respectively represent studying mature topics, popular topics, and diversified combination of topics, and are exploitation strategies. The identification method proposed by Huang et al. (2022) is used to identify . In short, if the metrics of π calculated based on a paper, i, in Table 1 exceed the median value in the field, it indicates that π has been executed in i; that is, when equals 1, a scientist, α, publishes i by executing the research strategy, π. Different from the Q model, the quality of a paper is determined by Qα, , , and pi simultaneously, where and are newly introduced random variables. The joint distribution of the above generative process is shown in Eq. 2. Finally, consistent with Sinatra et al.’s (2016) experiments, we also utilize the number of citations of a paper 10 years after publication () to gauge its scientific impact, which indirectly measures its quality. Other metrics, such as PageRank score (Chen, Xie et al., 2007), novelty score (Luo, Lu et al., 2022), and F1000 (Bornmann & Haunschild, 2015) can also be employed to gauge .
In the RSQ model, there are the observed data (X), hidden variables (Z), and model parameters (θ). The observed data include Nα, Cα,i, and . The hidden variables include Qα, pi, and . The model parameters include λ, , μQ, σQ, μπ, σπ, μp, and σp. Importantly, Nα and are independent of the hidden variables, and can be seen as ancillary variables (Blei, Ng, & Jordan, 2003). Hence, their randomness can be ignored. λ and represent the average productivity of scientists in a specific field and α’s research strategy preference for π. λ, is not our focus. μQ, σQ indicate the average and standard deviation of research ability of scientists in a specific field. μπ, σπ suggest the benefits and risk features of π. μp, σp represent the randomness of the quality of a paper, in which μp is set to 0. Moreover, pi can be substituted by Eq. 1, and therefore Qα and are hidden variables that need to be inferred (gray zone in Figure 1). These variables and parameters are summarized in Table 2.
The notations and definitions utilized in the Research Strategy Q model
Type . | Notation . | Definition . |
---|---|---|
Model parameters | λ | The average productivity of scientists in a specific field |
Model parameters | μQ | The average of scientists’ research ability in a specific field |
Model parameters | σQ | The standard deviation of research ability in a specific field |
Model parameters | μπ | The average benefit of the research strategy, π, in a specific field |
Model parameters | σπ | The risk degree of the research strategy, π, in a specific field |
Model parameters | Scientist α’s probability of adopting π | |
Model parameters | μP | The average of randomness of the quality of papers |
Model parameters | σP | The standard deviation of randomness of the quality of papers |
Hidden variable | Qα | The research ability of a scientist, α |
Hidden variable | The research strategy ability of α on π | |
Hidden variable | pi | The randomness of the quality of a paper, i |
Observed data | Nα | The productivity of a scientist, α |
Observed data | Cα,i | The quality of a paper, i, authored by a scientist, α |
Observed data | The indicative variable denoting whether α executes π in i |
Type . | Notation . | Definition . |
---|---|---|
Model parameters | λ | The average productivity of scientists in a specific field |
Model parameters | μQ | The average of scientists’ research ability in a specific field |
Model parameters | σQ | The standard deviation of research ability in a specific field |
Model parameters | μπ | The average benefit of the research strategy, π, in a specific field |
Model parameters | σπ | The risk degree of the research strategy, π, in a specific field |
Model parameters | Scientist α’s probability of adopting π | |
Model parameters | μP | The average of randomness of the quality of papers |
Model parameters | σP | The standard deviation of randomness of the quality of papers |
Hidden variable | Qα | The research ability of a scientist, α |
Hidden variable | The research strategy ability of α on π | |
Hidden variable | pi | The randomness of the quality of a paper, i |
Observed data | Nα | The productivity of a scientist, α |
Observed data | Cα,i | The quality of a paper, i, authored by a scientist, α |
Observed data | The indicative variable denoting whether α executes π in i |
3.3. Inference and Parameter Estimation
To solve the RSQ model, we employ the BBVI-EM algorithm proposed by Huang et al. (2023b) to approximate the posterior distribution (Eq. 3). The BBVI-EM is a generic variational inference algorithm, in which the variational parameters (ϕ) and model parameters (θ) are iteratively updated by maximizing the Evidence Lower Bound (ELBO) of the log likelihood (Eq. 4).
The BBVI-EM algorithm for solving the Research Strategy Q model
Input: | |
The observed data X(Nα, Cα,i, ) | |
Initial model parameters ψ(μQ, σQ, μP, σP, μπ, σπ) | |
Initial variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α) | |
Maxitersgrad, ϵ1(ψ), ϵ2(ϕ), and MaxitersEM | |
Output: | |
Model parameters ψ(μQ, σQ, μP, σP, μπ, σπ) | |
Variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α) | |
1 | iters = 0; |
2 | While True |
3 | iters + = 1; E_iters = 0; M_iters = 0; |
4 | While True # E-step |
5 | E_iters+ = 1 |
6 | |
7 | |
8 | If ∣(μQ,α, σQ,α, μπ,α, σπ,α)m − (μQ,α, σQ,α, μπ,α, σπ,α)m−1∣ED < ϵ2orE_iters < Maxitersgradthen |
break | |
9 | While True # M-step |
10 | M_iters+ = 1 |
11 | |
12 | |
13 | If ∣(μQ, σQ, μπ, σπ, μP, σP)m − (μQ, σQ, μπ, σπ, μP, σP)m−1∣ED < ϵ1orM_iters < Maxitersgradthen |
Break | |
14 | If ∣(μ, σ, μP, σP, μπ, σπ)m − (μ, σ, μP, σP, μπ, σπ)m−1∣ED < ϵ1and |
oriters < MaxitersEMthen | |
Break | |
15 | End |
Input: | |
The observed data X(Nα, Cα,i, ) | |
Initial model parameters ψ(μQ, σQ, μP, σP, μπ, σπ) | |
Initial variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α) | |
Maxitersgrad, ϵ1(ψ), ϵ2(ϕ), and MaxitersEM | |
Output: | |
Model parameters ψ(μQ, σQ, μP, σP, μπ, σπ) | |
Variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α) | |
1 | iters = 0; |
2 | While True |
3 | iters + = 1; E_iters = 0; M_iters = 0; |
4 | While True # E-step |
5 | E_iters+ = 1 |
6 | |
7 | |
8 | If ∣(μQ,α, σQ,α, μπ,α, σπ,α)m − (μQ,α, σQ,α, μπ,α, σπ,α)m−1∣ED < ϵ2orE_iters < Maxitersgradthen |
break | |
9 | While True # M-step |
10 | M_iters+ = 1 |
11 | |
12 | |
13 | If ∣(μQ, σQ, μπ, σπ, μP, σP)m − (μQ, σQ, μπ, σπ, μP, σP)m−1∣ED < ϵ1orM_iters < Maxitersgradthen |
Break | |
14 | If ∣(μ, σ, μP, σP, μπ, σπ)m − (μ, σ, μP, σP, μπ, σπ)m−1∣ED < ϵ1and |
oriters < MaxitersEMthen | |
Break | |
15 | End |
Note. The superscript E denotes the Euclidean distance between two vectors.
4. DATA
4.1. Synthetic Data Generation
The actual Qα and of a scholar can never be known in advance, as they are unobserved individual abilities. Similar to our recent study (Huang et al., 2023b), we generate the synthetic data to evaluate whether the RSQ model can untangle the effect of the research ability and research strategy on the quality of papers and effectively quantify Qα and . To achieve this aim, the generative process of the RSQ model is used to create the synthetic data, by which the true value of the hidden variables and model parameters can be known in advance.
As shown in Table 3, we introduce three experimental setups for the synthetic data. In simulation 1, the number of authors (∣𝒜∣) is set to 500, and the average number of publications of scientists (λ) is set to 50. Thus, there are about 25,000 papers. For each scientist α ∈ 𝒜, Qα is sampled from LogNormal(0, 1). For each paper i ∈ 𝒩, pi is sampled from LogNormal(0, e−1). There is only one research strategy, π1, and is 0.5. Because μπ1 > 0, this indicates that adopting π1 tends to improve the quality of the paper . In simulation 2, in addition to π1, there is another research strategy, π2 with μπ2 < 0, which means that adopting π2 tends to stifle . Finally, in simulation 3, in addition to π1 and π2, the research strategy π3 with μπ3 = 0 is introduced. π3 tends to have a neutral impact on . The remaining configurations in simulation 2 and simulation 3 are consistent with those in simulation 1. In addition, we also generate synthetic data by sampling Qα and from student’s t distribution, the chi-square distribution, and the F distribution in our Supplementary material. Under each simulation setting, 10 groups of synthetic data are generated. Overall, simulation 1 is the simplest inferential task, with one strategy, while simulation 3 is the most difficult inferential task, with three strategies. The estimated Qα and through the BBVI-EM algorithm can be compared with their true value.
The experimental setup in the synthetic data generation
Parameter . | Simulation 1 . | Simulation 2 . | Simulation 3 . |
---|---|---|---|
∣𝒜∣ | 500 | 500 | 500 |
λ | 50 | 50 | 50 |
μQ | 0 | 0 | 0 |
σQ | e0 | e0 | e0 |
μπ1 | 2 | 2 | 2 |
σπ1 | e0 | e0 | e0 |
0.5 | 0.5 | 0.5 | |
μπ2 | \ | −2 | −2 |
σπ2 | \ | e0 | e0 |
\ | 0.5 | 0.5 | |
μπ3 | \ | \ | 0 |
σπ3 | \ | \ | e0 |
\ | \ | 0.5 | |
μP | 0 | 0 | 0 |
σP | e−1 | e−1 | e−1 |
Parameter . | Simulation 1 . | Simulation 2 . | Simulation 3 . |
---|---|---|---|
∣𝒜∣ | 500 | 500 | 500 |
λ | 50 | 50 | 50 |
μQ | 0 | 0 | 0 |
σQ | e0 | e0 | e0 |
μπ1 | 2 | 2 | 2 |
σπ1 | e0 | e0 | e0 |
0.5 | 0.5 | 0.5 | |
μπ2 | \ | −2 | −2 |
σπ2 | \ | e0 | e0 |
\ | 0.5 | 0.5 | |
μπ3 | \ | \ | 0 |
σπ3 | \ | \ | e0 |
\ | \ | 0.5 | |
μP | 0 | 0 | 0 |
σP | e−1 | e−1 | e−1 |
4.2. Empirical Data Collection
We employ the Microsoft Academic Graph (MAG)2 (Zhao, Bu, & Li, 2021) as our data source, which contains over 200 million papers as of 2020 and generates six levels of “field of study” (i.e., FoSL0, FoSL1, …, FoSL5) (Shen, Ma, & Wang, 2018). We extract papers in the physics, chemistry, and computer science fields to construct three empirical data sets.
Specifically, we use FoSL0 (“physics,” “chemistry,” and “computer science”) created in MAG to extract 11,714,663, 18,986,736, and 23,315,998 papers, and employ FoSL2 to represent research topics. Subsequently, we adopt the approach proposed by Huang et al. (2022) to respectively identify the five research strategies (πM, πP, πD, πE, πC). Taking a research policy, πD, as an example, two metrics, Average distance (Xα,i) and Average volume (Xα,i), of a paper, i, are reduced to a scalar, DSα,i, based on Principal Component Analysis (PCA). If DSα,i exceeds the median value of DS of papers in a specific field, πD is adopted in i ( = 1), and otherwise it is not adopted ( = 0). Hence, Nα, , and are observed data (i.e., empirical data), while Qα, are hidden variables, which we are interested in. The RSQ model connects the observed data with the hidden variables through Eq. 2. We choose scientists who published their first paper between 1990 and 2000 and have published at least 30 papers up till 2010 as our research objects. Hence, for each paper, the number of citations 10 years after publication can be computed in our empirical data sets. The statistics of the three empirical data sets are shown in Table 4. There are 26,992, 26,937, and 17,750 scientists in the physics, chemistry, and computer science fields, respectively. The adoption probability of π indicates /∣𝒩∣. In the following experiments, we separately conducted experiments on the three empirical data sets.
The empirical data used in the Research Strategy model
Statistics . | Physics . | Chemistry . | Computer Science . |
---|---|---|---|
Number of authors (∣𝒜∣) | 26,992 | 26,937 | 17,750 |
Number of papers (∣𝒩∣) | 773,814 | 1,087,593 | 816,881 |
Adoption probability of πM | 0.559 | 0.511 | 0.444 |
Adoption probability of πP | 0.812 | 0.643 | 0.753 |
Adoption probability of πD | 0.835 | 0.564 | 0.578 |
Adoption probability of πE | 0.752 | 0.576 | 0.814 |
Adoption probability of πC | 0.653 | 0.528 | 0.682 |
Time | 1990–2010 | 1990–2010 | 1990–2010 |
Statistics . | Physics . | Chemistry . | Computer Science . |
---|---|---|---|
Number of authors (∣𝒜∣) | 26,992 | 26,937 | 17,750 |
Number of papers (∣𝒩∣) | 773,814 | 1,087,593 | 816,881 |
Adoption probability of πM | 0.559 | 0.511 | 0.444 |
Adoption probability of πP | 0.812 | 0.643 | 0.753 |
Adoption probability of πD | 0.835 | 0.564 | 0.578 |
Adoption probability of πE | 0.752 | 0.576 | 0.814 |
Adoption probability of πC | 0.653 | 0.528 | 0.682 |
Time | 1990–2010 | 1990–2010 | 1990–2010 |
5. EXPERIMENTS AND RESULTS
5.1. Evaluation Metrics
5.2. Evaluating the Estimation Performance on the Synthetic Data
We use the synthetic data mentioned to evaluate the estimation performance of the RSQ model for quantifying Qα and . The variational parameters μQ,α and μπ,α obtained by the BBVI-EM algorithm are used to estimate Qα and , respectively. The Q model is employed as a baseline, in which the effect of research strategies is entirely neglected. As shown in Table 5, we report the average results on 10 groups of the synthetic data under each simulation setting. The Pearsonr, R2, RMSE, and MAE are utilized to evaluate the estimation performance.
The estimation performance of the Research Strategy Q model on the synthetic data
Simulation . | Metrics . | Q model . | Research Strategy Q model . | |
---|---|---|---|---|
Qα . | . | Qα . | ||
Simulation 1 | Pearsonr | 0.9840 | 0.9516 | 0.9973*** |
R2 | 0.8566 | 0.8966 | 0.9825*** | |
RMSE | 0.3804 | 0.3142 | 0.1224*** | |
MAE | 0.3379 | 0.2514 | 0.1064*** | |
Simulation 2 | Pearsonr | 0.9722 | 0.9507/0.9521 | 0.9961*** |
R2 | 0.9413 | 0.8749/0.8784 | 0.9860*** | |
RMSE | 0.2470 | 0.3443/0.3462 | 0.1163*** | |
MAE | 0.1963 | 0.2771/0.2749 | 0.0963*** | |
Simulation 3 | Pearsonr | 0.9576 | 0.9481/0.9475/0.9482 | 0.9942*** |
R2 | 0.9085 | 0.8613/0.8648/0.8615 | 0.9717*** | |
RMSE | 0.3012 | 0.3362/0.3595/0.3659 | 0.1632*** | |
MAE | 0.2383 | 0.2963/0.2884/0.2945 | 0.1370*** |
Simulation . | Metrics . | Q model . | Research Strategy Q model . | |
---|---|---|---|---|
Qα . | . | Qα . | ||
Simulation 1 | Pearsonr | 0.9840 | 0.9516 | 0.9973*** |
R2 | 0.8566 | 0.8966 | 0.9825*** | |
RMSE | 0.3804 | 0.3142 | 0.1224*** | |
MAE | 0.3379 | 0.2514 | 0.1064*** | |
Simulation 2 | Pearsonr | 0.9722 | 0.9507/0.9521 | 0.9961*** |
R2 | 0.9413 | 0.8749/0.8784 | 0.9860*** | |
RMSE | 0.2470 | 0.3443/0.3462 | 0.1163*** | |
MAE | 0.1963 | 0.2771/0.2749 | 0.0963*** | |
Simulation 3 | Pearsonr | 0.9576 | 0.9481/0.9475/0.9482 | 0.9942*** |
R2 | 0.9085 | 0.8613/0.8648/0.8615 | 0.9717*** | |
RMSE | 0.3012 | 0.3362/0.3595/0.3659 | 0.1632*** | |
MAE | 0.2383 | 0.2963/0.2884/0.2945 | 0.1370*** |
Indicates p < 0.05.
Indicates p < 0.01.
Indicates p < 0.001.
In simulation 1, only one research strategy, π1, is simulated. Both models give a satisfactory Pearsonr, which means that the estimated value and the real value are highly linearly correlated. However, the RSQ model gets a significantly better estimation performance than the Q model in terms of R2, RMSE, and MAE. The t-test is used to test the significant difference, and “***” indicates the significance level (p < 0.001). The bold text indicates the better results.
In simulation 2, two different research strategies, π1 and π2, are considered, and our model also gives satisfactory results. The estimated results of and are separated by “/”. For π1 and π2 with opposite characteristics, the RSQ model also effectively estimates Qα and , and has a significantly better estimation accuracy than the Q model in terms of all metrics.
In simulation 3, three different strategies, π1, π2, and π3, are considered. The RSQ model accurately quantifies and Qα. Specifically, Pearsonr and R2 of our model are close to 1, and are obviously greater than those of the Q model. RMSE and MAE of our model are significantly lower than those of the Q model. Notably, as the variety of strategies increases, the performance of the baseline obviously declines, while our model remains almost unchanged.
In our Supplementary material, we report more results on the synthetic data. Our results show that our model can untangle the effect of research ability and different research strategy ability on research performance and accurately estimate them in the synthetic data.
5.3. Examining the General Characteristics of Research Strategies
To disclose the general characteristics of the five research strategies (πM, πP, πD, πE, πC), we employ the RSQ model to examine three empirical data sets, and use the model parameters (μπ, σπ) to quantitatively analyze the benefit and risk features of a research strategy, π. As mentioned above, μπ represents the average impact of adopting π on the quality of a paper, , and σπ indicates the standard deviation of this impact. When μπ > 0, π tends to improve , while, when μπ < 0, π tends to stifle .
As shown in Table 6, we report the estimated benefits and risk features of πM, πP, πD, πE, πC, in three empirical data sets. For mature topic selection strategy (πM), in the physics, chemistry, and computer science fields, μπM is always less than 0, which suggests that studying mature topics is more likely to stifle scientists’ research performance. Indeed, a crucial evaluation criterion for scientific papers is novelty and originality. Mature topics have been thoroughly studied, and therefore limit an individual’s ability to produce breakthrough research (Huang et al., 2022; Liu et al., 2021). Notably, σπM gets a large value, which suggests that high-quality studies might also originate from mature topics, such as the revival of deep learning. For diverse topic selection strategy (πD) in the three fields, μπD is also less than 0. Studying topics with diverse knowledge also negatively affects scientists’ research performance. This may be because πD requires scholars to have rich knowledge reserves and its research process tends to be more challenging. For example, cross-disciplinary research is generally more time-consuming and requires handling high levels of complexity (Schaltegger, Beckmann, & Hansen, 2013; Xu, Ding, & Malic, 2015). For the popular topic selection strategy (πP) in the three fields, μπP always gets the maximum value, which shows that following the academic frontier is the most effective way to enhance academic impact. Moreover, for the emerging topic selection strategy (πE) and the combinatorial innovation topic selection strategy (πC), we find that both μπE and μπC are positive, which suggests that studying emerging topics and conducting combinatorial innovation tend to produce high-quality studies. Previous studies have shown that emerging topics represent research priorities and frontiers (Rotolo, Hicks, & Martin, 2015; Yang, Lu et al., 2022), and combination innovation is a scientific and reasonable innovation mode (Foster et al., 2015; Luo et al., 2022). Notably, σπE and σπC both take the larger value, which means that the exploration research strategy is a high-risk and high-yield strategy (Liu et al., 2021; Yu, Szymanski, & Jia, 2021).
The estimated benefit and risk features of the five research strategies
Strategy . | The benefit degree of π (μπ) . | The risk degree of π (σπ) . | ||||
---|---|---|---|---|---|---|
Physics . | Chemistry . | Computer Science . | Physics . | Chemistry . | Computer Science . | |
πM | −1.1793 | −1.2461 | −0.4073 | 1.3080 | 1.2375 | 1.2082 |
πP | 2.0366 | 2.6538 | 1.7841 | 1.4028 | 1.3592 | 1.2742 |
πD | −0.6053 | −0.2765 | −0.3215 | 1.2844 | 1.2363 | 1.2084 |
πE | 0.6733 | 0.8498 | 1.2041 | 1.3202 | 1.2498 | 1.3077 |
πC | 0.8435 | 1.2187 | 1.2179 | 1.3247 | 1.2553 | 1.2670 |
Strategy . | The benefit degree of π (μπ) . | The risk degree of π (σπ) . | ||||
---|---|---|---|---|---|---|
Physics . | Chemistry . | Computer Science . | Physics . | Chemistry . | Computer Science . | |
πM | −1.1793 | −1.2461 | −0.4073 | 1.3080 | 1.2375 | 1.2082 |
πP | 2.0366 | 2.6538 | 1.7841 | 1.4028 | 1.3592 | 1.2742 |
πD | −0.6053 | −0.2765 | −0.3215 | 1.2844 | 1.2363 | 1.2084 |
πE | 0.6733 | 0.8498 | 1.2041 | 1.3202 | 1.2498 | 1.3077 |
πC | 0.8435 | 1.2187 | 1.2179 | 1.3247 | 1.2553 | 1.2670 |
Note. Negative values are shown in bold.
In conclusion, πM and πD are weakened strategies with μπ < 0, which tend to stifle scientists’ research performance, while πP, πE, and πC are enhanced strategies with μπ > 0, which tend to improve scientists’ research performance. The model parameters of the RSQ model present a novel approach to disclose the unique characteristics of different research strategies, which helps researchers to understand the topic selection behavior from the perspective of exploration-exploitation.
5.4. Analyzing the Individual-Level Ability
The model parameters, μπ and σπ, reflect the overall characteristics of a research strategy, π, based on all scientists in the empirical data sets. However, for each scientist, α, the research strategy ability is a sample. α may have research strategies that α is good at or not good at. The RSQ model presents a new approach to evaluate scientists from the perspective of research ability (Qα) and research strategy ability ().
First, we employ Pearsonr to analyze the relationship among Qα, and some traditional author-level evaluation metrics (i.e., Nα, Cα, , Hα). Nα, Cα, , and Hα respectively denote the number of papers authored by α, total number of citations acquired by α’s papers, number of citations of α’s most-cited paper, and h-index of α. As shown in Figure 2, in the three fields, Pearsonr between any and Qα is lower than 0.6. This means that our model, to some extent, untangles the effect of and Qα on α’s research performance, and therefore and Qα measure the individual abilities in different aspects. In addition, there is a certain linear correlation among of different research strategies. To be specific, and are highly correlated (over 0.70), as they both measure the exploration strategy ability. has the lowest correlation with and (about 0.24 and 0.17), because it mainly measures the exploitation strategy ability. The low positive correlation also shows that exploration and exploitation are not contradictory but complementary from the perspective of proficiency (Zacher, Robinson, & Rosing, 2016). is highly correlated with and . This may be because scholars who are good at following research frontiers are better at innovating. There is a certain correlation between and other , as diverse topic knowledge involves various types of topics (Huang et al., 2022). Finally, Qα and Nα are almost unrelated. The Pearsonr values between Qα and Cα, , and Hα are only about 0.39, 0.30, and 0.40, which also shows that Qα can serve as a supplementary indicator for traditional metrics (Sinatra et al., 2016). Moreover, the five are also unrelated to Nα, and have a moderate correlation with Cα, , and Hα. Therefore, and Qα can be used as the complementary author-level metrics.
The Pearson correlation coefficient among Qα, , and traditional author-level metrics.
The Pearson correlation coefficient among Qα, , and traditional author-level metrics.
Subsequently, we also used Pearsonr to analyze the relationship among Qα, , and research strategy preference (), as shown in Figure 3. denotes the ratio of to Nα, in which indicates the number of times α adopts π in his or her Nα papers. As shown in Figure 4, we find an interesting result; that is, is nearly unrelated to . This means that scholars who prefer to execute a specific research strategy π may not be good at executing π, which further shows the unique attribute of . Qα is also unrelated to . In addition, consistent with our previous results (Huang et al., 2022), is negatively related to and in the three fields, which suggests that choosing between exploration and exploitation is an inevitable dilemma for scientists.
The case for clarifying the research strategy ability of scientists.
To further check the predictive power of Qα and , we conduct regression analysis, in which Qα and are employed as the independent variables, while traditional author-level metrics (Nα, Cα, , Hα) are respectively employed as the dependent variable. Specifically, in our regression analysis, we only use scientists’ papers published before 2000 to train the RSQ model, and use their papers published before 2010 to calculate Nα, Cα, , and Hα. Thus, this regression analysis examines the predictive ability of Qα and . As shown in Tables 7, 8, and 9, we report the regression coefficients in three empirical data sets. When log Nα (the logarithm of Nα) is the dependent variable, R2 is especially lower (less than 0.1), which also suggests that Qα and are linearly unrelated to future productivity of α. However, when logCα, log, and logHα are the dependent variables, the maximum value of R2 can reach 0.48. The coefficients of Qα and the five are nearly always significantly positive. This means that Qα and can be used to predict the future academic impact of α. The negative value of coefficients (e.g., the coefficient of ) is caused by the multicollinearity problem. As mentioned above, these research strategy abilities have a certain degree of linear correlation.
The regression coefficients in the physics field
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | −0.8017*** | 0.3351*** | 0.4469*** | 0.1153*** |
−0.0141*** | 0.0820*** | 0.0543*** | 0.0362*** | |
0.2565*** | 0.4145*** | 0.3158*** | 0.1962*** | |
−0.2549*** | −0.3278*** | −0.2673*** | −0.1662*** | |
0.1482*** | 0.4425*** | 0.3230*** | 0.2263*** | |
0.1740*** | 0.2440*** | 0.1697*** | 0.1167*** | |
R2 | 0.098 | 0.144 | 0.079 | 0.162 |
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | −0.8017*** | 0.3351*** | 0.4469*** | 0.1153*** |
−0.0141*** | 0.0820*** | 0.0543*** | 0.0362*** | |
0.2565*** | 0.4145*** | 0.3158*** | 0.1962*** | |
−0.2549*** | −0.3278*** | −0.2673*** | −0.1662*** | |
0.1482*** | 0.4425*** | 0.3230*** | 0.2263*** | |
0.1740*** | 0.2440*** | 0.1697*** | 0.1167*** | |
R2 | 0.098 | 0.144 | 0.079 | 0.162 |
Indicates p < 0.05.
Indicates p < 0.01.
Indicates p < 0.001.
The regression coefficients in the chemistry field
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | 0.3129*** | 2.2090*** | 1.4531*** | 1.0129*** |
−0.0026 | 0.1619*** | 0.0979*** | 0.0743*** | |
−0.0235*** | 0.0195 | 0.0656*** | 0.0195*** | |
−0.0631*** | −0.1490*** | −0.1098*** | −0.0672*** | |
−0.0866*** | −0.0984*** | −0.0480** | −0.0421*** | |
−0.0151* | 0.2825*** | 0.2347*** | 0.1298*** | |
R2 | 0.034 | 0.348 | 0.238 | 0.367 |
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | 0.3129*** | 2.2090*** | 1.4531*** | 1.0129*** |
−0.0026 | 0.1619*** | 0.0979*** | 0.0743*** | |
−0.0235*** | 0.0195 | 0.0656*** | 0.0195*** | |
−0.0631*** | −0.1490*** | −0.1098*** | −0.0672*** | |
−0.0866*** | −0.0984*** | −0.0480** | −0.0421*** | |
−0.0151* | 0.2825*** | 0.2347*** | 0.1298*** | |
R2 | 0.034 | 0.348 | 0.238 | 0.367 |
Indicates p < 0.05.
Indicates p < 0.01.
Indicates p < 0.001.
The regression coefficients in the computer science field
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | 0.0366 | 0.9250*** | 0.8078*** | 0.4054*** |
0.0021 | 0.1174*** | 0.0970*** | 0.0523*** | |
0.2036*** | 0.2507*** | 0.2102*** | 0.1069*** | |
−0.0181*** | 0.1389*** | 0.1228*** | 0.0602*** | |
0.0956*** | 0.6919*** | 0.5458*** | 0.3168*** | |
0.0139 | 0.3619*** | 0.3030*** | 0.1714*** | |
R2 | 0.076 | 0.463 | 0.345 | 0.480 |
Independent . | Dependent . | |||
---|---|---|---|---|
logNα . | logCα . | log . | logHα . | |
Qα | 0.0366 | 0.9250*** | 0.8078*** | 0.4054*** |
0.0021 | 0.1174*** | 0.0970*** | 0.0523*** | |
0.2036*** | 0.2507*** | 0.2102*** | 0.1069*** | |
−0.0181*** | 0.1389*** | 0.1228*** | 0.0602*** | |
0.0956*** | 0.6919*** | 0.5458*** | 0.3168*** | |
0.0139 | 0.3619*** | 0.3030*** | 0.1714*** | |
R2 | 0.076 | 0.463 | 0.345 | 0.480 |
Indicates p < 0.05.
Indicates p < 0.01.
Indicates p < 0.001.
Then we randomly select four case scientists (α1, α2, α3, α4) in each field, and display their in the radar chart, as shown in Figure 4. have been normalized by . As shown in the left subfigure of Figure 4, α4 (green) is especially good at executing πM and πD, while she or he is not good at πP, πE, and πC. Hence, compared to exploration, α4 is better at exploitation. In contrast to α4, α3 (purple) is good at πP, πE, and πC, while she or he is not adept at πM and πD. Thus, α3 is skilled at exploration. However, α1 (orange) presents a relatively balanced ability in terms of the five research strategies, and α1’s radar chart is like a pentagon. Hence, α1 can balance exploration and exploitation. Moreover, α3’s radar chart (red) is completely wrapped by α1’s radar chart, which means that α3’s abilities to conduct the five research strategies are weaker than those of α1. The values in parentheses of the legend represent the standardized Qα. We also find that Qα1 > Qα3. For the case study in the chemistry and computer science fields, we find similar results. In conclusion, one scholar may have distinct proficiency levels for different research strategies (e.g., ≠ ), while a number of scholars may have distinct proficiency levels for the same research strategy (e.g., ≠ ). Hence, can work as a novel metric to evaluate scholars from the perspective of exploration and exploitation.
Finally, we also present a case for guiding how to use and Qα. Taking scientists in the physics field as an example, as shown in Table 10, scholars are ranked according to Qα or in descending order. The governments, academic institutions, and funding agencies may refer to the ranking list to select scholars who are suitable for completing their research projects. For example, engineering projects requiring scholars to proficiently exploit mature knowledge can be handed over to scholars with a low rank of , , and , while scientific research projects requiring scholars to creatively explore new knowledge can be assigned to scholars with a low rank of and .
The ranking list of scientists in the physics field
Rank . | Qα . | . | . | . | . | . |
---|---|---|---|---|---|---|
1 | Matthew J. Holman (1.873) | Aniruddha R. Thakar (2.827) | R. Paul Butler (7.000) | David J. Schlegel (3.227) | A. Saggion (4.643) | D. A. Frail (5.417) |
2 | R. Paul Butler (1.811) | T. Ishii (2.826) | J. J. Engelen (6.122) | Jarle Brinchmann (2.498) | Guinevere Kauffmann (4.441) | Edo Berger (5.254) |
3 | Thomas Thiemann (1.808) | A. Kibayashi (2.771) | J. Allyn Smith (6.112) | Michael R. Blanton (2.463) | Adrian Pope (4.422) | Derek W. Fox (4.268) |
4 | Roberto Emparan (1.784) | C. Mauger (2.769) | James Annis (5.970) | Robert J. Brunner (2.444) | Paolo Ciliegi (4.330) | Paul A. Price (4.024) |
5 | Jonathan L. Feng (1.783) | E. Kearns (2.746) | Georg Weiglein (5.961) | Mariangela Bernardi (2.369) | Roberto Emparan (4.296) | Tyler L. Bourke (4.012) |
6 | A. A. Moiseev (1.778) | M. D. Messier (2.666) | Ralf Kaiser (5.946) | Darren Madgwick (2.326) | A. Robert (4.191) | Weidong Li (4.004) |
7 | Darren Madgwick (1.746) | A. L. Stachyra (2.593) | Douglas P. Finkbeiner (5.931) | Benjamin J. Weiner (2.194) | Idit Zehavi (4.133) | S. A. Yost (3.890) |
8 | David J. Schlegel (1.739) | E. Sharkey (2.563) | Gavin P. Salam (5.909) | Heidi Jo Newberg (2.173) | D. Kranich (4.029) | David Berenstein (3.800) |
9 | David Nesvorný (1.739) | B. Viren (2.441) | A. Kisselev (5.888) | Idit Zehavi (2.116) | G. Hinshaw (3.916) | Amihay Hanany (3.732) |
10 | Nima Arkani-Hamed (1.736) | M. Etoh (2.431) | P. Di Nezza (5.886) | Michael A. Carr (2.088) | Daniel C. Long (3.852) | Alicia M. Soderberg (3.724) |
Rank . | Qα . | . | . | . | . | . |
---|---|---|---|---|---|---|
1 | Matthew J. Holman (1.873) | Aniruddha R. Thakar (2.827) | R. Paul Butler (7.000) | David J. Schlegel (3.227) | A. Saggion (4.643) | D. A. Frail (5.417) |
2 | R. Paul Butler (1.811) | T. Ishii (2.826) | J. J. Engelen (6.122) | Jarle Brinchmann (2.498) | Guinevere Kauffmann (4.441) | Edo Berger (5.254) |
3 | Thomas Thiemann (1.808) | A. Kibayashi (2.771) | J. Allyn Smith (6.112) | Michael R. Blanton (2.463) | Adrian Pope (4.422) | Derek W. Fox (4.268) |
4 | Roberto Emparan (1.784) | C. Mauger (2.769) | James Annis (5.970) | Robert J. Brunner (2.444) | Paolo Ciliegi (4.330) | Paul A. Price (4.024) |
5 | Jonathan L. Feng (1.783) | E. Kearns (2.746) | Georg Weiglein (5.961) | Mariangela Bernardi (2.369) | Roberto Emparan (4.296) | Tyler L. Bourke (4.012) |
6 | A. A. Moiseev (1.778) | M. D. Messier (2.666) | Ralf Kaiser (5.946) | Darren Madgwick (2.326) | A. Robert (4.191) | Weidong Li (4.004) |
7 | Darren Madgwick (1.746) | A. L. Stachyra (2.593) | Douglas P. Finkbeiner (5.931) | Benjamin J. Weiner (2.194) | Idit Zehavi (4.133) | S. A. Yost (3.890) |
8 | David J. Schlegel (1.739) | E. Sharkey (2.563) | Gavin P. Salam (5.909) | Heidi Jo Newberg (2.173) | D. Kranich (4.029) | David Berenstein (3.800) |
9 | David Nesvorný (1.739) | B. Viren (2.441) | A. Kisselev (5.888) | Idit Zehavi (2.116) | G. Hinshaw (3.916) | Amihay Hanany (3.732) |
10 | Nima Arkani-Hamed (1.736) | M. Etoh (2.431) | P. Di Nezza (5.886) | Michael A. Carr (2.088) | Daniel C. Long (3.852) | Alicia M. Soderberg (3.724) |
6. DISCUSSION AND CONCLUSIONS
This study analyzes the impact of the research ability (Qα) and research strategy ability () on scientists’ research performance through a novel RSQ model, in which Qα indicates the fundamental ability of a scientist α, while describes the exploration and exploitation strategy ability of α. The BBVI-EM algorithm (Huang et al., 2023b) is utilized to solve the RSQ model. We use synthetic data and three empirical data sets in the physics, chemistry, and computer science fields to check the effectiveness of our model, and disclose the characteristics of five research strategies proposed by our previous study from the individual and overall perspectives. Overall, our model can not only be taken as a novel author-level evaluation tool, but it also sheds light on topic selection behavior in terms of exploration and exploitation. Here we provide a detailed summary of the theoretical and practical implications, as well as the limitations of our research.
6.1. Theoretical Implications
This article has the following theoretical implications. First, we propose a novel probabilistic graphical model, the RSQ model, which explain how scientists’ research ability and proficiency in different research strategies affect their research performance. Based on synthetic data and empirical data, we show that our model can untangle the relationships among research ability (Qα) and different research strategy abilities (), and effectively estimate them. Compared with the Q model (Sinatra et al., 2016), overlooking the impact of research strategies, our model achieves a better estimation performance of Qα, which suggests that the coupling relationships among Qα and should be considered.
Second, we employ the model parameters of the RSQ model to disclose the overall characteristics of five research strategies (πM, πP, πD, πE, πC) and shed light on the topic selection behavior from the perspective of exploration and exploitation. Specifically, based on our empirical data sets, our experimental results show that studying mature topics (πM) tends to stifle scientists from producing high-impact papers, because scientific research requires innovation. Studying diverse knowledge topics (πD) also stifles scientists’ performance, because diverse knowledge generally requires the handling of high levels of complexity (Schaltegger et al., 2013; Xu et al., 2015). However, following the academic frontier (πP) is the most effective way to enhance the academic impact. Exploration research strategies (πE and πC) are high risk and high yield (Liu et al., 2021; Yu et al., 2021). In conclusion, πM and πD are weakened strategies which tend to stifle scientists’ research performance, while πP, πE, and πC are enhanced strategies which tend to improve their research performance.
Third, we employ the variational parameters introduced by the BBVI-EM algorithm to analyze the author-level abilities through quantifying Qα and . We check the linear correlation among Qα, and some traditional author-level evaluation metrics. We find that Qα and are linearly unrelated to the productivity (Nα), and have a slight correlation with the accumulative citations (Cα), citations of the most-cited paper (), and the h-index (Hα), which means that they can be used as complementary metrics for author-level evaluation. We also analyze the relationship among Qα, , and , and find that scholars who prefer to execute a specific research strategy, π, may not necessarily be good at executing π. Importantly, our research thoroughly explains the relationship between exploration and exploitation; that is, exploration ability (, ) and exploitation ability (, , ) are positively correlated, indicating that scholars who are good at exploiting are often good at exploring, and vice versa. Indeed, only scholars who are good at exploring can understand and exploit cutting-edge knowledge, and only scholars who are good at exploiting have time to explore new knowledge. Therefore, exploration and exploitation are not contradictory but complementary (Zacher et al., 2016). However, the exploration preference (, ) and exploitation preference () are negatively correlated, which suggests that choosing between exploration and exploitation is an inevitable dilemma for scientists (Cohen et al., 2007). This is due to the limited attention of scholars, who need to make a choice between exploration and exploitation during their working hours. In addition, Qα and have predictive power for forecasting scholars’ future research performance. Finally, we show that one scholar may have sperate proficiency levels for different research strategies, and different scholars may have distinct proficiency levels for the same research strategy.
6.2. Practical Implications
This article also has the following practical implications. First, the research ability (Qα) and research strategy ability () of different research strategies quantified by the RSQ model can be used to evaluate scientists’ ability from different aspects. Qα evaluates the comprehensive abilities of a scholar to publish high-quality papers, while a series of evaluates the exploration and exploitation strategy ability of the scholar. Therefore, governments, academic institutions, and funding agencies may employ our proposed metrics to select excellent scholars, by which scientists with the greatest potential to produce path-breaking research can and should be rewarded according to their abilities (Madsen & Aagaard, 2020). For example, engineering projects can be assigned to scholars who are good at exploiting mature knowledge, while exploration projects can be handed over to scholars who are skilled at exploring new knowledge.
Second, five research strategies (πM, πP, πD, πE, πC) presented by Huang et al. (2022) are adopted in the RSQ model. Actually, any suitable definition of research strategies, such as a series of research strategies present by Foster et al. (2015), can also be adopted in our model. Hence, our model may be employed in other related studies to help researchers deeply understand exploration and exploitation behaviors. Overall, the current study not only sheds light on the characteristics of exploration and exploitation, but also proposes a new method and perspective for academic evaluation and performance assessment.
6.3. Limitations and Future Work
There are still some limitations in this study. First, in our correlation analysis and regression analysis, we observe a certain degree of linear correlation among the research ability and research strategy abilities. Therefore, a more sophisticated model may be proposed to better decouple and explain their complex coupling relationship. In addition, more factors influencing research performance (e.g., mentorship, team size, platform) should be included in our regression analysis. Second, our empirical data is collected from typical STEM disciplines. In the future, the empirical data in HSS disciplines and more sophisticated synthetic data need to be further analyzed to ensure the generalization performance of our model. Third, the RSQ model may be extended from static to dynamic to study the evolution of scientists’ abilities and strategies preferences. Finally, this paper employs citations to indirectly measure the quality of papers. The performance of our model based on other metrics measuring the quality of papers should be further explored.
AUTHOR CONTRIBUTIONS
Shengzhi Huang: Conceptualization, Formal analysis, Methodology, Writing—original draft. Wei Lu: Methodology, Supervision, Validation. Qikai Cheng: Data curation, Investigation. Li Zhang: Investigation, Validation. Yong Huang: Funding acquisition, Writing—review & editing.
COMPETING INTERESTS
The authors have no competing interests.
FUNDING INFORMATION
This work was supported by the Postdoctoral Fellowship Program of CPSF under grant number GZB20240565 and Youth Science Foundation of the National Natural Science Foundation of China (grant no. 72004168).
Notes
REFERENCES
Author notes
Handling Editor: Li Tang