This study quantifies and analyzes the individual-level abilities of scientists utilizing either an exploration or an exploitation strategy. Specifically, we present a Research Strategy Q model, which untangles the coupling effect of scientists’ research ability (Qα) and research strategy ability (Eαπ) on research performance. Qα indicates scientists’ fundamental ability to publish high-quality papers, while Eαπ indicates scientists’ proficiency in terms of exploration and exploitation strategies. Five research strategies proposed by our previous study are employed. We generate synthetic data and collect empirical data as our experimental data set. We show that these research strategies present different benefit and risk characteristics. Adopting some exploitation strategies tends to stifle research performance, while exploration strategies are high risk and high yield. Qα and Eαπ have predictive power for research performance. Moreover, we find that, first, scholars who prefer to execute a research strategy, π, may not necessarily be better at executing π. Second, some scholars have differences in their abilities towards different strategies, while other scholars have differences in their abilities towards the same strategy. Third, exploration and exploitation are not contradictory but complementary from the perspective of proficiency, while they are mutually exclusive from the perspective of selection preference.

Nowadays, an increasing number of scholars are devoting themselves to scientific research and publish a huge number of papers yearly. Evaluating and understanding scholars’ academic performance is a crucial research direction in Science of Science (Zhou, Wang et al., 2020), which serves academic contexts, such as grant applications, performance evaluation, and personnel promotion (Wang, Zhou, & Zeng, 2023). Fair and diversified scientific evaluation is necessary for promoting scientific progress (Ain, Riaz, & Afzal, 2019; Sinatra, Wang et al., 2016). The exploration and exploitation dilemma inevitably faced by scholars makes exploration-exploitation the crucial dimension by which to evaluate individual scholars (Yang, Hu et al., 2023).

With the development of big data analysis techniques, data-driven bibliometrics-based evaluation methods have become a powerful supplement to peer review. Bibliometrics-based methods mainly analyze scientists’ research performance from number of citations and papers, citation and coauthorship network, and academic texts. However, traditional author-level evaluation metrics generally evaluate scholars from a single-dimensional view of their impact (Zhang, Wang et al., 2023), cannot measure the innovation level of scientists (Petersen, Riccaboni et al., 2012), and neglect the research focus of scholars. Recent studies show that the exploration-exploitation research strategies present significantly different return and risk characteristics (Jia, Wang, & Szymanski, 2017; Yin, Wang et al., 2019), and scholars show distinct preferences for them (Foster, Rzhetsky, & Evans, 2015; Huang, Lu et al., 2022). Inspired by these studies, this study aims to address the following research issues:

  1. propose an explanatory model to untangle the effect of individual-level ability and five research strategies proposed by Huang et al. (2022) on research performance;

  2. quantitatively examine the return and risk characteristics of these research strategies; and

  3. comparatively analyze scientists’ abilities towards different strategies.

In short, the goal of our study is to provide a deeper understanding of topic selection behavior and present a new method and perspective for academic evaluation and performance assessment.

In this paper, we propose a Research Strategy Q model (RSQ) that untangles and quantifies the effect of scientists’ research ability (Qα) and research strategy ability (Eαπ) on their research performance. Qα is defined as the underlying fundamental ability of a scientist to publish high-quality papers (Huang, Lu et al., 2023b; Sinatra et al., 2016), while Eαπ is defined as the scientist’s proficiency in exploration and exploitation strategies. The five research strategies proposed by Huang et al. (2022) are adopted. We generate synthetic data and separately select about 20,000 scholars in the physics, chemistry, and computer science fields as our data sets. Our results show that our model can effectively uncouple the distinct abilities of a scientist and disclose that different research strategies have obviously distinct benefits and risk features. Moreover, a scholar may have separate proficiency levels for different research strategies (e.g., EαπMEαπE), while a number of scholars may have separate proficiency levels for the same specific research strategy (e.g., Eα1πMEα2πM). Importantly, exploration and exploitation are not contradictory but complementary from the perspective of proficiency, while they are mutually exclusive from the perspective of selection preference. Our code is open and accessible1.

This study has the following theoretical and practical implications. The RSQ model is presented to explain how individual research ability and proficiency in five research strategies affect scientists’ research performance. On the one hand, the RSQ model discloses the unique characteristics of different research strategies, and sheds light on the topic selection behavior from the perspective of exploration and exploitation. On the other hand, the RSQ model can be utilized as a novel tool to evaluate scientists. Governments, academic institutions, and funding agencies may employ our model to evaluate scholars, by which they can improve their talent assessment mechanism.

The rest of this paper is organized as follows. In Section 2, we review related studies. In Section 3, we present the problem definition and our model. In Section 4, we clarify our data set. In Section 5, we provide the analysis of experimental results. In Section 6, we conclude with theoretical and practical implications, limitations, and points for future research.

2.1. Reviews on Scholar Evaluation

Peer review and bibliometrics-based methods are two mainstream methods for evaluating scholars’ research performance. In peer review, scholars and their publications are manually evaluated by domain experts or peers based on their knowledge and experience. This subjective method is effective in evaluating small numbers of scientists. However, with more and more scholars devoting themselves to scientific research and the increasing volume of publications, evaluation purely based on peer review is not feasible because it is both time and effort consuming (Bornmann & Daniel, 2009a). In addition, peer review is inevitably influenced by subjective factors, such as cognitive distortions (Protasiewicz, Pedrycz et al., 2016). Hence, peer review is often combined with bibliometrics-based methods.

Bibliometrics-based methods generally evaluate the productivity and impact of scholars based on their bibliometric information, such as citation count, number of papers, h-index, and variants of the h-index (Ain et al., 2019). The h-index, proposed by Hirsch (2005), simultaneously measures the productivity and impact of scholars, and shows a certain degree of consistency with peer review (Johnson & Lovegrove, 2008). However, the h-index suffers from some obvious problems. For example, it cannot completely reflect the citation distribution of a scientist’s papers (Bornmann & Daniel, 2009b); it is affected by self-citation (Bartneck & Kokkelmans, 2011); it puts new scholars at a disadvantage; and it is insensitive to changes in research performance (Rousseau & Leuven, 2008). To solve these issues, several variants of the h-index have been proposed. For instance, Egghe (2006) proposed the g-index, defined as the largest number such that the top g papers received at least g2 citations. The g-index is more able to take into account highly cited papers. Alonso, Cabrerizo et al. (2010) computed the hg-index by taking the geometric mean of the h-index and g-index. Schreiber (2018) developed a timely h-index, ht, that only considers a scientist’s publication and citation records of the previous t years. ht alleviates the time bias caused by the dominance of rather old papers in the h-index. Khurana and Sharma (2022) presented hc by adding the weight of the highest cited paper to the h-index, which improves the evaluation performance for low-ranked scholars. Ain et al. (2019) examined the differences between the h-index and its citation intensity-based variants, and found that a high correlation exists among most of them. However, these metrics overlook the impact of research topics and strategies on scientists’ research performance.

PageRank-based algorithms have been utilized to rank scientists, in which citation networks and coauthorship networks are constructed. For example, Sidiropoulos and Manolopoulos (2006) ranked scholars based on the average PageRank score of their papers in the citation network. Gao, Wang et al. (2016) combined the h-index and PageRank algorithm to present the PR-index, which provides a more balanced impact measure than existing indices. Yan and Ding (2011) introduced a weighted PageRank algorithm based on a coauthorship network to measure author impact. Zhang et al. (2023) proposed a RelRank algorithm based on the PageRank algorithm, in which a coauthorship network consisting of the author–venue relationship is developed. RelRank can identify high-impact scholars who are devoted to specific publication outlets. However, the above author-level evaluation indicators still evaluate scientists from a single-dimensional view of their impact (Zhang et al., 2023) and cannot evaluate exploration-exploitation performance of scientists.

Sinatra et al. (2016) proposed a novel Q model, which uncouples the effects of productivity, research ability, and luck on scholars’ research performance. The Q parameter estimated by the probabilistic model is a time-invariant indicator that can truly account for a scientist’s individual ability. However, the Q model neglects the effect of research topics. Yang et al. (2023) presented the disruptive h-index and consolidating h-index based on disruption metrics (Funk & Owen-Smith, 2017) to comprehensively evaluate the disruptive and consolidating impact of scientists. However, the disruptive/consolidating h-index is hard to interpret, because the h-index simply combines two numbers of two distinct units. Inspired by their studies, I employ an interpretable probabilistic graphical model to evaluate the exploration-exploitation performance of scholars.

2.2. Reviews on Exploration and Exploitation

The conflicting demands between productive tradition and risky innovation bring an “essential tension” to scientists (Kuhn, 1977), and scientists inevitably face a dilemma between exploration and exploitation (Cohen, McClure, & Yu, 2007). Consolidating studies (i.e., exploitation) tend to be more easily accepted by the current scientist evaluation system, and therefore have a greater probability of being published (Jia et al., 2017). Innovating studies (i.e., exploration) generally have extremely high time cost and sunk cost, but are more likely to have greater impact in the long run (Yin et al., 2019). Many studies have studied the choice of exploration and exploitation in terms of scientists’ topic selection behavior.

Jia et al. (2017) disclosed three fundamental properties dominating changes in research interests (i.e., heterogeneity, recency, and subject proximity), and also proposed a random-walk-based to reproduce the empirical observations, by which they shed light on the interplay between exploitation and exploration. Huang, Huang et al. (2023a) developed a random walk-based model based on reinforcement learning theory to examine the role of productivity and impact in the evolution of scientists’ research interests, and found that scientists prefer to select topics that help them improve their productivity and impact. Franco, Malhotra, and Simonovits (2014) showed that scholars prefer to submit papers that reject the null hypothesis, to ensure publication. Liu, Wang et al. (2018) and Liu, Dehmamy et al. (2021) examined the career histories of artists, film directors, and scientists, and showed that successful individuals generally undergo extensive exploration before engaging in sufficient exploitation. Zeng, Shen et al. (2019) analyzed the topic-switching dynamics of physicists, and showed that they shift more frequently among topics than they did in the past. Chen and Ding (2023) analyzed the publication records of 117 Nobel laureates in physics, and found that they generally explore about three topics alternately in different periods and focus on exploiting their core topics throughout their careers. Foster et al. (2015) introduced five research strategies (i.e., jump, new consolidation, repeat consolidation, new bridge, repeat bridge) based on chemical entities networks. They demonstrated that the aggregate distribution of research strategies remains extremely stable, and risky innovation strategies are especially rare. Recently, Huang et al. (2022) proposed five novel research strategies under exploration and exploitation behavior; that is, mature topic selection strategy (πM), popular topic selection strategy (πP), diverse topic selection strategy (πD), emerging topic selection strategy (πE), and combinatorial innovation topic selection strategy (πC). They also presented a series of metrics to identify the research strategies adopted by a paper. They found that impactful scientists frequently execute πD, πP, πE, πC but execute πM less often. The five strategies and corresponding metrics are summarized in Table 1.

Table 1.

The five research strategies proposed by Huang et al. (2022) 

TypeStrategyDefinitionMetrics
Exploitation Mature topic selection strategy (πMScientists consolidate mature topics that have been extensively studied −logP1(Xα,i; tα,i) and −logP2(Xα,i; tα,i
Exploitation Popular topic selection strategy (πPScientists follow hot topics that have recently received widespread attention Productivity ratio (Xα,i, tα,i) and Impact ratio (Xα,i, tα,i
Exploitation Diverse topic selection strategy (πDScientists study topics that cover diverse knowledge in semantic space Average distance (Xα,i) and Average volume (Xα,i
Exploration Emerging topic selection strategy (πEScientists explore emerging topics identified by novelty index μ(Novelty index, Xα,i, tα,i) and σ(Novelty index, Xα,i, tα,i
Exploration Combinatorial innovation topic selection strategy (πCScientists creatively combine dissimilar topics μ(Innovation index, Xα,i, tα,i) and σ(Innovation index, Xα,i, tα,i
TypeStrategyDefinitionMetrics
Exploitation Mature topic selection strategy (πMScientists consolidate mature topics that have been extensively studied −logP1(Xα,i; tα,i) and −logP2(Xα,i; tα,i
Exploitation Popular topic selection strategy (πPScientists follow hot topics that have recently received widespread attention Productivity ratio (Xα,i, tα,i) and Impact ratio (Xα,i, tα,i
Exploitation Diverse topic selection strategy (πDScientists study topics that cover diverse knowledge in semantic space Average distance (Xα,i) and Average volume (Xα,i
Exploration Emerging topic selection strategy (πEScientists explore emerging topics identified by novelty index μ(Novelty index, Xα,i, tα,i) and σ(Novelty index, Xα,i, tα,i
Exploration Combinatorial innovation topic selection strategy (πCScientists creatively combine dissimilar topics μ(Innovation index, Xα,i, tα,i) and σ(Innovation index, Xα,i, tα,i

Note. Xα,i and tα,i denote the topics and publication time of a paper nα,i.

Although related studies have achieved fruitful results in understanding exploration and exploitation, to the best of our knowledge, no relevant studies have tried to quantify the exploration and exploitation strategy abilities of scholars. As mentioned, exploration and exploitation strategies have significantly different return and risk characteristics (Jia et al., 2017; Yin et al., 2019), and scholars show distinct preferences for them (Foster et al., 2015; Huang et al., 2022). This study employs five strategies presented by Huang et al. (2022), and proposes an explanatory model by which scholars’ proficiency in terms of different strategies can be decoupled and analyzed.

3.1. Problem Definition

Previous studies demonstrate that the research ability of a scientist, α, directly or indirectly affects α’s research performance (Li, Zhang et al., 2022; Sinatra et al., 2016), and the research strategy, π, adopted by α also has an impact on his or her research performance (Foster et al., 2015; Huang et al., 2022). To our minds, a scholar may present different proficiency levels for distinct research strategies. For example, some scholars may be better at exploring new topics, while others may be better at consolidating mature topics. In this paper, the research ability of a scientist, α, denoted as Qα, is defined as a relatively stable underlying ability of α to publish high-quality papers by using their available knowledge (Huang et al., 2023b; Sinatra et al., 2016). High-quality papers are those that made contributions in a certain field and inspired subsequent research, thus usually with high scientific impact. The proficiency level of a research strategy, π, of α is defined as the research strategy ability, denoted as Eαπ, which represents the ability of α to make good use of π. Hence, Qα is mainly related to individual ability, while Eαπ is related to individual ability and the characteristics of the research strategy. Qα and Eαπ jointly affect scholars’ research performance. Decoupling the relationship among them and quantifying them are interesting and meaningful research questions.

Specifically, the current study aims to untangle and analyze the impact of research ability (Qα) and research strategy ability (Eαπ) on α’s research performance, by which we provide a novel approach for evaluating scholars from exploration and exploitation and shed light on the benefits and risk features of different research strategies. Notably, Qα and Eαπ are both seen as unobserved inherent attributes of α, and therefore the probabilistic graphical model is a suitable tool to analyze them.

3.2. Research Strategy Q Model

We propose the Research Strategy Q (RSQ) model based on the Q model proposed by Sinatra et al. (2016) and five research strategies (πM, πP, πD, πE, πC) proposed by Huang et al. (2022). The generative process of the RSQ model is to model the process of scholars publishing scientific papers through adopting different research strategies. Specifically, for each scientist, α, individual publication records, citation records, and research strategy execution records are generated in the following steps.

  • Step 1. For each scientist (α), α’s number of publications (Nα) is sampled from a Poisson distribution, Nα ∼ Poisson(λ), in which λ represents the average productivity of scholars in a specific field.

  • Step 2. For each scientist (α), α’s research ability (Qα) is sampled from a lognormal distribution, Qα ∼ lognormal(μQ, σQ2), in which μQ represents the average research ability of scholars in a specific field.

  • Step 3. For each scientist (α), α’s research strategy ability on a research strategy, π (Eαπ), is also sampled from a lognormal distribution, Eαπ ∼ lognormal(μπ, σπ2), in which μπ and σπ indicate the benefit and risk features of π. Five research strategies are studied, and therefore there are five Eαπ, ππ, π = {πM, πP, πD, πE, πC}.

  • Step 4. For each paper (i) authored by α, whether a research strategy, π, is executed or not is sampled from a Bernoulli distribution, 1iπ ∼ Bernoulli(pαπ), in which pαπ indicates the probability of executing π. The Bernoulli distributions of different research strategies are independent of each other.

  • Step 5. For each paper (i) authored by α, the quality of the paper (Cα,i) is simultaneously determined by Qα, Eαπ, 1iπ, and pi based on a multiplicative process, as shown in Eq. 1. pi indicates the random factor in i’s quality (i.e., luck), and is sampled from a lognormal distribution, pi ∼ lognormal(μP, σP2). Cαi represents the quality of the paper and is indirectly measured by its citations.
    (1)

The above generative process is shown in Figure 1. We assume that the productivity (Nα) follows a Poisson distribution, which is suitable for count data. The citations (Cαi) follows a lognormal distribution, which has been proven to be reasonable in related studies (Huang et al., 2023b; Sinatra et al., 2016). The research strategy execution variable (1iπ) is a binary variable, and therefore the Bernoulli distribution is employed. We follow the hypothesis of the Q model that the research ability (Qα) and the luck (pi) follow a lognormal distribution. To make Eq. 1 reasonable, research strategy ability (Eαπ) is also assumed to follow a lognormal distribution. Notably, five research strategies—mature topic selection strategy (πM), popular topic selection strategy (πP), diverse topic selection strategy (πD), emerging topic selection strategy (πE), and combinatorial innovation topic selection strategy (πC)—are adopted in the RSQ model. The definitions and quantitative indicators of these strategies have been clearly introduced in Table 1. In a nutshell, πE and πC respectively represent studying emerging topics and combinatorial innovation of topics, and are exploration strategies. πM, πP, and πD respectively represent studying mature topics, popular topics, and diversified combination of topics, and are exploitation strategies. The identification method proposed by Huang et al. (2022) is used to identify 1iπ. In short, if the metrics of π calculated based on a paper, i, in Table 1 exceed the median value in the field, it indicates that π has been executed in i; that is, when 1iπ equals 1, a scientist, α, publishes i by executing the research strategy, π. Different from the Q model, the quality of a paper is determined by Qα, Eαπ, 1iπ, and pi simultaneously, where Eαπ and 1iπ are newly introduced random variables. The joint distribution of the above generative process is shown in Eq. 2. Finally, consistent with Sinatra et al.’s (2016) experiments, we also utilize the number of citations of a paper 10 years after publication (Cαi) to gauge its scientific impact, which indirectly measures its quality. Other metrics, such as PageRank score (Chen, Xie et al., 2007), novelty score (Luo, Lu et al., 2022), and F1000 (Bornmann & Haunschild, 2015) can also be employed to gauge Cαi.

Figure 1.

The generative process of the Research Strategy Q model.

Figure 1.

The generative process of the Research Strategy Q model.

Close modal

In the RSQ model, there are the observed data (X), hidden variables (Z), and model parameters (θ). The observed data include Nα, Cα,i, and 1iπ. The hidden variables include Qα, pi, and Eαπ. The model parameters include λ, pαπ, μQ, σQ, μπ, σπ, μp, and σp. Importantly, Nα and 1iπ are independent of the hidden variables, and can be seen as ancillary variables (Blei, Ng, & Jordan, 2003). Hence, their randomness can be ignored. λ and pαπ represent the average productivity of scientists in a specific field and α’s research strategy preference for π. λ, pαπ is not our focus. μQ, σQ indicate the average and standard deviation of research ability of scientists in a specific field. μπ, σπ suggest the benefits and risk features of π. μp, σp represent the randomness of the quality of a paper, in which μp is set to 0. Moreover, pi can be substituted by Eq. 1, and therefore Qα and Eαπ are hidden variables that need to be inferred (gray zone in Figure 1). These variables and parameters are summarized in Table 2.

Table 2.

The notations and definitions utilized in the Research Strategy Q model

TypeNotationDefinition
Model parameters λ The average productivity of scientists in a specific field 
Model parameters μQ The average of scientists’ research ability in a specific field 
Model parameters σQ The standard deviation of research ability in a specific field 
Model parameters μπ The average benefit of the research strategy, π, in a specific field 
Model parameters σπ The risk degree of the research strategy, π, in a specific field 
Model parameters pαπ Scientist α’s probability of adopting π 
Model parameters μP The average of randomness of the quality of papers 
Model parameters σP The standard deviation of randomness of the quality of papers 
Hidden variable Qα The research ability of a scientist, α 
Hidden variable Eαπ The research strategy ability of α on π 
Hidden variable pi The randomness of the quality of a paper, i 
Observed data Nα The productivity of a scientist, α 
Observed data Cα,i The quality of a paper, i, authored by a scientist, α 
Observed data 1iπ The indicative variable denoting whether α executes π in i 
TypeNotationDefinition
Model parameters λ The average productivity of scientists in a specific field 
Model parameters μQ The average of scientists’ research ability in a specific field 
Model parameters σQ The standard deviation of research ability in a specific field 
Model parameters μπ The average benefit of the research strategy, π, in a specific field 
Model parameters σπ The risk degree of the research strategy, π, in a specific field 
Model parameters pαπ Scientist α’s probability of adopting π 
Model parameters μP The average of randomness of the quality of papers 
Model parameters σP The standard deviation of randomness of the quality of papers 
Hidden variable Qα The research ability of a scientist, α 
Hidden variable Eαπ The research strategy ability of α on π 
Hidden variable pi The randomness of the quality of a paper, i 
Observed data Nα The productivity of a scientist, α 
Observed data Cα,i The quality of a paper, i, authored by a scientist, α 
Observed data 1iπ The indicative variable denoting whether α executes π in i 
The inferential problem of the RSQ model is to compute the posterior distribution of the hidden Qα and Eαπ given the observed Nα, Cα,i, and 1iπ, as shown in Eq. 3. The BBVI-EM algorithm (Huang et al., 2023b) is employed to solve our model.
(2)
(3)

3.3. Inference and Parameter Estimation

To solve the RSQ model, we employ the BBVI-EM algorithm proposed by Huang et al. (2023b) to approximate the posterior distribution (Eq. 3). The BBVI-EM is a generic variational inference algorithm, in which the variational parameters (ϕ) and model parameters (θ) are iteratively updated by maximizing the Evidence Lower Bound (ELBO) of the log likelihood (Eq. 4).

For each scientist α, to infer his or her research ability (Qα) and five research strategy abilities (Eαπ, ππ, π = {πM, πP, πD, πE, πC}), six variational distributions are introduced. Specifically, for α ∈ 𝒜, we employ a lognormal(μQ,α, σQ,α2) to approximate the posterior of Qα. For each ππ, we respectively employ a lognormal (μπ,α, σπ,α2) to approximate the posterior of Eαπ. Hence, μQ,α and μπ,α can be used to estimate Qα and Eαπ, respectively. μQ,α, σQ,α, μπ,α, σπ,α are variational parameters (ϕ), while μQ, σQ, μπ, σπ, μP, σP are model parameters (θ). The variational parameters can be used to evaluate individual-level characteristics, while the model parameters can be used to evaluate overall characteristics in a specific field. The BBVI-EM algorithm is to iteratively update ϕ and θ based on stochastic optimization of ELBO, as shown in Algorithm 1. The 4th to 8th rows of Algorithm 1 show the E-step of BBVI-EM, by which ϕ is updated with fixed θ. The 9th to 13th rows of Algorithm 1 show the M-step of BBVI-EM, by which θ is updated with fixed ϕ. The 14th row is the convergence condition. For a more detailed introduction to the BBVI-EM algorithm, please review our previous research.
(4)
Algorithm 1.

The BBVI-EM algorithm for solving the Research Strategy Q model

Input
The observed data X(Nα, Cα,i, 1iπ
Initial model parameters ψ(μQ, σQ, μP, σP, μπ, σπ
Initial variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α
Maxitersgrad, ϵ1(ψ), ϵ2(ϕ), and MaxitersEM 
  
Output
Model parameters ψ(μQ, σQ, μP, σP, μπ, σπ
Variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α
  
iters = 0; 
While True 
  iters + = 1; E_iters = 0; M_iters = 0; 
  While True # E-step 
   E_iters+ = 1 
𝓛μQ,ασQ,αμπ,ασπ,α=1Ss=1SlogqQα,sEαπμQ,ασQ,αμπ,ασπ,αlogPNαCα,i1iπQα,sEαπμQσQμπσπμPσPlogqQα,sEαπμQ,ασQ,αμπ,ασπ,α 
μQ,ασQ,αμπ,ασπ,αm=μQ,ασQ,αμπ,ασπ,αm1+ρ𝓛μQ,ασQ,αμπ,ασπ,αm1 
  If ∣(μQ,α, σQ,α, μπ,α, σπ,α)m − (μQ,α, σQ,α, μπ,α, σπ,α)m−1ED < ϵ2orE_iters < Maxitersgradthen 
   break 
  While True # M-step 
10    M_iters+ = 1 
11 𝓛μQσQμπσπμPσP=1Ss=1SqQα,sEαπμQ,ασQ,αμαπσαπlogPNαCα,i1iπQα,sEαπμQσQμπσπμPσP 
12 μσμPσPμπσπm=μσμPσPμπσπm1+ρ𝓛μσμPσPμπσπm1 
13   If ∣(μQ, σQ, μπ, σπ, μP, σP)m − (μQ, σQ, μπ, σπ, μP, σP)m−1ED < ϵ1orM_iters < Maxitersgradthen 
Break 
14   If ∣(μ, σ, μP, σP, μπ, σπ)m − (μ, σ, μP, σP, μπ, σπ)m−1ED < ϵ1andμασαμαπσαπmμασαμαπσαπm1ED<ϵ2 
  oriters < MaxitersEMthen 
   Break 
15 End 
Input
The observed data X(Nα, Cα,i, 1iπ
Initial model parameters ψ(μQ, σQ, μP, σP, μπ, σπ
Initial variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α
Maxitersgrad, ϵ1(ψ), ϵ2(ϕ), and MaxitersEM 
  
Output
Model parameters ψ(μQ, σQ, μP, σP, μπ, σπ
Variational parameters ϕ(μQ,α, σQ,α, μπ,α, σπ,α
  
iters = 0; 
While True 
  iters + = 1; E_iters = 0; M_iters = 0; 
  While True # E-step 
   E_iters+ = 1 
𝓛μQ,ασQ,αμπ,ασπ,α=1Ss=1SlogqQα,sEαπμQ,ασQ,αμπ,ασπ,αlogPNαCα,i1iπQα,sEαπμQσQμπσπμPσPlogqQα,sEαπμQ,ασQ,αμπ,ασπ,α 
μQ,ασQ,αμπ,ασπ,αm=μQ,ασQ,αμπ,ασπ,αm1+ρ𝓛μQ,ασQ,αμπ,ασπ,αm1 
  If ∣(μQ,α, σQ,α, μπ,α, σπ,α)m − (μQ,α, σQ,α, μπ,α, σπ,α)m−1ED < ϵ2orE_iters < Maxitersgradthen 
   break 
  While True # M-step 
10    M_iters+ = 1 
11 𝓛μQσQμπσπμPσP=1Ss=1SqQα,sEαπμQ,ασQ,αμαπσαπlogPNαCα,i1iπQα,sEαπμQσQμπσπμPσP 
12 μσμPσPμπσπm=μσμPσPμπσπm1+ρ𝓛μσμPσPμπσπm1 
13   If ∣(μQ, σQ, μπ, σπ, μP, σP)m − (μQ, σQ, μπ, σπ, μP, σP)m−1ED < ϵ1orM_iters < Maxitersgradthen 
Break 
14   If ∣(μ, σ, μP, σP, μπ, σπ)m − (μ, σ, μP, σP, μπ, σπ)m−1ED < ϵ1andμασαμαπσαπmμασαμαπσαπm1ED<ϵ2 
  oriters < MaxitersEMthen 
   Break 
15 End 

Note. The superscript E denotes the Euclidean distance between two vectors.

4.1. Synthetic Data Generation

The actual Qα and Eαπ of a scholar can never be known in advance, as they are unobserved individual abilities. Similar to our recent study (Huang et al., 2023b), we generate the synthetic data to evaluate whether the RSQ model can untangle the effect of the research ability and research strategy on the quality of papers and effectively quantify Qα and Eαπ. To achieve this aim, the generative process of the RSQ model is used to create the synthetic data, by which the true value of the hidden variables and model parameters can be known in advance.

As shown in Table 3, we introduce three experimental setups for the synthetic data. In simulation 1, the number of authors (∣𝒜∣) is set to 500, and the average number of publications of scientists (λ) is set to 50. Thus, there are about 25,000 papers. For each scientist α ∈ 𝒜, Qα is sampled from LogNormal(0, 1). For each paper i ∈ 𝒩, pi is sampled from LogNormal(0, e−1). There is only one research strategy, π1, and pαπ1 is 0.5. Because μπ1 > 0, this indicates that adopting π1 tends to improve the quality of the paper Cαi. In simulation 2, in addition to π1, there is another research strategy, π2 with μπ2 < 0, which means that adopting π2 tends to stifle Cαi. Finally, in simulation 3, in addition to π1 and π2, the research strategy π3 with μπ3 = 0 is introduced. π3 tends to have a neutral impact on Cαi. The remaining configurations in simulation 2 and simulation 3 are consistent with those in simulation 1. In addition, we also generate synthetic data by sampling Qα and Eαπ from student’s t distribution, the chi-square distribution, and the F distribution in our Supplementary material. Under each simulation setting, 10 groups of synthetic data are generated. Overall, simulation 1 is the simplest inferential task, with one strategy, while simulation 3 is the most difficult inferential task, with three strategies. The estimated Qα and Eαπ through the BBVI-EM algorithm can be compared with their true value.

Table 3.

The experimental setup in the synthetic data generation

ParameterSimulation 1Simulation 2Simulation 3
∣𝒜∣ 500 500 500 
λ 50 50 50 
μQ 
σQ e0 e0 e0 
μπ1 
σπ1 e0 e0 e0 
pαπ1 0.5 0.5 0.5 
μπ2 −2 −2 
σπ2 e0 e0 
pαπ2 0.5 0.5 
μπ3 
σπ3 e0 
pαπ3 0.5 
μP 
σP e−1 e−1 e−1 
ParameterSimulation 1Simulation 2Simulation 3
∣𝒜∣ 500 500 500 
λ 50 50 50 
μQ 
σQ e0 e0 e0 
μπ1 
σπ1 e0 e0 e0 
pαπ1 0.5 0.5 0.5 
μπ2 −2 −2 
σπ2 e0 e0 
pαπ2 0.5 0.5 
μπ3 
σπ3 e0 
pαπ3 0.5 
μP 
σP e−1 e−1 e−1 

4.2. Empirical Data Collection

We employ the Microsoft Academic Graph (MAG)2 (Zhao, Bu, & Li, 2021) as our data source, which contains over 200 million papers as of 2020 and generates six levels of “field of study” (i.e., FoSL0, FoSL1, …, FoSL5) (Shen, Ma, & Wang, 2018). We extract papers in the physics, chemistry, and computer science fields to construct three empirical data sets.

Specifically, we use FoSL0 (“physics,” “chemistry,” and “computer science”) created in MAG to extract 11,714,663, 18,986,736, and 23,315,998 papers, and employ FoSL2 to represent research topics. Subsequently, we adopt the approach proposed by Huang et al. (2022) to respectively identify the five research strategies (πM, πP, πD, πE, πC). Taking a research policy, πD, as an example, two metrics, Average distance (Xα,i) and Average volume (Xα,i), of a paper, i, are reduced to a scalar, DSα,i, based on Principal Component Analysis (PCA). If DSα,i exceeds the median value of DS of papers in a specific field, πD is adopted in i (1iπ = 1), and otherwise it is not adopted (1iπ = 0). Hence, Nα, Cαi, and 1iπ are observed data (i.e., empirical data), while Qα, Eαπ are hidden variables, which we are interested in. The RSQ model connects the observed data with the hidden variables through Eq. 2. We choose scientists who published their first paper between 1990 and 2000 and have published at least 30 papers up till 2010 as our research objects. Hence, for each paper, the number of citations 10 years after publication can be computed in our empirical data sets. The statistics of the three empirical data sets are shown in Table 4. There are 26,992, 26,937, and 17,750 scientists in the physics, chemistry, and computer science fields, respectively. The adoption probability of π indicates i𝒩1iπ/∣𝒩∣. In the following experiments, we separately conducted experiments on the three empirical data sets.

Table 4.

The empirical data used in the Research Strategy model

StatisticsPhysicsChemistryComputer Science
Number of authors (∣𝒜∣) 26,992 26,937 17,750 
Number of papers (∣𝒩∣) 773,814 1,087,593 816,881 
Adoption probability of πM 0.559 0.511 0.444 
Adoption probability of πP 0.812 0.643 0.753 
Adoption probability of πD 0.835 0.564 0.578 
Adoption probability of πE 0.752 0.576 0.814 
Adoption probability of πC 0.653 0.528 0.682 
Time 1990–2010 1990–2010 1990–2010 
StatisticsPhysicsChemistryComputer Science
Number of authors (∣𝒜∣) 26,992 26,937 17,750 
Number of papers (∣𝒩∣) 773,814 1,087,593 816,881 
Adoption probability of πM 0.559 0.511 0.444 
Adoption probability of πP 0.812 0.643 0.753 
Adoption probability of πD 0.835 0.564 0.578 
Adoption probability of πE 0.752 0.576 0.814 
Adoption probability of πC 0.653 0.528 0.682 
Time 1990–2010 1990–2010 1990–2010 

5.1. Evaluation Metrics

In our following experiments, first, we check whether our model can effectively untangle the coupling relationship among Qα and five Eαπ, and accurately quantify them on the synthetic data. To achieve this aim, we employ the Pearson correlation coefficient (Pearsonr), the goodness of fit (R2), root mean square error (RMSE), and mean absolute error (MAE) to evaluate the linear correlation and the difference between the predicted value and real value. The mathematical equations are shown in Eqs. 58. The larger values of Pearsonr and R2 as well as the smaller values of RMSE and MAE indicate better estimation performance. In addition, Pearsonr is also employed to check the linear correlation among Qα, Eαπ and traditional author-level evaluation metrics. R2 is also employed to analyze the results of our regression analysis.
(5)
(6)
(7)
(8)

5.2. Evaluating the Estimation Performance on the Synthetic Data

We use the synthetic data mentioned to evaluate the estimation performance of the RSQ model for quantifying Qα and Eαπ. The variational parameters μQ,α and μπ,α obtained by the BBVI-EM algorithm are used to estimate Qα and Eαπ, respectively. The Q model is employed as a baseline, in which the effect of research strategies is entirely neglected. As shown in Table 5, we report the average results on 10 groups of the synthetic data under each simulation setting. The Pearsonr, R2, RMSE, and MAE are utilized to evaluate the estimation performance.

Table 5.

The estimation performance of the Research Strategy Q model on the synthetic data

SimulationMetricsQ modelResearch Strategy Q model
QαEαπQα
Simulation 1 Pearsonr 0.9840 0.9516 0.9973*** 
R2 0.8566 0.8966 0.9825*** 
RMSE 0.3804 0.3142 0.1224*** 
MAE 0.3379 0.2514 0.1064*** 
  
Simulation 2 Pearsonr 0.9722 0.9507/0.9521 0.9961*** 
R2 0.9413 0.8749/0.8784 0.9860*** 
RMSE 0.2470 0.3443/0.3462 0.1163*** 
MAE 0.1963 0.2771/0.2749 0.0963*** 
  
Simulation 3 Pearsonr 0.9576 0.9481/0.9475/0.9482 0.9942*** 
R2 0.9085 0.8613/0.8648/0.8615 0.9717*** 
RMSE 0.3012 0.3362/0.3595/0.3659 0.1632*** 
MAE 0.2383 0.2963/0.2884/0.2945 0.1370*** 
SimulationMetricsQ modelResearch Strategy Q model
QαEαπQα
Simulation 1 Pearsonr 0.9840 0.9516 0.9973*** 
R2 0.8566 0.8966 0.9825*** 
RMSE 0.3804 0.3142 0.1224*** 
MAE 0.3379 0.2514 0.1064*** 
  
Simulation 2 Pearsonr 0.9722 0.9507/0.9521 0.9961*** 
R2 0.9413 0.8749/0.8784 0.9860*** 
RMSE 0.2470 0.3443/0.3462 0.1163*** 
MAE 0.1963 0.2771/0.2749 0.0963*** 
  
Simulation 3 Pearsonr 0.9576 0.9481/0.9475/0.9482 0.9942*** 
R2 0.9085 0.8613/0.8648/0.8615 0.9717*** 
RMSE 0.3012 0.3362/0.3595/0.3659 0.1632*** 
MAE 0.2383 0.2963/0.2884/0.2945 0.1370*** 
*

Indicates p < 0.05.

**

Indicates p < 0.01.

***

Indicates p < 0.001.

In simulation 1, only one research strategy, π1, is simulated. Both models give a satisfactory Pearsonr, which means that the estimated value and the real value are highly linearly correlated. However, the RSQ model gets a significantly better estimation performance than the Q model in terms of R2, RMSE, and MAE. The t-test is used to test the significant difference, and “***” indicates the significance level (p < 0.001). The bold text indicates the better results.

In simulation 2, two different research strategies, π1 and π2, are considered, and our model also gives satisfactory results. The estimated results of Eαπ1 and Eαπ2 are separated by “/”. For π1 and π2 with opposite characteristics, the RSQ model also effectively estimates Qα and Eαπ, and has a significantly better estimation accuracy than the Q model in terms of all metrics.

In simulation 3, three different strategies, π1, π2, and π3, are considered. The RSQ model accurately quantifies Eαπ and Qα. Specifically, Pearsonr and R2 of our model are close to 1, and are obviously greater than those of the Q model. RMSE and MAE of our model are significantly lower than those of the Q model. Notably, as the variety of strategies increases, the performance of the baseline obviously declines, while our model remains almost unchanged.

In our Supplementary material, we report more results on the synthetic data. Our results show that our model can untangle the effect of research ability and different research strategy ability on research performance and accurately estimate them in the synthetic data.

5.3. Examining the General Characteristics of Research Strategies

To disclose the general characteristics of the five research strategies (πM, πP, πD, πE, πC), we employ the RSQ model to examine three empirical data sets, and use the model parameters (μπ, σπ) to quantitatively analyze the benefit and risk features of a research strategy, π. As mentioned above, μπ represents the average impact of adopting π on the quality of a paper, Cαi, and σπ indicates the standard deviation of this impact. When μπ > 0, π tends to improve Cαi, while, when μπ < 0, π tends to stifle Cαi.

As shown in Table 6, we report the estimated benefits and risk features of πM, πP, πD, πE, πC, in three empirical data sets. For mature topic selection strategy (πM), in the physics, chemistry, and computer science fields, μπM is always less than 0, which suggests that studying mature topics is more likely to stifle scientists’ research performance. Indeed, a crucial evaluation criterion for scientific papers is novelty and originality. Mature topics have been thoroughly studied, and therefore limit an individual’s ability to produce breakthrough research (Huang et al., 2022; Liu et al., 2021). Notably, σπM gets a large value, which suggests that high-quality studies might also originate from mature topics, such as the revival of deep learning. For diverse topic selection strategy (πD) in the three fields, μπD is also less than 0. Studying topics with diverse knowledge also negatively affects scientists’ research performance. This may be because πD requires scholars to have rich knowledge reserves and its research process tends to be more challenging. For example, cross-disciplinary research is generally more time-consuming and requires handling high levels of complexity (Schaltegger, Beckmann, & Hansen, 2013; Xu, Ding, & Malic, 2015). For the popular topic selection strategy (πP) in the three fields, μπP always gets the maximum value, which shows that following the academic frontier is the most effective way to enhance academic impact. Moreover, for the emerging topic selection strategy (πE) and the combinatorial innovation topic selection strategy (πC), we find that both μπE and μπC are positive, which suggests that studying emerging topics and conducting combinatorial innovation tend to produce high-quality studies. Previous studies have shown that emerging topics represent research priorities and frontiers (Rotolo, Hicks, & Martin, 2015; Yang, Lu et al., 2022), and combination innovation is a scientific and reasonable innovation mode (Foster et al., 2015; Luo et al., 2022). Notably, σπE and σπC both take the larger value, which means that the exploration research strategy is a high-risk and high-yield strategy (Liu et al., 2021; Yu, Szymanski, & Jia, 2021).

Table 6.

The estimated benefit and risk features of the five research strategies

StrategyThe benefit degree of π (μπ)The risk degree of π (σπ)
PhysicsChemistryComputer SciencePhysicsChemistryComputer Science
πM −1.1793 −1.2461 −0.4073 1.3080 1.2375 1.2082 
πP 2.0366 2.6538 1.7841 1.4028 1.3592 1.2742 
πD −0.6053 −0.2765 −0.3215 1.2844 1.2363 1.2084 
πE 0.6733 0.8498 1.2041 1.3202 1.2498 1.3077 
πC 0.8435 1.2187 1.2179 1.3247 1.2553 1.2670 
StrategyThe benefit degree of π (μπ)The risk degree of π (σπ)
PhysicsChemistryComputer SciencePhysicsChemistryComputer Science
πM −1.1793 −1.2461 −0.4073 1.3080 1.2375 1.2082 
πP 2.0366 2.6538 1.7841 1.4028 1.3592 1.2742 
πD −0.6053 −0.2765 −0.3215 1.2844 1.2363 1.2084 
πE 0.6733 0.8498 1.2041 1.3202 1.2498 1.3077 
πC 0.8435 1.2187 1.2179 1.3247 1.2553 1.2670 

Note. Negative values are shown in bold.

In conclusion, πM and πD are weakened strategies with μπ < 0, which tend to stifle scientists’ research performance, while πP, πE, and πC are enhanced strategies with μπ > 0, which tend to improve scientists’ research performance. The model parameters of the RSQ model present a novel approach to disclose the unique characteristics of different research strategies, which helps researchers to understand the topic selection behavior from the perspective of exploration-exploitation.

5.4. Analyzing the Individual-Level Ability

The model parameters, μπ and σπ, reflect the overall characteristics of a research strategy, π, based on all scientists in the empirical data sets. However, for each scientist, α, the research strategy ability Eαπ is a sample. α may have research strategies that α is good at or not good at. The RSQ model presents a new approach to evaluate scientists from the perspective of research ability (Qα) and research strategy ability (Eαπ).

First, we employ Pearsonr to analyze the relationship among Qα, Eαπ and some traditional author-level evaluation metrics (i.e., Nα, Cα, Cα*, Hα). Nα, Cα, Cα*, and Hα respectively denote the number of papers authored by α, total number of citations acquired by α’s papers, number of citations of α’s most-cited paper, and h-index of α. As shown in Figure 2, in the three fields, Pearsonr between any Eαπ and Qα is lower than 0.6. This means that our model, to some extent, untangles the effect of Eαπ and Qα on α’s research performance, and therefore Eαπ and Qα measure the individual abilities in different aspects. In addition, there is a certain linear correlation among Eαπ of different research strategies. To be specific, EαπE and EαπC are highly correlated (over 0.70), as they both measure the exploration strategy ability. EαπM has the lowest correlation with EαπE and EαπC (about 0.24 and 0.17), because it mainly measures the exploitation strategy ability. The low positive correlation also shows that exploration and exploitation are not contradictory but complementary from the perspective of proficiency (Zacher, Robinson, & Rosing, 2016). EαπP is highly correlated with EαπE and EαπC. This may be because scholars who are good at following research frontiers are better at innovating. There is a certain correlation between EαπD and other Eαπ, as diverse topic knowledge involves various types of topics (Huang et al., 2022). Finally, Qα and Nα are almost unrelated. The Pearsonr values between Qα and Cα, Cα*, and Hα are only about 0.39, 0.30, and 0.40, which also shows that Qα can serve as a supplementary indicator for traditional metrics (Sinatra et al., 2016). Moreover, the five Eαπ are also unrelated to Nα, and have a moderate correlation with Cα, Cα*, and Hα. Therefore, Eαπ and Qα can be used as the complementary author-level metrics.

Figure 2.

The Pearson correlation coefficient among Qα, Eαπ, and traditional author-level metrics.

Figure 2.

The Pearson correlation coefficient among Qα, Eαπ, and traditional author-level metrics.

Close modal

Subsequently, we also used Pearsonr to analyze the relationship among Qα, Eαπ, and research strategy preference (pαπ), as shown in Figure 3. pαπ denotes the ratio of Nα,π+ to Nα, in which Nα,π+ indicates the number of times α adopts π in his or her Nα papers. As shown in Figure 4, we find an interesting result; that is, Eαπ is nearly unrelated to pαπ. This means that scholars who prefer to execute a specific research strategy π may not be good at executing π, which further shows the unique attribute of Eαπ. Qα is also unrelated to pαπ. In addition, consistent with our previous results (Huang et al., 2022), pαπM is negatively related to pαπE and pαπC in the three fields, which suggests that choosing between exploration and exploitation is an inevitable dilemma for scientists.

Figure 3.

The Pearson correlation coefficient among Qα, Eαπ, and pαπ.

Figure 3.

The Pearson correlation coefficient among Qα, Eαπ, and pαπ.

Close modal
Figure 4.

The case for clarifying the research strategy ability of scientists.

Figure 4.

The case for clarifying the research strategy ability of scientists.

Close modal

To further check the predictive power of Qα and Eαπ, we conduct regression analysis, in which Qα and Eαπ are employed as the independent variables, while traditional author-level metrics (Nα, Cα, Cα*, Hα) are respectively employed as the dependent variable. Specifically, in our regression analysis, we only use scientists’ papers published before 2000 to train the RSQ model, and use their papers published before 2010 to calculate Nα, Cα, Cα*, and Hα. Thus, this regression analysis examines the predictive ability of Qα and Eαπ. As shown in Tables 7, 8, and 9, we report the regression coefficients in three empirical data sets. When log Nα (the logarithm of Nα) is the dependent variable, R2 is especially lower (less than 0.1), which also suggests that Qα and Eαπ are linearly unrelated to future productivity of α. However, when logCα, logCα*, and logHα are the dependent variables, the maximum value of R2 can reach 0.48. The coefficients of Qα and the five Eαπ are nearly always significantly positive. This means that Qα and Eαπ can be used to predict the future academic impact of α. The negative value of coefficients (e.g., the coefficient of EαπD) is caused by the multicollinearity problem. As mentioned above, these research strategy abilities have a certain degree of linear correlation.

Table 7.

The regression coefficients in the physics field

IndependentDependent
logNαlogCαlogCα*logHα
Qα −0.8017*** 0.3351*** 0.4469*** 0.1153*** 
EαπM −0.0141*** 0.0820*** 0.0543*** 0.0362*** 
EαπP 0.2565*** 0.4145*** 0.3158*** 0.1962*** 
EαπD −0.2549*** −0.3278*** −0.2673*** −0.1662*** 
EαπE 0.1482*** 0.4425*** 0.3230*** 0.2263*** 
EαπC 0.1740*** 0.2440*** 0.1697*** 0.1167*** 
  
R2 0.098 0.144 0.079 0.162 
IndependentDependent
logNαlogCαlogCα*logHα
Qα −0.8017*** 0.3351*** 0.4469*** 0.1153*** 
EαπM −0.0141*** 0.0820*** 0.0543*** 0.0362*** 
EαπP 0.2565*** 0.4145*** 0.3158*** 0.1962*** 
EαπD −0.2549*** −0.3278*** −0.2673*** −0.1662*** 
EαπE 0.1482*** 0.4425*** 0.3230*** 0.2263*** 
EαπC 0.1740*** 0.2440*** 0.1697*** 0.1167*** 
  
R2 0.098 0.144 0.079 0.162 
*

Indicates p < 0.05.

**

Indicates p < 0.01.

***

Indicates p < 0.001.

Table 8.

The regression coefficients in the chemistry field

IndependentDependent
logNαlogCαlogCα*logHα
Qα 0.3129*** 2.2090*** 1.4531*** 1.0129*** 
EαπM −0.0026 0.1619*** 0.0979*** 0.0743*** 
EαπP −0.0235*** 0.0195 0.0656*** 0.0195*** 
EαπD −0.0631*** −0.1490*** −0.1098*** −0.0672*** 
EαπE −0.0866*** −0.0984*** −0.0480** −0.0421*** 
EαπC −0.0151* 0.2825*** 0.2347*** 0.1298*** 
  
R2 0.034 0.348 0.238 0.367 
IndependentDependent
logNαlogCαlogCα*logHα
Qα 0.3129*** 2.2090*** 1.4531*** 1.0129*** 
EαπM −0.0026 0.1619*** 0.0979*** 0.0743*** 
EαπP −0.0235*** 0.0195 0.0656*** 0.0195*** 
EαπD −0.0631*** −0.1490*** −0.1098*** −0.0672*** 
EαπE −0.0866*** −0.0984*** −0.0480** −0.0421*** 
EαπC −0.0151* 0.2825*** 0.2347*** 0.1298*** 
  
R2 0.034 0.348 0.238 0.367 
*

Indicates p < 0.05.

**

Indicates p < 0.01.

***

Indicates p < 0.001.

Table 9.

The regression coefficients in the computer science field

IndependentDependent
logNαlogCαlogCα*logHα
Qα 0.0366 0.9250*** 0.8078*** 0.4054*** 
EαπM 0.0021 0.1174*** 0.0970*** 0.0523*** 
EαπP 0.2036*** 0.2507*** 0.2102*** 0.1069*** 
EαπD −0.0181*** 0.1389*** 0.1228*** 0.0602*** 
EαπE 0.0956*** 0.6919*** 0.5458*** 0.3168*** 
EαπC 0.0139 0.3619*** 0.3030*** 0.1714*** 
  
R2 0.076 0.463 0.345 0.480 
IndependentDependent
logNαlogCαlogCα*logHα
Qα 0.0366 0.9250*** 0.8078*** 0.4054*** 
EαπM 0.0021 0.1174*** 0.0970*** 0.0523*** 
EαπP 0.2036*** 0.2507*** 0.2102*** 0.1069*** 
EαπD −0.0181*** 0.1389*** 0.1228*** 0.0602*** 
EαπE 0.0956*** 0.6919*** 0.5458*** 0.3168*** 
EαπC 0.0139 0.3619*** 0.3030*** 0.1714*** 
  
R2 0.076 0.463 0.345 0.480 
*

Indicates p < 0.05.

**

Indicates p < 0.01.

***

Indicates p < 0.001.

Then we randomly select four case scientists (α1, α2, α3, α4) in each field, and display their Eαπ in the radar chart, as shown in Figure 4. Eαπ have been normalized by EαπminEαπ/maxEαπminEαπ. As shown in the left subfigure of Figure 4, α4 (green) is especially good at executing πM and πD, while she or he is not good at πP, πE, and πC. Hence, compared to exploration, α4 is better at exploitation. In contrast to α4, α3 (purple) is good at πP, πE, and πC, while she or he is not adept at πM and πD. Thus, α3 is skilled at exploration. However, α1 (orange) presents a relatively balanced ability in terms of the five research strategies, and α1’s radar chart is like a pentagon. Hence, α1 can balance exploration and exploitation. Moreover, α3’s radar chart (red) is completely wrapped by α1’s radar chart, which means that α3’s abilities to conduct the five research strategies are weaker than those of α1. The values in parentheses of the legend represent the standardized Qα. We also find that Qα1 > Qα3. For the case study in the chemistry and computer science fields, we find similar results. In conclusion, one scholar may have distinct proficiency levels for different research strategies (e.g., EαπMEαπE), while a number of scholars may have distinct proficiency levels for the same research strategy (e.g., Eα1πMEα2πM). Hence, Eαπ can work as a novel metric to evaluate scholars from the perspective of exploration and exploitation.

Finally, we also present a case for guiding how to use Eαπ and Qα. Taking scientists in the physics field as an example, as shown in Table 10, scholars are ranked according to Qα or Eαπ in descending order. The governments, academic institutions, and funding agencies may refer to the ranking list to select scholars who are suitable for completing their research projects. For example, engineering projects requiring scholars to proficiently exploit mature knowledge can be handed over to scholars with a low rank of EαπM, EαπP, and EαπD, while scientific research projects requiring scholars to creatively explore new knowledge can be assigned to scholars with a low rank of EαπE and EαπC.

Table 10.

The ranking list of scientists in the physics field

RankQαEαπMEαπPEαπDEαπEEαπC
Matthew J. Holman (1.873) Aniruddha R. Thakar (2.827) R. Paul Butler (7.000) David J. Schlegel (3.227) A. Saggion (4.643) D. A. Frail (5.417) 
R. Paul Butler (1.811) T. Ishii (2.826) J. J. Engelen (6.122) Jarle Brinchmann (2.498) Guinevere Kauffmann (4.441) Edo Berger (5.254) 
Thomas Thiemann (1.808) A. Kibayashi (2.771) J. Allyn Smith (6.112) Michael R. Blanton (2.463) Adrian Pope (4.422) Derek W. Fox (4.268) 
Roberto Emparan (1.784) C. Mauger (2.769) James Annis (5.970) Robert J. Brunner (2.444) Paolo Ciliegi (4.330) Paul A. Price (4.024) 
Jonathan L. Feng (1.783) E. Kearns (2.746) Georg Weiglein (5.961) Mariangela Bernardi (2.369) Roberto Emparan (4.296) Tyler L. Bourke (4.012) 
A. A. Moiseev (1.778) M. D. Messier (2.666) Ralf Kaiser (5.946) Darren Madgwick (2.326) A. Robert (4.191) Weidong Li (4.004) 
Darren Madgwick (1.746) A. L. Stachyra (2.593) Douglas P. Finkbeiner (5.931) Benjamin J. Weiner (2.194) Idit Zehavi (4.133) S. A. Yost (3.890) 
David J. Schlegel (1.739) E. Sharkey (2.563) Gavin P. Salam (5.909) Heidi Jo Newberg (2.173) D. Kranich (4.029) David Berenstein (3.800) 
David Nesvorný (1.739) B. Viren (2.441) A. Kisselev (5.888) Idit Zehavi (2.116) G. Hinshaw (3.916) Amihay Hanany (3.732) 
10 Nima Arkani-Hamed (1.736) M. Etoh (2.431) P. Di Nezza (5.886) Michael A. Carr (2.088) Daniel C. Long (3.852) Alicia M. Soderberg (3.724) 
RankQαEαπMEαπPEαπDEαπEEαπC
Matthew J. Holman (1.873) Aniruddha R. Thakar (2.827) R. Paul Butler (7.000) David J. Schlegel (3.227) A. Saggion (4.643) D. A. Frail (5.417) 
R. Paul Butler (1.811) T. Ishii (2.826) J. J. Engelen (6.122) Jarle Brinchmann (2.498) Guinevere Kauffmann (4.441) Edo Berger (5.254) 
Thomas Thiemann (1.808) A. Kibayashi (2.771) J. Allyn Smith (6.112) Michael R. Blanton (2.463) Adrian Pope (4.422) Derek W. Fox (4.268) 
Roberto Emparan (1.784) C. Mauger (2.769) James Annis (5.970) Robert J. Brunner (2.444) Paolo Ciliegi (4.330) Paul A. Price (4.024) 
Jonathan L. Feng (1.783) E. Kearns (2.746) Georg Weiglein (5.961) Mariangela Bernardi (2.369) Roberto Emparan (4.296) Tyler L. Bourke (4.012) 
A. A. Moiseev (1.778) M. D. Messier (2.666) Ralf Kaiser (5.946) Darren Madgwick (2.326) A. Robert (4.191) Weidong Li (4.004) 
Darren Madgwick (1.746) A. L. Stachyra (2.593) Douglas P. Finkbeiner (5.931) Benjamin J. Weiner (2.194) Idit Zehavi (4.133) S. A. Yost (3.890) 
David J. Schlegel (1.739) E. Sharkey (2.563) Gavin P. Salam (5.909) Heidi Jo Newberg (2.173) D. Kranich (4.029) David Berenstein (3.800) 
David Nesvorný (1.739) B. Viren (2.441) A. Kisselev (5.888) Idit Zehavi (2.116) G. Hinshaw (3.916) Amihay Hanany (3.732) 
10 Nima Arkani-Hamed (1.736) M. Etoh (2.431) P. Di Nezza (5.886) Michael A. Carr (2.088) Daniel C. Long (3.852) Alicia M. Soderberg (3.724) 

This study analyzes the impact of the research ability (Qα) and research strategy ability (Eαπ) on scientists’ research performance through a novel RSQ model, in which Qα indicates the fundamental ability of a scientist α, while Eαπ describes the exploration and exploitation strategy ability of α. The BBVI-EM algorithm (Huang et al., 2023b) is utilized to solve the RSQ model. We use synthetic data and three empirical data sets in the physics, chemistry, and computer science fields to check the effectiveness of our model, and disclose the characteristics of five research strategies proposed by our previous study from the individual and overall perspectives. Overall, our model can not only be taken as a novel author-level evaluation tool, but it also sheds light on topic selection behavior in terms of exploration and exploitation. Here we provide a detailed summary of the theoretical and practical implications, as well as the limitations of our research.

6.1. Theoretical Implications

This article has the following theoretical implications. First, we propose a novel probabilistic graphical model, the RSQ model, which explain how scientists’ research ability and proficiency in different research strategies affect their research performance. Based on synthetic data and empirical data, we show that our model can untangle the relationships among research ability (Qα) and different research strategy abilities (Eαπ), and effectively estimate them. Compared with the Q model (Sinatra et al., 2016), overlooking the impact of research strategies, our model achieves a better estimation performance of Qα, which suggests that the coupling relationships among Qα and Eαπ should be considered.

Second, we employ the model parameters of the RSQ model to disclose the overall characteristics of five research strategies (πM, πP, πD, πE, πC) and shed light on the topic selection behavior from the perspective of exploration and exploitation. Specifically, based on our empirical data sets, our experimental results show that studying mature topics (πM) tends to stifle scientists from producing high-impact papers, because scientific research requires innovation. Studying diverse knowledge topics (πD) also stifles scientists’ performance, because diverse knowledge generally requires the handling of high levels of complexity (Schaltegger et al., 2013; Xu et al., 2015). However, following the academic frontier (πP) is the most effective way to enhance the academic impact. Exploration research strategies (πE and πC) are high risk and high yield (Liu et al., 2021; Yu et al., 2021). In conclusion, πM and πD are weakened strategies which tend to stifle scientists’ research performance, while πP, πE, and πC are enhanced strategies which tend to improve their research performance.

Third, we employ the variational parameters introduced by the BBVI-EM algorithm to analyze the author-level abilities through quantifying Qα and Eαπ. We check the linear correlation among Qα, Eαπ and some traditional author-level evaluation metrics. We find that Qα and Eαπ are linearly unrelated to the productivity (Nα), and have a slight correlation with the accumulative citations (Cα), citations of the most-cited paper (Cα*), and the h-index (Hα), which means that they can be used as complementary metrics for author-level evaluation. We also analyze the relationship among Qα, Eαπ, and pαπ, and find that scholars who prefer to execute a specific research strategy, π, may not necessarily be good at executing π. Importantly, our research thoroughly explains the relationship between exploration and exploitation; that is, exploration ability (EαπE, EαπC) and exploitation ability (EαπM, EαπD, EαπP) are positively correlated, indicating that scholars who are good at exploiting are often good at exploring, and vice versa. Indeed, only scholars who are good at exploring can understand and exploit cutting-edge knowledge, and only scholars who are good at exploiting have time to explore new knowledge. Therefore, exploration and exploitation are not contradictory but complementary (Zacher et al., 2016). However, the exploration preference (pαπE, pαπC) and exploitation preference (pαπM) are negatively correlated, which suggests that choosing between exploration and exploitation is an inevitable dilemma for scientists (Cohen et al., 2007). This is due to the limited attention of scholars, who need to make a choice between exploration and exploitation during their working hours. In addition, Qα and Eαπ have predictive power for forecasting scholars’ future research performance. Finally, we show that one scholar may have sperate proficiency levels for different research strategies, and different scholars may have distinct proficiency levels for the same research strategy.

6.2. Practical Implications

This article also has the following practical implications. First, the research ability (Qα) and research strategy ability (Eαπ) of different research strategies quantified by the RSQ model can be used to evaluate scientists’ ability from different aspects. Qα evaluates the comprehensive abilities of a scholar to publish high-quality papers, while a series of Eαπ evaluates the exploration and exploitation strategy ability of the scholar. Therefore, governments, academic institutions, and funding agencies may employ our proposed metrics to select excellent scholars, by which scientists with the greatest potential to produce path-breaking research can and should be rewarded according to their abilities (Madsen & Aagaard, 2020). For example, engineering projects can be assigned to scholars who are good at exploiting mature knowledge, while exploration projects can be handed over to scholars who are skilled at exploring new knowledge.

Second, five research strategies (πM, πP, πD, πE, πC) presented by Huang et al. (2022) are adopted in the RSQ model. Actually, any suitable definition of research strategies, such as a series of research strategies present by Foster et al. (2015), can also be adopted in our model. Hence, our model may be employed in other related studies to help researchers deeply understand exploration and exploitation behaviors. Overall, the current study not only sheds light on the characteristics of exploration and exploitation, but also proposes a new method and perspective for academic evaluation and performance assessment.

6.3. Limitations and Future Work

There are still some limitations in this study. First, in our correlation analysis and regression analysis, we observe a certain degree of linear correlation among the research ability and research strategy abilities. Therefore, a more sophisticated model may be proposed to better decouple and explain their complex coupling relationship. In addition, more factors influencing research performance (e.g., mentorship, team size, platform) should be included in our regression analysis. Second, our empirical data is collected from typical STEM disciplines. In the future, the empirical data in HSS disciplines and more sophisticated synthetic data need to be further analyzed to ensure the generalization performance of our model. Third, the RSQ model may be extended from static to dynamic to study the evolution of scientists’ abilities and strategies preferences. Finally, this paper employs citations to indirectly measure the quality of papers. The performance of our model based on other metrics measuring the quality of papers should be further explored.

Shengzhi Huang: Conceptualization, Formal analysis, Methodology, Writing—original draft. Wei Lu: Methodology, Supervision, Validation. Qikai Cheng: Data curation, Investigation. Li Zhang: Investigation, Validation. Yong Huang: Funding acquisition, Writing—review & editing.

The authors have no competing interests.

This work was supported by the Postdoctoral Fellowship Program of CPSF under grant number GZB20240565 and Youth Science Foundation of the National Natural Science Foundation of China (grant no. 72004168).

Ain
,
Q.
,
Riaz
,
H.
, &
Afzal
,
M. T.
(
2019
).
Evaluation of h-index and its citation intensity based variants in the field of mathematics
.
Scientometrics
,
119
,
187
211
.
Alonso
,
S.
,
Cabrerizo
,
F.
,
Herrera-Viedma
,
E.
, &
Herrera
,
F.
(
2010
).
hg-index: A new index to characterize the scientific output of researchers based on the h- and g-indices
.
Scientometrics
,
82
,
391
400
.
Bartneck
,
C.
, &
Kokkelmans
,
S.
(
2011
).
Detecting h-index manipulation through self-citation analysis
.
Scientometrics
,
87
,
85
98
. ,
[PubMed]
Blei
,
D. M.
,
Ng
,
A. Y.
, &
Jordan
,
M. I.
(
2003
).
Latent Dirichlet allocation
.
Journal of Machine Learning Research
,
3
,
993
1022
.
Bornmann
,
L.
, &
Daniel
,
H.-D.
(
2009a
).
Reviewer and editor biases in journal peer review: An investigation of manuscript refereeing at Angewandte Chemie International Edition
.
Research Evaluation
,
18
,
262
272
.
Bornmann
,
L.
, &
Daniel
,
H. D.
(
2009b
).
The state of h index research. Is the h index the ideal way to measure research performance?
EMBO Reports
,
10
,
2
6
. ,
[PubMed]
Bornmann
,
L.
, &
Haunschild
,
R.
(
2015
).
Which people use which scientific papers? An evaluation of data from F1000 and Mendeley
.
Journal of Informetrics
,
9
,
477
487
.
Chen
,
P.
,
Xie
,
H.
,
Maslov
,
S.
, &
Redner
,
S.
(
2007
).
Finding scientific gems with Google’s PageRank algorithm
.
Journal of Informetrics
,
1
,
8
15
.
Chen
,
Y.
, &
Ding
,
J.
(
2023
).
Exploitation and exploration: An analysis of the research pattern of Nobel laureates in Physics
.
Journal of Informetrics
,
17
,
101428
.
Cohen
,
J. D.
,
McClure
,
S. M.
, &
Yu
,
A. J.
(
2007
).
Should I stay or should I go? How the human brain manages the trade-off between exploitation and exploration
.
Philosophical Transactions of the Royal Society B: Biological Sciences
,
362
,
933
942
. ,
[PubMed]
Egghe
,
L.
(
2006
).
Theory and practise of the g-index
.
Scientometrics
,
69
,
131
152
.
Foster
,
J. G.
,
Rzhetsky
,
A.
, &
Evans
,
J. A.
(
2015
).
Tradition and innovation in scientists’ research strategies
.
American Sociological Review
,
80
,
875
908
.
Franco
,
A.
,
Malhotra
,
N.
, &
Simonovits
,
G.
(
2014
).
Publication bias in the social sciences: Unlocking the file drawer
.
Science
,
345
,
1502
1505
. ,
[PubMed]
Funk
,
R. J.
, &
Owen-Smith
,
J.
(
2017
).
A dynamic network measure of technological change
.
Management Science
,
63
,
791
817
.
Gao
,
C.
,
Wang
,
Z.
,
Li
,
X.
,
Zhang
,
Z.
, &
Zeng
,
W.
(
2016
).
PR-index: Using the h-index and PageRank for determining true impact
.
PLOS ONE
,
11
,
e0161755
. ,
[PubMed]
Hirsch
,
J. E.
(
2005
).
An index to quantify an individual’s scientific research output
.
Proceedings of the National Academy of Sciences
,
102
,
16569
16572
. ,
[PubMed]
Huang
,
S.
,
Huang
,
Y.
,
Bu
,
Y.
,
Luo
,
Z.
, &
Lu
,
W.
(
2023a
).
Disclosing the interactive mechanism behind scientists’ topic selection behavior from the perspective of the productivity and the impact
.
Journal of Informetrics
,
17
,
101409
.
Huang
,
S.
,
Lu
,
W.
,
Bu
,
Y.
, &
Huang
,
Y.
(
2022
).
Revisiting the exploration-exploitation behavior of scholars’ research topic selection: Evidence from a large-scale bibliographic database
.
Information Processing & Management
,
59
,
103110
.
Huang
,
S.
,
Lu
,
W.
,
Huang
,
Y.
, &
Luo
,
Z.
(
2023b
).
Quantifying scientists’ research ability by taking institutions’ scientific impact as priori information
.
Journal of Information Science
.
Jia
,
T.
,
Wang
,
D.
, &
Szymanski
,
B. K.
(
2017
).
Quantifying patterns of research-interest evolution
.
Nature Human Behaviour
,
1
,
0078
.
Johnson
,
S. D.
, &
Lovegrove
,
B. G.
(
2008
).
Assessment of research performance in biology: How well do peer review and bibliometry correlate?
BioScience
,
58
,
160
164
.
Khurana
,
P.
, &
Sharma
,
K.
(
2022
).
Impact of h-index on author’s rankings: An improvement to the h-index for lower-ranked authors
.
Scientometrics
,
127
,
4483
4498
.
Kuhn
,
T. S.
(
1977
).
The essential tension: Selected studies in scientific tradition and change
.
Chicago, IL
:
University of Chicago Press
.
Li
,
W.
,
Zhang
,
S.
,
Zheng
,
Z.
,
Cranmer
,
S. J.
, &
Clauset
,
A.
(
2022
).
Untangling the network effects of productivity and prominence among scientists
.
Nature Communications
,
13
,
4907
. ,
[PubMed]
Liu
,
L.
,
Dehmamy
,
N.
,
Chown
,
J.
,
Giles
,
C. L.
, &
Wang
,
D.
(
2021
).
Understanding the onset of hot streaks across artistic, cultural, and scientific careers
.
Nature Communications
,
12
,
5392
. ,
[PubMed]
Liu
,
L.
,
Wang
,
Y.
,
Sinatra
,
R.
,
Giles
,
C. L.
,
Song
,
C.
, &
Wang
,
D.
(
2018
).
Hot streaks in artistic, cultural, and scientific careers
.
Nature
,
559
,
396
399
. ,
[PubMed]
Luo
,
Z.
,
Lu
,
W.
,
He
,
J.
, &
Wang
,
Y.
(
2022
).
Combination of research questions and methods: A new measurement of scientific novelty
.
Journal of Informetrics
,
16
,
101282
.
Madsen
,
E. B.
, &
Aagaard
,
K.
(
2020
).
Concentration of Danish research funding on individual researchers and research topics: Patterns and potential drivers
.
Quantitative Science Studies
,
1
,
1159
1181
.
Petersen
,
A. M.
,
Riccaboni
,
M.
,
Stanley
,
H. E.
, &
Pammolli
,
F.
(
2012
).
Persistence and uncertainty in the academic career
.
Proceedings of the National Academy of Sciences
,
109
,
5213
5218
. ,
[PubMed]
Protasiewicz
,
J.
,
Pedrycz
,
W.
,
Kozłowski
,
M.
,
Dadas
,
S.
,
Stanisławek
,
T.
, …
Gałężewska
,
M.
(
2016
).
A recommender system of reviewers and experts in reviewing problems
.
Knowledge-Based Systems
,
106
,
164
178
.
Rotolo
,
D.
,
Hicks
,
D.
, &
Martin
,
B. R.
(
2015
).
What is an emerging technology?
Research Policy
,
44
,
1827
1843
.
Rousseau
,
R.
, &
Leuven
,
K. U.
(
2008
).
Reflections on recent developments of the h-index and h-type indices
.
COLLNET Journal of Scientometrics and Information Management
,
2
(
1
),
1
8
.
Schaltegger
,
S.
,
Beckmann
,
M.
, &
Hansen
,
E. G.
(
2013
).
Transdisciplinarity in corporate sustainability: Mapping the field
.
Business Strategy and the Environment
,
22
,
219
229
.
Schreiber
,
M.
(
2018
).
A skeptical view on the Hirsch index and its predictive power
.
Physica Scripta
,
93
,
102501
.
Shen
,
Z.
,
Ma
,
H.
, &
Wang
,
K.
(
2018
).
A Web-scale system for scientific knowledge exploration
.
arXiv
.
Sidiropoulos
,
A.
, &
Manolopoulos
,
Y.
(
2006
).
Generalized comparison of graph-based ranking algorithms for publications and authors
.
Journal of Systems and Software
,
79
,
1679
1700
.
Sinatra
,
R.
,
Wang
,
D.
,
Deville
,
P.
,
Song
,
C.
, &
Barabási
,
A.-L.
(
2016
).
Quantifying the evolution of individual scientific impact
.
Science
,
354
,
aaf5239
. ,
[PubMed]
Wang
,
R.
,
Zhou
,
Y.
, &
Zeng
,
A.
(
2023
).
Evaluating scientists by citation and disruption of their representative works
.
Scientometrics
,
128
,
1689
1710
.
Xu
,
J.
,
Ding
,
Y.
, &
Malic
,
V.
(
2015
).
Author credit for transdisciplinary collaboration
.
PLOS ONE
,
10
,
e0137968
. ,
[PubMed]
Yan
,
E.
, &
Ding
,
Y.
(
2011
).
Discovering author impact: A PageRank perspective
.
Information Processing & Management
,
47
,
125
134
.
Yang
,
A. J.
,
Hu
,
H.
,
Zhao
,
Y.
,
Wang
,
H.
, &
Deng
,
S.
(
2023
).
From consolidation to disruption: A novel way to measure the impact of scientists and identify laureates
.
Information Processing & Management
,
60
,
103420
.
Yang
,
J.
,
Lu
,
W.
,
Hu
,
J.
, &
Huang
,
S.
(
2022
).
A novel emerging topic detection method: A knowledge ecology perspective
.
Information Processing & Management
,
59
,
102843
.
Yin
,
Y.
,
Wang
,
Y.
,
Evans
,
J. A.
, &
Wang
,
D.
(
2019
).
Quantifying the dynamics of failure across science, startups and security
.
Nature
,
575
,
190
194
. ,
[PubMed]
Yu
,
X.
,
Szymanski
,
B. K.
, &
Jia
,
T.
(
2021
).
Become a better you: Correlation between the change of research direction and the change of scientific performance
.
Journal of Informetrics
,
15
,
101193
.
Zacher
,
H.
,
Robinson
,
A. J.
, &
Rosing
,
K.
(
2016
).
Ambidextrous leadership and employees’ self-reported innovative performance: The role of exploration and exploitation behaviors
.
Journal of Creative Behavior
,
50
,
24
46
.
Zeng
,
A.
,
Shen
,
Z.
,
Zhou
,
J.
,
Fan
,
Y.
,
Di
,
Z.
, …
Havlin
,
S.
(
2019
).
Increasing trend of scientists to switch between topics
.
Nature Communications
,
10
,
3439
. ,
[PubMed]
Zhang
,
Y.
,
Wang
,
M.
,
Zipperle
,
M.
,
Abbasi
,
A.
, &
Tani
,
M.
(
2023
).
RelRank: A relevance-based author ranking algorithm for individual publication venues
.
Information Processing & Management
,
60
,
103156
.
Zhao
,
Z.
,
Bu
,
Y.
, &
Li
,
J.
(
2021
).
Characterizing scientists leaving science before their time: Evidence from mathematics
.
Information Processing & Management
,
58
,
102661
.
Zhou
,
Y.
,
Wang
,
R.
,
Zeng
,
A.
, &
Zhang
,
Y.-C.
(
2020
).
Identifying prize-winning scientists by a competition-aware ranking
.
Journal of Informetrics
,
14
,
101038
.

Author notes

Handling Editor: Li Tang

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. For a full description of the license, please visit https://creativecommons.org/licenses/by/4.0/legalcode.

Supplementary data