Disclosing the interactive mechanism behind scientists’ topic selection behavior from the perspective of the productivity and the impact
Keywords
1. Introduction
It is particularly crucial to quantitatively figure out what factors affect scientists’ topic selection behavior and how they do so, because this not only affects how scientists are trained, and funded (Foster et al., 2015; Jia et al., 2017; Sinatra et al., 2016; Wang et al., 2013), but also implicitly reflects the process of collective discovery of new knowledge and how collective decisions shape the science (Yu et al., 2021; Zeng et al., 2019).
Previous studies have found various internal factors (e.g. gender, research interest, etc.) and external factors (e.g. topic value, topic novelty, funding, etc.) affecting topic selection behavior (Bu et al., 2018; Fadhly et al., 2018; Keshavarz & Shekari, 2020; Laudel, 2006; Li et al., 2017; Wei et al., 2013). However, few researches quantitively study whether and how the productivity and the impact acquired by studying a topic shape the evolution of scientists’ research interest. The interaction process between scientists and the external environment behind topic selection behavior is still unclear.
Jia et al. (2017) recently simulated the process of evolution of research interests, and proposed the seashore walk model, in which one of the basic assumptions is that scientists execute a random walk in topic selection. However, scientists actually may modify their research strategy when they receive the incentive from the external environment. As shown in Fig. 1, the environment includes other scientists, reviewers, journals, and etc. Scientists interact with the environment to improve the productivity and the impact, which are considered as critical evaluation factors of research performance evaluation of scientists in some countries (e.g., China). For example, in the process of a paper from submission to publication, authors need to communicate with reviewers, and apart from self-citation, an author can only accumulate citations when other scientists cite his/her papers. There are other evaluation factors as well. Under the conflict defined as “the essential tension” (Kuhn, 1977) and “publish or perish” (Qiu, 2010), scientists may shift their research focus to chase their goals. This study aims to examine whether the productivity and the impact acquired by studying a topic is related to the evolution of scientists’ research interest in a data-driven way, and disclose and simulate the interactive mechanism behind topic selection behavior.
In this paper, we selected over 20,000 scientists in the computer science field as the analysis objects, and employed “field of study” generated from Microsoft Academic Graph (MAG) (Sinha et al., 2015; Tang et al., 2008; Zhang et al., 2019) to represent topics. We proposed two correlation indicators, by which we disclose that the productivity and the impact shape the correlation in the evolution of research interest. To further explore how the two factors affect scientists’ topic selection behavior, we also proposed a variant of the seashore walk model (Jia et al., 2017), Q seashore walk model based on the Q-learning algorithm (Watkins, 1989; Watkins & Dayan, 1992), to simulate the process of topic choice based on the interactive mechanism hypothesis. To test the validity of our model, we compared the metrics distribution generated by the empirical data and that generated by the simulation data, and analyzed the role of rewards in scientists’ research performance.
The current study has the following theoretical and practical implications. We present that the productivity and the impact shape the correlation in the evolution of research interest of scientists, and used a simulation model to explain how this happens. Therefore, this study may help researchers deeply understand scientists’ topic selection behavior, and provides a theoretical basis for formulating research and development policy. The spirit of our model provides an interesting perspective to examine scientists’ decision-making process, and contributes to the theoretical advancement of applying the reinforcement learning theory to analyze scientists’ behavior.
The rest of this paper is organized as follows. Section 2 reviews the related work on scientists’ topic selection behavior and reinforcement learning theory. The problem definition and methodology are clarified in Section 3. The data set is clearly introduced in Section 4. In Section 5, we provide the experimental setup and analysis of experimental results. In Section 6, we offer the contributions and limitations of this study. Finally, we summarize the paper in Section 7.
2. Background
2.1. The factors affecting topic selection behavior
How scientists select their research topics is an area of continuous focus, and previous studies have found a variety of internal and external factors affecting this behavior. In this paper, we simply define external factors as the objective environmental factors (e.g., topic nature, funding, etc.), and internal factors as the influencing factors closely related to the scientists themselves (e.g., gender, research interest, etc.).
Many studies focused on analyzing external factors. For example, Wallace and Ràfols (2018) conducted analysis of avian influenza research by qualitative interviews and quantitative bibliometric approaches, and confirmed that institutional pressures and funding pressures have a powerful influence on the topic choice of researchers. Laudel (2006) also found that physicists adapt their research content to obtain funding. Hoonlor et al. (2013) also revealed that funding may raise interest in the supported disciplines. After interviewing three Indonesian scientists, Fadhly et al. (2018) identified 11 ways (e.g. research trends, available data, etc.) to select research topics, and concluded that scientists from different disciplines have their own approaches to select topics. Wei et al. (2013) and Li et al. (2017) found that scientists from physics, mathematics, economics, and biomedicine tend to follow large and hot topics, and Chinese scientists prefer to follow large topics. Lakeh and Ghaffarzadegan (2017) analyzed the abstracts of over 200,000 HIV/AIDS scientific papers, and found that where scientists live affect their research topics. Buehling (2021) aimed to examine whether journal rankings may influence scientists’ topic selection process. After checking the Handelsblatt Ranking (HBR) for economists, they confirmed that there is no significant relationship between the two.
Some studies explored to analyze internal factors, and disclose the inherent characteristics of topic selection behavior. Duch et al. (2012) showed that gender difference may result in the inequality in academic resource allocation, which drives topic selection behavior. Jia et al. (2017) found that the evolution of research interests follows an exponential distribution, which is determined by three fundamental factors (i.e., heterogeneity, recency, and subject proximity). Bu et al. (2018) analyzed the scientific careers of scientists from the computer science field, and found that the collaborators of high-impact scientists tend to study diverse research topics. Keshavarz and Shekari (2020) analyzed data collected through a questionnaire survey of 391 postgraduate students, and revealed four factors affecting topic selection (i.e. personal issues, topic nature, information resources, and research operability). In addition, Zeng et al. (2019) analyzed the co-citing network of papers of 3420 scientists, and showed that scientists nowadays move more frequently among topics than those in the past. Yu et al. (2021) performed an analysis of publication records from over 14,000 scientists in physics, and disclosed that the change of research interest may help scientists create papers with increased impact, but it is not associated with productivity. Huang et al. (2022b) recently proposed five novel research strategies under exploration and exploitation, and uncovered the relationship between scientists’ research performance (i.e., productivity and impact) and their preference for these strategies. However, they did not further explore how different incentives and/or rewards drive topic selection behavior.
In short, various internal factors and external factors affect scientists’ research topic selection process. However, previous studies generally neglected to explore whether and how the productivity and the impact affect the evolution of scientists’ research interests. The interaction mechanism between scientists and the environment in the process of topic selection is still understudied. In this study, we aim to answer the above questions.
2.2. Reinforcement learning theory
The fundamental goal of reinforcement learning is to find good policies, which can resolve sequential decisions problems for agents, by optimizing cumulative future rewards (Sutton & Barto, 2018). The Q- learning algorithm (Watkins, 1989; Watkins & Dayan, 1992) is one of the classical valued-based algorithms, which teaches the agents to learn how to act in the controlled Markovian environment, and the convergence of the Q-learning algorithm has been strictly proven. In recent years, deep learning algorithm and Q-learning algorithm have been effectively combined. Mnih et al. (2013) first combined the neural network with reinforcement learning and proposed the Deep Q-Network (DQN), which achieved excellent performance in seven Atari games. Van Hasselt et al. (2016) analyzed and tackled the DQN's overestimation of action value by presenting the Double Q-learning algorithm (DDQN), which achieves better performances in some Atari games. Wang et al. (2016) presented the dueling network, which separately estimates the state value function and the state-dependent action advantage function. Their model exceeds the state of the art in the Atari 2600 domain. Moreover, reinforcement learning has also been widely utilized in various fields. For example, AlphaGo Zero has achieved a superhuman performance in Go, and also defeated previous AlphaGo models (Silver et al., 2017). Duan et al. (2021) employed the Q-learning algorithm to solve the log anomaly detection task, and proposed the QLLog model. Ciranka et al. (2022) recently employed the reinforcement learning policy to imitate the human behavior in relational learning. They found that human learner used symmetric learning policy under full feedback contexts, but asymmetric learning policy under partial feedback contexts.
To the best of our knowledge, the interaction mechanism between scientists and the environment behind their topic selection behavior generally is relatively understudied, and therefore remains unclear. Inspired by Jia et al. (2017) and Ciranka et al. (2022)’s research, we proposed a novel Q seashore walk model based on the Q-learning algorithm, in which the process of topic selection is imitated. Our simulation model aims to help researchers better understand scientists’ topic selection behavior, and provide evidence for policy intervention in scientific research.
3. Methodology
3.1. Problem definition
The productivity and the impact are two widely employed aspects to evaluate the research performance of a scientist, and are generally quantified by number of publications and citation frequency (Huang et al., 2021, 2022a; Khosrowjerdi & Bornmann, 2021; Perianes-Rodriguez & Ruiz-Castillo, 2018; Sinatra et al., 2016; Zhu et al., 2021). Due to scientists’ pursuit of career advancement, the productivity and the impact may be seen as the incentive and/ or reward from external environment. Specifically, the authors should interact with the environment to publish papers and accumulate citations. For example, in the process of a paper from submission to publication, authors should consult with reviewers. Apart from self-citation, an author can only accumulate citations when other scientists cite his/her papers. Hence, there may exist the interaction mechanism between scientists and the environment, which drives scientists’ topic selection behavior, and topics that bring richer benefits may be more attractive to scientists.
This study aims to explore whether and how the productivity and the impact acquired by studying a topic shape scientists’ topic selection behavior in a data-driven way, and disclose and simulate the interactive mechanism composed of the productivity and the impact behind topic selection. In this paper, firstly, we propose TDC (Topic distribution correlation) and TFC (Topic frequency correlation) metrics to show that the productivity and the impact shape the correlation of topic distribution in scientists’ publication sequence. Second, we propose the Q seashore walk model to further verify and explain our results. Finally, we also conduct a series of robustness tests.
3.2. Topic distribution correlation metrics and topic frequency correlation metrics
To disclose the correlation among the evolution of scientists’ research interests from the perspective of the productivity and the impact, we proposed two correlation metrics. Before formally providing the formulas of the two metrics, we first introduce some definitions and notations used in this study.
For each scientist, , we sorted ’s publication sequence by the publication year of each paper (i.e., ), where the research content of each paper is represented by a list of topics (e.g., FoS in MAG dataset). We proposed two division approaches to split the publication sequence into adjacent windows. In the first division method, the publication sequence is divided by the time span, , and papers published between and are grouped into the same group. In the second division method, the publication sequence is divided by the number of publications, , and consecutive papers are grouped into the same group. To make full use of the publication sequence, we employed the sliding window to traverse the publication sequence from beginning to end. To clarify the two division methods, we provide a simple example, as shown in Fig. 2. has published papers from 2000 to 2020. and are set to 2 and 3, respectively. Thus, in the first division method, papers published in 2000 and 2001(i.e., ) constitute the first window, and constitute the second window, and so on. In the second division method, the first three papers (i.e., ) constitute the first window, and also are grouped into the second window, and so on.
Subsequently, we introduce the topic distribution correlation metric (TDC) and topic frequency correlation metrics (TFC). Specifically, topic vector (Jia et al., 2017) is employed in TDC and TFC. To evaluate the level of importance of topics from the perspective of the productivity and the impact, we expand the definition of topic vector. We respectively employed the frequency of each topic adopted by and the citation count acquired by the papers studying the topic authored by in each window () to count each topic. The topic vectors calculated by the usage frequency and citation frequency in are denoted as and , respectively, and are a K-dimensional vector. is the number of unique topics in ’s publication sequence. For example, as shown in the Fig. 3, in the first window, is , because “Java” is adopted twice in two papers, and “Machine learning” and “Computer vision” are respectively adopted once in one paper. is , because the papers adopting “Java” and “Machine learning” are cited only once in a citation window, .
Because we focus on the difference of the importance ranking of topics in the topic vectors, the Spearman correlation coefficient is adopted. Specifically, in the first division method, we calculated the Spearman correlation coefficient between and , . The gap, , avoids the overlapping of two paper sets from two windows. Similarly, we also calculated in the second division method. The positive correlation value suggests that scientists prefer to continue to study the topic that previously generated more papers. However, the zero and negative values mean that productivity have nothing to do with the scientists’ topic selection behavior. Subsequently, we calculated and .The positive value suggests that scientists tend to choose the topic that previously had a greater impact, and the zero and negative values mean that the impact is a negligible factor for the scientists’ topic selection behavior. Notably, to consider the whole career life span of a scientist, the Spearman correlation coefficient between all pairs of windows is calculated, and these coefficients with significance level above a certain level, , are summed and divided by the total number of pairs of windows, ( and). Finally, the formulas of can be found in Eqs. (1)-(4).(1)(2)(3)(4)
Unlike TDC, which measures the relative importance of different topics, TFC aims to gage the level of involvement of a single topic. Specifically, for th topic, we count its usage frequency sequence and citation frequency sequence in windows, denoted as , and . For example, in Fig. 3, is . Then, in the first division method, we calculate the Pearson correlation coefficient between and , . Notably, the gap, , avoids the overlapping of two paper sets from two adjacent windows. Similarly, we calculated in the second division method. The positive value suggests that scientists continuously select the topic that previously generated more productivity. We also calculate and . The positive value means that scientists continuously study the topic that previously generated more impact. To consider all topics adopted by a scientist, for each topic, the Pearson correlation coefficient is calculated, and these coefficients with significance level above a certain level, , are summed and divided by the total number of topics, . The formulas of can be found in Eqs. (5)–(8). The notations used in TDC and TFC and its definitions can be found in Table 1.(5)(6)(7)(8)
Notation | Definition |
---|---|
The length of time span required in the first division method | |
The number of publications required in the second division method | |
The total number of windows in the first division method | |
The total number of windows in the second division method | |
The length of citation window | |
The index of th window | |
Topic vector calculated based on the productivity in | |
Topic vector calculated based on the impact in | |
Topic frequency calculated based on the productivity in | |
Topic frequency calculated based on the impact in | |
Significant level | |
The index of th topic | |
The total number of unique topics in a publication sequence |
3.3. Q seashore walk model
In this subsection, we aim to explore how the productivity and the impact affect the scientists’ topic selection behavior. We employed a simulation model to imitate topic selection behavior based on the interactive mechanism hypothesis, in which the productivity and the impact are regarded as rewards.
Inspired by Isaac Newton's retrospection that his scientific career had been like “a boy playing on the seashore… finding… a prettier shell than ordinary” (Jia et al., 2017). Mandelbrote (2001) proposed the seashore walk model, which models the evolution of a scientist's research interests. However, the seashore walk model simply assumes that scientists perform an unbiased random walk on the 1-D lattice, which means that they follow a random exploration policy. However, we argued that scientists might chase goals, and shift their research focus according to the benefits gained by their previous strategies. Therefore, we proposed the Q seashore walk model, which aim to further figure out the role of the productivity and the impact in the process of selecting topics.
Before introducing our model, we simply review the seashore walk model. A scientist, , starts from a random position, and performs an unbiased random walk () on a seashore, which is a 1-D lattice with piles of shells located on some sites, as shown in the Fig. 4. Different-colored shapes represent shells with different combinations of topics. The walker picks a shell at one site, corresponding to publishing a paper. The probability that shells exist on a site is . The number of shells at the site, , follows a power law distribution, , if there are any shells. is the maximum number of shells in a site. The total number of steps of a walker, , followed a truncated log-normal distribution, . The total length of the seashore is sites, and the length of each topic tool is sites. Each topic pool has three distinct topics, and two neighboring topic pools (e.g., and ) vary by one topic. The topics of papers from are the random combination of topics from the topic tool.
In the Q seashore walk model, instead of assuming an random strategy, we employed a classical and effective reinforcement learning algorithm, the Q-learning algorithm (Watkins, 1989; Watkins & Dayan, 1992), to quantify the decision-making policy adopted by . Specifically, our model is established according to the following ideas: tends to stay at the topic pool that brings them rich rewards, and depart from the topic pool with poor benefits. In short, if picks one shell, would prefer to stay at the current position, otherwise try to leave. Specifically, the scientist is the agent, , and the seashore is the environment, . The action space, , is composed of moving one step left or right, denoted as and . The state space, , is composed of the left side of the topic pool and the right side of the topic pool, denoted as and . The rewards gained by depend on three factors (i.e., action, state, and shell), as shown in Table 2. For example, as shown in Fig. 5, when at state, , takes the action and then picks a shell, he will get a negative reward, , which suggests that leaving the topic pool is a wrong decision. Moreover, to consider the productive differences among topics, we divide the topic pool into high-yield topic pool and general topic pool with the probability, and . The difference between the two is that the former has a larger . To model the impact of simulated papers, we employed a log-normal distribution, to imitate the citation distribution of simulated papers.
Empty Cell | ||||
Any shells | ||||
Yes | ||||
No |
Finally, we introduced the definition of Q value, , which gauges the long-term benefits of taking an action at a state, and which can be estimated iteratively by Bellman equation, as shown in Eq. (9). takes an action by with the probability, , and executes an unbiased random walker with the probability, . For example, as shown in the upper of Fig. 5, due to , the agent takes . When , we used the random strategy. Thus, the Q seashore walk model degenerates into the seashore walk model, when the random strategy is always adopted. In our model, there are only four combinations of the action and state, and therefore only takes four values. In each , the , are employed to manage the policy of , andare initialized with zero values. We also introduced three rules to update . (1). when does not leave , and remains unchanged (e.g., ), Eq. (9) is utilized to update . (2). when does not leave , but changes (e.g., ), we update and to zero values. This is due to the fact that has been at the middle of the topic pool, and both actions seem to be good decisions. (3). when leaves (e.g., arrives at ), Eq. (9) is utilized to update , Eq. (10) is employed to update , and then and are updated to zero value. indicates the opposite direction of . Actually, Eq. (10) indicates the intensity of 's desire to leave . The notations used in the Q seashore walk model can be found in Table 3.(9)(10)
Notation | Definition |
---|---|
The agent (walker, scientist) | |
Action space () | |
State space () | |
Environment (seashore) | |
The long-term benefits of taking an action at a state | |
The reward when a shell has been found | |
The reward when no shell is found | |
The discount parameter in the Bellman equation | |
The probability that a site contains any shells | |
The number of shells at the site () | |
The maximum number of shells in a site | |
The probability that the topic pool is high-yield | |
The total number of steps of a walker () | |
C | Citation distribution of simulated papers () |
The maximum number of steps for a walker | |
The probability that the walker adopts an exploration action | |
The topic pool (The length of a topic pool) | |
The length of all topic pools () | |
The number of 1-D lattices | |
The number of scientists sampled in each 1-D lattice |
3.4. Data collection and preprocessing
In this paper, we utilized Microsoft Academic Graph (MAG) data from Open Academic Graph (OAG) 2.1 (Sinha et al., 2015; Tang et al., 2008; Zhang et al., 2019) as our data source. MAG comprise over 200 million papers, and is one of the most widely used data sets in bibliometrics (De Domenico et al., 2016; Jin et al., 2021). In MAG, there are totally 228,251 “field of study” (FoS), which are nested within 293 sub-disciplines, which are nested within 19 disciplines (Shen et al., 2018). Each paper collected from MAG has been associated with a list of FoS (i.e. topics).
Since we are familiar with the computer science field (CS), we extracted papers from this field by the FoS keyword, “computer science”, as our data set (hereafter, ). There are 23,471,718 papers, and 29,073,597 authors who have published at least one paper in .To analyze the role of the productivity and the impact in the evolution of scientists’ research interests, we considered scientists who have published at least 100 papers in the CS field. This is because scientists who have published fewer papers cannot provide sufficient data for our analysis, and introduce randomness into the results. Finally, 21,239 scientists constitute our analysis object. Our dataset will be described in detail in the following “Descriptive statistics” subsection.
4. Experiments and results
4.1. Descriptive statistics
The distribution characteristics of is shown in Fig. 6. The annual number of publications in the CS field experienced exponential growth from 1950 to 2019, as shown in Fig. 6(a). The number of new authors in the CS field per year also experienced substantial growth, as shown in Fig. 6(b). However, the majority of authors (77.09%) only published one paper, and the proportion of authors with less than four papers is over 90%. As we mentioned previously, only 21,239 authors with over 100 papers () are selected as our analysis object. Fig. 6(c) presents the annual number of new topics, which increased rapidly from 1950 to 2010 and then slowed down from 2011 to 2020, which means that the CS field has gradually become mature, but it is still developing. Notably, the lower number of papers, authors, and topics in 2020 is due to the truncation of data collection. The number of publications in each topic can be found in Fig. 6(d). The only topic with over 3 million papers is “artificial intelligence”, and the majority of topics have less than 100,000 papers. Actually, many topics have only been selected a few times during the scientific career of . Hence, topics adopted by less than 10 times () are filtered out, when we calculated the TDC and TFC. In the “Robustness tests” subsection, we also repeated our experiments with different value of and .
4.2. Correlation analysis based on TDC and TFC metrics
We employed TDC and TFC metrics to show that the productivity and the impact shape the correlation in the evolution of research interest. To analyze the magnitude of the correlation coefficient, we built a control group by shuffling scientists’ publication sequence. Specifically, we randomly sort the order of a scientist's papers on the premise that annual number of papers in the reshuffled sequence is the same as that in the original sequence. In the following experiments, and in two division methods are set to 1 and 8, respectively. is set to 0.1. TDC and TFC distributions of over 20,000 scientists based on the empirical data and the randomized data are shown in Figs. 7 and 8.
As shown in Fig. 7(a) and (b), distribution based on the empirical data (the red bar) are bell-shaped distributions, and their mean values are greater than zero (0.135 and 0.166), which means that scientists prefer to continuously study topics that bring more productivity. We also found that distribution based on the shuffled data (the gray bar) obviously shift to the negative half of the x-axis, and their mean values are smaller than those based on the empirical data (0.067 and 0.111). Moreover, as shown in Fig. 7(c) and (d), distribution based on the empirical data also are bell-shaped distributions with mean values greater than zero (0.133 and 0.105), which also means that scientists prefer to study the topics that help them produce papers. Not surprisingly, distribution based on the shuffled data also obviously shifts to the negative half of the x-axis, and their mean values are 0.078 and −0.080. Notably, in Fig. 7, TDC and TFC distributions based on the empirical data and those based on the randomized data are significantly different under the KS test with .
Subsequently, we used the citation count acquired by a paper in following 3 years to count the topic vector. As shown in Fig. 8(a) and (b), we found that distribution based on empirical data presents the bell-shaped distribution, and are at the right side of distribution based on shuffled data. As shown in Fig. 8(c) and (d), we get the similar results on . In addition, distribution based on empirical data and its control group are also significantly different under the KS test with . Thus, scientists tend to continuously study the topic that produce the greater impact.
4.3. Robustness tests
To test the robustness of our results on empirical data, we respectively filtered out the high-frequency topics, changed the values for and , changed the values for and , and did not use the sliding window method, and repeated the above experiments.
Specifically, (1). considering that some topics are adopted frequently by a scientist and the control group may keep the order of these high-frequency topics, we simply filter out the topics studied by more than 60% of the total number of papers published by a scientist, and repeated the experiments. (2). We change the values for and ( and ), and repeat our experiments. (3). We change the values for , employ authors who have published from 60 to 100 papers as our analysis objects, and repeat our experiments. (4). We change the values for , allow more low frequency topics, and repeat our experiments. (5). We do not use the sliding window method in two division methods, and repeat our experiments. For example, in the second division method, when , we have , instead of .
In the above robustness tests, our results remain unchanged. Specifically, the difference between the average value of TDC (TFC) distribution based on the empirical data and that based on the randomized data is always positive, which means that the random distribution shifts to the negative half of the x-axis. The results are listed in Table 4. Moreover, in all the aforementioned experimental results, TDC (TFC) kernel density based on empirical data and its control group are significantly different under the KS test with .
Robustness tests | (1) | (2) | (3) | (4) | (5) |
---|---|---|---|---|---|
0.072⁎⁎⁎ | 0.034⁎⁎⁎ | 0.037⁎⁎⁎ | 0.086⁎⁎⁎ | 0.069⁎⁎⁎ | |
0.058⁎⁎⁎ | 0.047⁎⁎⁎ | 0.018⁎⁎⁎ | 0.067⁎⁎⁎ | 0.053⁎⁎⁎ | |
0.069⁎⁎⁎ | 0.055⁎⁎⁎ | 0.051⁎⁎⁎ | 0.074⁎⁎⁎ | 0.060⁎⁎⁎ | |
0.068⁎⁎⁎ | 0.078⁎⁎⁎ | 0.045⁎⁎⁎ | 0.076⁎⁎⁎ | 0.072⁎⁎⁎ | |
0.057⁎⁎⁎ | 0.041⁎⁎⁎ | 0.038⁎⁎⁎ | 0.061⁎⁎⁎ | 0.045⁎⁎⁎ | |
0.186⁎⁎⁎ | 0.203⁎⁎⁎ | 0.168⁎⁎⁎ | 0.142⁎⁎⁎ | 0.131⁎⁎⁎ | |
0.048⁎⁎⁎ | 0.052⁎⁎⁎ | 0.032⁎⁎⁎ | 0.052⁎⁎⁎ | 0.032⁎⁎⁎ | |
0.114⁎⁎⁎ | 0.127⁎⁎⁎ | 0.094⁎⁎⁎ | 0.096⁎⁎⁎ | 0.078⁎⁎⁎ |
Note: *** indicates pvalue< 0.001.
4.4. Modeling the process of the topic selection by the Q seashore walk model
In this subsection, we employed the Q seashore walk model to imitate the process of selecting topics. The simulation experiment aims to further explore how the productivity and the impact affect the evolution of scientists’ research interests, and help us deeply comprehend the role of rewards for scientists in scientific research.
Specifically, after filtering out topics adopted by a scientist less than 10 times in his publication sequence, each paper from 21,239 scientists covers an average of 3.21 topics. Thus, the number of unique topics in each topic pool is set to three. To reproduce the real TDC and TFC distribution, we set different parameter settings of the Q seashore model, as shown in Table 5. First, to reproduce the real TDC distribution, we adopted the set of parameters in the first column of Table 5. There are totally 3 million pieces of simulated publication sequence, in which 8339 walkers have more than 100 shells. Since there is no publication year in the simulated data, we only calculated the in the simulation 1 data. is also set to 8. As shown in Fig. 9, the real TDC distribution and the simulated one are almost overlapping, and have the similar mean and standard deviation. Second, to reproduce the real TFC distribution, we adopted the set of parameters in the second column of Table 5. There are totally 15 million pieces of simulated publication sequence, in which only 4880 walkers have more than 100 shells. We calculated the in the simulation 2 data. As shown in Fig. 10, the real TFC distribution and the simulated one are broadly similar, and have the similar mean and standard deviation. Notably, the difference between the kernel density of the real distribution and that of the simulated distribution is not significant under the KS test with , and the can be found in Figs. 9 and 10. Considering that simulating scientists’ behavior is a complex and challenging task, we argue that the Q seashore walk model has achieved satisfactory simulation results. Thus, our model effectively imitates the scientists’ topic selection process based on the interactive mechanism. The interactive mechanism is a reasonable explanation of how the productivity and the impact affect the scientists’ topic selection behavior. Specifically, the scientist may perceive the productivity and the impact as a kind of rewards, and then adjust their current topic selection policy for their goals, as shown in Fig. 1.
Notations | Simulation 1 | Simulation 2 | Simulation 3 |
---|---|---|---|
5.0 | 5.0 | [−10,25] | |
1.0 | 1.0 | 1.0 | |
1.0 | 1.0 | 1.0 | |
0.15 | 0.035 | 0.035 | |
10 | 10 | 10 | |
0.15 | 0.15 | 0.0 | |
10 | 10 | 10 | |
C | |||
5000 | 5000 | 20,000 | |
0.1 | 0.0 | 0.0 | |
20 | 20 | 50 | |
3000 | 3000 | 5000 | |
300 | 300 | 150 | |
10,000 | 50,000 | 50,000 |
To figure out how the magnitude of rewards affect scientists’ topic selection behavior and then affect their research performance, we repeated our experiments under the set of parameters in the third column of Table 5, in which the value of ranges from −10 to 25 with the interval equal to 1. As shown in Fig. 11(a) and (b), the x-axis indicates , and the y-axis represents the average accumulative number of papers and the average accumulative number of citations in all simulated scientists. With the increase of , the average number of papers per simulated scientist rises first and then decreases (red dot), and the trend can be perfectly fitted by a quadratic function (black line). Since there exist a linear correlation between the cumulative number of papers of a scientist and the cumulative number of citations of the scientist, Fig. 11(b) shows the same trend as Fig. 11(a). To be specific, when the gradually becomes positive, the average number of papers (citations) increase, which is in line with our common sense that proper rewards promote the scientists’ research performance. However, when is too large, rewards have a negative effect on the productivity and the impact of scientists, which is also line with our common sense that excessive rewards make researchers become satisfied with the status quo. We summarize this phenomenon as “too much is as bad as too little”. The optimal value of is about 7, which means there exists a theoretical optimal reward to promote scientists’ research performance by coordinating the interactive mechanism. Moreover, in Fig. 11(c), the y-axis represents the average unique topics adopted by the simulated scientists. We found that gray points can be fitted well by a linear function with a negative slope, −0.0014 (), which means that excessive rewards also discourage scientists from exploring new topics. However, exploration is one of the key factors in scientific innovation (Foster et al., 2015), and therefore excessive rewards may undermine the chances of originality. Hence, we provide theoretical evidence that a proper rewards mechanism may help the government and institutions adjust the research strategy of scientists, so as to better promote scientific development. Notably, because the majority of simulated scientists do not have any shells, and therefore the average number of papers in Fig. 11(a) is less than 1.
5. Discussion
Under the conflict defined as “the essential tension” (Kuhn, 1977) and “publish or perish” (Qiu, 2010), the productivity and the impact may be seen as the intermediate goals for scientists, to pursue the career advancement. In this paper, we explore to disclose and model the interactive mechanism composed of the productivity and the impact. In the following paper, we discuss the theoretical and practical implications and limitations of this study.
This paper has the following theoretical implications. First, we propose two correlation metrics to disclose the correlation in term of topic distribution from scientists' publication sequence, which verifies that the productivity and the impact is related to the evolution of scientists’ research interests. Specifically, some scientists tend to continuously select topics which help them produce more productivity and impact. Second, we propose a novel Q seashore walk model. The simulated publication sequences generated by our model also present the correlation consistent with that in the empirical data. Therefore, our model effectively imitates the process of topic selection, and further confirms the above conclusions. More importantly, our model shows that the productivity and the impact may be seem as a reward, and the interactive mechanism drives scientist's topic selection behavior. In addition, the spirit of our model provides an interesting perspective to examine scientists’ decision-making process, and contributes to the theoretical advancement of applying the reinforcement learning theory to analyze scientists’ behavior. Third, based on the simulated data, we also prove that proper rewards effectively stimulate the performance of scientists, but excessive rewards inhibit their performance. This research may help researchers deeply understand the process of topic selection, and provide a theoretical basis for research and development policy formulation.
This study has the following practical implications. According to our simulation results, we report evidence that the research performance of scientists may be affected by the reward, and follows the trend of increasing first and then decreasing. Hence, government departments and institutions may adjust their research and development policy, and help scientists to build a benign interactive mechanism, so as to better promote the scientific development. Taking the revolutionary scientific research evaluation mechanism in China in recent years for example, Academic Representative Work System is advocated and gradually implemented, which effectively prevents scientists from blindly pursuing the number of publications rather than papers’ quality.
However, there are still some limitations in this study. First, there still exists other many factors affecting scientists’ topic selection behavior, which may or may not affect the interactive mechanism. Hence, in future studies, other factors need to be further analyzed. Second, our model confirms that there exists an optimal reward value to maximize the research performance of scientists. However, due to the fact that rewards from the real world are complex, the theoretical optimal reward is not equivalent to the real one. Thus, a sophisticated reward mechanism should be proposed, by which our model may provide more detailed guidance for research and development policy formulation. Furthermore, we only analyzed scientists from the CS field. Actually, scientists from different fields and from scientific strata may attach different importance to the same type of reward, or completely focus on a different type of reward, and therefore constitute a completely different interactive mechanism. Identifying what kinds of rewards scientists pursue can help government departments, academic publishing companies, and policy makers put forward practical proposals to promote the development of science.
6. Conclusion
In this study, we proposed two correlation metrics (TDC and TFC) to analyze 20,000 scientists’ publication sequence from the computer science field. We disclose that the productivity and the impact is related to the evolution of scientists’ research interests. Indeed, career advancement needs a steady stream of publications (Jia et al., 2017), and therefore scientists tend to continuously select topics which help them produce more productivity and impact. Notably, it should be emphasized that the observed results are statistical, not definitive, and individual scientists might exhibit a diversity of behaviors. To further figure out how the productivity and the impact drive the evolution of scientists’ research interests, we proposed a novel Q seashore walk model, which reported the evidence that the interactive mechanism is a reasonable explanation of how the productivity and the impact affect the scientists’ topic selection behavior. We also analyzed the role of reward for scientists’ research performance, and proved that proper rewards stimulate the performance of scientists, but excessive rewards inhibit their performance. We summarize this phenomenon as “too much is as bad as too little”. Thus, this study may help researchers figure out the motivations behind scientists’ decision-making process, deeply understand their topic selection behavior, and provide theoretical evidences for policy intervention in scientific research.
Declaration of Competing Interest
The authors declare no competing interests.
Acknowledgments
This work was supported by the Youth Science Foundation of the National Natural Science Foundation of China (grant no. 72004168).
References
- Bu et al., 2018Understanding success through the diversity of collaborators and the milestone of careerJournal of the Association for Information Science and Technology, 69 (2018), pp. 87-97
- Buehling, 2021Changing research topic trends as an effect of publication rankings–The case of German economists and the Handelsblatt RankingJournal of Informetrics, 15 (2021), Article 101199
- Ciranka et al., 2022Asymmetric reinforcement learning facilitates human inference of transitive relationsNature Human Behaviour, 6 (2022), pp. 555-564
- De Domenico et al., 2016Quantifying the diaspora of knowledge in the last centuryApplied Network Science, 1 (2016), pp. 1-13
- Duan et al., 2021QLLog: A log anomaly detection method based on Q-learning algorithmInformation Processing & Management, 58 (2021), Article 102540
- Duch et al., 2012The possible role of resource requirements and academic career-choice risk on gender differences in publication rate and impactPLOS One, 7 (2012), p. e51332
- Fadhly et al., 2018Exploring cognitive process of research topic selection in academic writingEnglish Review: Journal of English Education, 7 (2018), pp. 157-166
- Foster et al., 2015Tradition and innovation in scientists’ research strategiesAmerican Sociological Review, 80 (2015), pp. 875-908
- Hoonlor et al., 2013Trends in computer science researchCommunications of the ACM, 56 (2013), pp. 74-83
- Huang et al., 2022aFine-grained citation count prediction via a transformer-based model with among-attention mechanismInformation Processing & Management, 59 (2022), Article 102799
- Huang et al., 2022bRevisiting the exploration-exploitation behavior of scholars’ research topic selection: Evidence from a large-scale bibliographic databaseInformation Processing & Management, 59 (2022), Article 103110
- Huang et al., 2021Disclosing the relationship between citation structure and future impact of a publicationJournal of the Association for Information Science and Technology, 73 (7) (2021), pp. 1025-1042
- Jia et al., 2017Quantifying patterns of research-interest evolutionNature Human Behaviour, 1 (2017), pp. 1-7
- Jin et al., 2021Scientific prizes and the extraordinary growth of scientific topicsNature Communications, 12 (2021), pp. 1-11
- Keshavarz and Shekari, 2020Factors affecting topic selection for theses and dissertations in library and information science: A national scale studyLibrary & Information Science Research, 42 (2020), Article 101052
- Khosrowjerdi and Bornmann, 2021Is culture related to strong science? An empirical investigationJournal of Informetrics, 15 (2021), Article 101160
- Kuhn, 1977The essential tension: Selected studies in scientific tradition and changeUniversity Of Chicago Press, Chicago, Illinois (1977)
- Lakeh and Ghaffarzadegan, 2017Global trends and regional variations in studies of HIV/AIDSScientific Reports, 7 (2017), pp. 1-8
- Laudel, 2006The art of getting funded: How scientists adapt to their funding conditionsScience and Public Policy, 33 (2006), pp. 489-504
- Li et al., 2017Do mathematicians, economists and biomedical scientists trace large topics more strongly than physicists?Journal of Informetrics, 11 (2017), pp. 598-607
- Mandelbrote, 2001Footprints of the lionCambridge Univ. Library (2001)
- Mnih et al., 2013Playing Atari with Deep Reinforcement Learning (2013)
ArXiv, abs/1312.5602.
- Perianes-Rodriguez and Ruiz-Castillo, 2018The impact of classification systems in the evaluation of the research performance of the Leiden Ranking universitiesJournal of the Association for Information Science and Technology, 69 (2018), pp. 1046-1053
- Qiu, 2010Publish or perish in China: The pressure to rack up publications in high-impact journals could encourage misconduct, some sayNature, 463 (2010), pp. 142-144
- Shen, Ma, & Wang, 2018A Web-scale system for scientific knowledge exploration. Annual Meeting of the Association for Computational Linguistics (2018)
- Silver et al., 2017Mastering the game of go without human knowledgeNature, 550 (2017), pp. 354-359
- Sinatra et al., 2016Quantifying the evolution of individual scientific impactScience (New York, N.Y.), 354 (2016), p. aaf5239
- Sinha et al., 2015An overview of microsoft academic service (mas) and applicationsProceedings of the 24th international conference on world wide web (2015), pp. 243-246
- Sutton and Barto, 2018Reinforcement learning: An introductionMIT Press (2018)
- Tang et al., 2008Arnetminer: Extraction and mining of academic social networksProceedings of the 14th ACM SIGKDD international conference on knowledge discovery and data mining (2008), pp. 990-998
- Van Hasselt et al., 2016Deep reinforcement learning with double q-learningProceedings of the AAAI conference on artificial intelligence (2016)
- Wallace and Ràfols, 2018Institutional shaping of research priorities: A case study on avian influenzaResearch Policy, 47 (2018), pp. 1975-1989
- Wang et al., 2013Quantifying long-term scientific impactScience (New York, N.Y.), 342 (2013), pp. 127-132
- Wang et al., 2016Dueling network architectures for deep reinforcement learningProceedings of the PMLR International Conference on Machine Learning (2016), pp. 1995-2003
- Watkins and Dayan, 1992Q-learningMachine learning, 8 (1992), pp. 279-292
- Watkins, 1989Learning from delayed rewardsRobotics & Autonomous Systems, 15 (4) (1989), pp. 233-235
- Wei et al., 2013Do scientists trace hot topics?Scientific Reports, 3 (2013), pp. 1-5
- Yu et al., 2021Become a better you: Correlation between the change of research direction and the change of scientific performanceJournal of Informetrics, 15 (2021), Article 101193
- Zeng et al., 2019Increasing trend of scientists to switch between topicsNature Communications, 10 (2019), pp. 1-11
- Zhang et al., 2019Oag: Toward linking large-scale heterogeneous entity graphsProceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining (2019), pp. 2585-2595
- Zhu, Liu and Yang, 2021Team size, research variety, and research performance: do coauthors’ coauthors matter?Journal of Informetrics, 15 (2021), Article 101205
Cited by (1)
An author credit allocation method with improved distinguishability and robustness
2023, Journal of Data and Information Science