- Review
- Open access
- Published:
From 2015 to 2023, eight years of empirical research on research integrity: a scoping review
Research Integrity and Peer Review volume 10, Article number: 5 (2025)
Abstract
Background
Research on research integrity (RI) has grown exponentially over the past several decades. Although the earliest publications emerged in the 1980 s, more than half of the existing literature has been produced within the last five years. Given that the most recent comprehensive literature review is now eight years old, the present study aims to extend and update previous findings.
Method
We conducted a systematic search of the Web of Science and Constellate databases for articles published between 2015 and 2023. To structure our overview and guide our inquiry, we addressed the following seven broad questions about the field:-What topics does the empirical literature on RI explore?
-
What are the primary objectives of the empirical literature on RI?
-
What methodologies are prevalent in the empirical literature on RI?
-
What populations or organizations are studied in the empirical literature on RI?
-
Where are the empirical studies on RI conducted?
-
Where is the empirical literature on RI published?
-
To what degree is the general literature on RI grounded in empirical research?
Additionally, we used the previous scoping review as a benchmark to identify emerging trends and shifts.
Results
Our search yielded a total of 3,282 studies, of which 660 articles met our inclusion criteria. All research questions were comprehensively addressed. Notably, we observed a significant shift in methodologies: the reliance on interviews and surveys decreased from 51 to 30%, whereas the application of meta-scientific methods increased from 17 to 31%. In terms of theoretical orientation, the previously dominant “Bad Apple” hypothesis declined from 54 to 30%, while the “Wicked System” hypothesis increased from 46 to 52%. Furthermore, there has been a pronounced trend toward testing solutions, rising from 31 to 56% at the expense of merely describing the problem, which fell from 69 to 44%.
Conclusion
Three gaps highlighted eight years ago by the previous scoping review remain unresolved. Research on decision makers (e.g., scientists in positions of power, policymakers, accounting for 3%), the private research sector and patents (4.7%), and the peer review system (0.3%) continues to be underexplored. Even more concerning, if current trends persist, these gaps are likely to become increasingly problematic.
Background
Since the 1990 sresearch integrity (RI) has emerged as a significant concern, undergone institutionalization, and evolved into a distinct field of research. The increasing number of retractions is now exceeding 10,000 per year [1]. This has sparked debates regarding whether this trend signals growing misconduct within the scientific community or reflects enhanced vigilance and improved regulatory mechanisms. For a broader discussion on these interpretations, see Oransky [2]. This debate is emblematic of the diverse issues addressed in the burgeoning RI literature, which now encompasses substantial corpora of both empirical and theoretical work. A simple search on Web of Science using the term “research integrity” returns thousands of results. We quickly realized that examining the theoretical and empirical literature would be too extensive for our study. As such, we chose to follow the footstep of the previous scoping review and focus on the empirical corpus. Young and rapidly expanding fields such as RI can greatly benefit from studies seeking to understand their evolution. Such research can identify emerging trends, pinpoint knowledge gaps, and enable researchers to adopt a more reflective stance. To that end, we structured our review around seven thematic questions:
-
Q1: What topics does the empirical literature on RI explore? Does it focus on fabrication, falsification, and plagiarism (FFP), more subtle malpractice like questionable research practices (QRP), or something else entirely?
-
Q2: What are the primary objectives of the empirical literature on RI? Is it looking for solutions or does it try to establish a diagnostic? Does it focus on individual transgressions or systemic flaws?
-
Q3: What methodologies are prevalent in the empirical literature on RI? How are empirical investigations conducted, and to what extent are experimental methods used?
-
Q4: What populations or organizations are studied in the empirical literature on RI? How much is it focused on researchers and which kinds?
-
Q5: Where are the empirical studies on RI conducted? Which countries publish the most research on RI?
-
Q6: Where is the empirical literature on RI published? Which journals publish the most empirical research on RI? Which disciplines contribute the most to RI literature?
-
Q7: To what degree is the general literature on RI grounded in empirical research? What is the proportion of empirical versus theoretical contributions?
The most suitable methodology for answering broad questions about a research field is a scoping review. Munn et al. [3] describe this method as follow: “scoping reviews are an ideal tool to determine the scope or coverage of a body of literature on a given topic and […] an overview (broad or detailed) of its focus”. Unfortunately, such studies are rare in the field of RI. The most recent general systematic review, A Decade of Empirical Research on Research Integrity: What Have We (Not) Looked At? by Aubert Bonn & Pinxten [4], covers the period from 2005 to 2015. Subsequent overviews, such as those by Neoh et al. [5], Mansour & Ruphy [6], and Ali et al. [7], either focus solely on co-citation pattern analysis, lack a systematic approach, or focus on academic integrity (e.g., student cheating or RI training below the PhD level). Given the surge in publications in recent years, our study fills an important gap by examining empirical RI research from 2015 to 2023 and contrasting these findings with earlier trends.
Methods
The methodology used in this study was designed with two objectives: to provide accurate answers to the thematic questions on RI and to maximize the validity of comparisons with the previous scoping review. To achieve these aims, we based our approach on the methodology of Aubert Bonn & Pinxten [4], making as few modifications as possible. Each modification and its rationale are detailed in tables in the Additional File 2. These tables can be used to convert the data between the two studies and estimate the validity of the comparisons. Every comparison made is valid according to the bias estimation reported in the Additional file 4.
As a scoping review, our methodology can be divided into four steps: selecting databases, automatically identifying studies through keyword searches, screening out irrelevant papers, and classifying the included studies. Both the screening and classification processes were carried out by the first author. For a more detailed description of standard scoping review methodology, see chapter 10 of the JBI Manual for Evidence Synthesis [8]. This study also adheres to the PRISMA guidelines for scoping reviews [9], with the checklist available in Additional file 1.
Database selection
We utilized two databases: Web of Science and Constellate. Web of Science indexes over 24,000 journals, making it indispensable for systematic reviews. However, as demonstrated by Martín-Martín et al. [10], Chavarro et al. [11], and Mongeon & Paul-Hus [12], works from the humanities and social sciences are not always possible to retrieve. To address this limitation, we included Constellate, a database curated by JSTOR that specializes in the humanities and social sciences.Footnote 1 For a discussion comparing our selected databases with those used by Aubert Bonn & Pinxten [4], see Additional file 4.
Identification
Aubert Bonn & Pinxten [4] noted two challenges for scoping reviews in the field of RI: inconsistent definitions of RI and diverse publication formats (pp. 2). As of 2024, there have been significant improvements in defining RI. Notable resources include national integrity codes and offices, such as the OFIS (Office Français de l’Integrité Scientifique) [13] and the European Code of Conduct for Research Integrity by ALLEA (All European Academies) [14]. The latter defines RI as:
“The basic responsibility of the research community to formulate the principles of research, to define the criteria for proper research behavior, to maximize the quality, reliability, and robustness of research and its results, and to respond adequately to threats to, or violations of, good research practices.” (pp 3).
Despite these advances, it is unclear whether this definition has been uniformly adopted by the empirical RI research community. To mitigate potential biases, we incorporated a broad set of keywords from related fields that might be used in RI literature.Footnote 2
Regarding the diversity of publication formats in the field, Aubert Bonn & Pinxten [4] identified documents classified as commentaries or notes that presented original empirical data. This means that by searching only for peer-reviewed articles we could miss includable papers. To address this, we broadened our search criteria to include any type of document allowed by the databases.Footnote 3
The keywords were chosen with two objectives in mind: to identify RI studies and to filter for empirical research. We assumed that empirical studies would consistently include headings such as “Method” and “Result”, while theoretical studies might not. To follow Aubert Bonn and Pinxten’s methodology and to manage the volume of studies given our available resources, only English-language works were included. The search results were then automatically filtered to include studies published between 2016 and 2023 included. The exact queries used for both databases can be found in Additional File 4 (Fig. 1).
Study flow diagram. Legend. Study flow diagram recommended by PRISMA [15] representing the flow of information through the phases of the scoping review. List of keywords: responsible innovation, questionable research practices, research misconduct, scientific integrity, responsible research, responsible conduct of research, scientific misconduct, research integrity, scientific fraud, metascience, method, methods, result, results
Screening
We included every study whose research topic aligns with ALLEA’s definition of RI. This encompasses research mentioning fabrication, falsification, and plagiarism (FFP), as well as questionable research practices (QRP).Footnote 4 It also includes FFP/QRP issues arising in the private research sector, as well as integrity training from the PhD student level. We excluded all other topics, even those closely adjacent. In particular, academic integrity (e.g., student cheating or RI training below the PhD level), field-specific methodological enhancements indirectly related to RI, and studies on research ethics (e.g., protecting the common good, addressing safety concerns, or strengthening the connection between society and science) or responsible scienceFootnote 5 (e.g., combining research ethics and RI).
A second set of inclusion criteria was applied to assess the empirical nature of the studies. Articles that did not involve data collection, manipulation or that utilized non-original data for illustrative or argumentative purposes, were excluded. This last condition often led to the exclusion of news articles and editorials. However, studies that designed a theoretical tool and testing it through examples or case studies were considered to generate new data and included. Lastly we conducted a brief correspondence test based on the Predatory Journal List [16] to exclude articles from predatory journals, as we did not want to indirectly promote these types of journals. Moreover, such papers are less likely to have undergone rigorous peer review.
Classification process
As previously mentioned, we based our classification process on the work of Aubert Bonn & Pinxten [4]. However, we implemented modifications to align the methodology with our seven thematic questions, ensure consistency with our definition of RI, or adhere to contemporary naming conventions. Details of these modifications and their justifications can be found in Additional file 2. In total, there are 10 categories and over 50 characteristics. Their detailed description is available in Additional file 4.
Results
Screening
From the 3,282 articles initially identified, 20% met the inclusion criteria and were incorporated into the study. The complete dataset can be found in Additional file 3. In the following sections, we present the answers to each of the seven questions. Comparisons with the previous scoping review are addressed in the Discussion section.
-
Q1: What topics does the empirical literature on RI explore? Does it focus on fabrication, falsification, and plagiarism (FFP), more subtle malpractice like questionable research practices (QRP), or something else entirely? (Fig. 2)
Unsurprisingly, the literature predominantly addresses the direct manifestations of the lack of RI, specifically FFP and QRP. Studies of this nature account for 62% of the total. QRP are featured in 43.8% of the literature (Only QRP 29,8% + FFP and QRP 28%/2) and are the most extensively researched topic, while FFP studies only appears in 19% of the articles (Only FFP 5% + FFP and QRP 28%/2). Another notable finding is the very low number of empirical studies on peer review, which account for only 0.3%
-
Q2: What are the primary objectives of the empirical literature on RI? Does the empirical literature on RI primarily focus on identifying individual transgressions or systemic flaws? Does it focus on individual transgressions or systemic flaws? (Fig. 3)
Fig. 3 Aims and origin of the misconduct. Legend. This category's objective is to characterize the aim of the studies. In addition, three aim characteristics: assess an approach efficacy, research of determinants and capacity building are sub-divided into systemic factors, individual factors and mixed. The sub-division goal is to determine where does the studies place the origin of the misconduct. *The “Total of the occurrence studies” striped line is the addition of lines present elsewhere in the figure (“Occurrence of QRP”, “Occurrence of FFP” and “Occurrence of FFP and QRP”). The total of articles (100%) is 660. The total of articles for the sub-division (100%) is 335.5. RI = research integrity; QRP = Questionable research practices; FFP = Fabrication, falsification, plagia
Most studies aim to identify the best approaches for improving research integrity (RI). More specifically, 19.6% focused on assessing the efficacy of an approach, while 18% seek to find new strategies (capacity building). Additionally, 19.3% of the studies look into the occurrence of misconduct. In terms of using a systemic or an individual framework, 52.3% of the studies focus exclusively on systems, 29.9% on individual, and 17.7% consider both perspectives.
-
Q3: What methodologies are prevalent in the empirical literature on RI? How does one empirically study RI? Are experiments possible? (Fig. 4)
The most favored methodology for studying research integrity is metascience, which accounts for 31.5% of the research. Close behind are quantitative surveys and qualitative interviews/focus groups, making up 29.7% of the studies. Surprisingly, experiment/simulation/quasi-experiment studies come third, representing 17.3% of the total.
-
Q4: What populations or organizations are studied in the empirical literature on RI? How much is it focused on researchers and which kinds? (Fig. 5)
As expected, an overwhelming majority of studies (73.3%) focus on scientists. Notably, the emphasis is primarily on public researchers, who constitute 67.9% of the total. In contrast, scientists from the private sector account only for a small fraction (2.6%). Research on decision-makers is likewise limited, with few studies on policymakers (1.5%), or researchers in leadership positions (1.4%). Similarly, research integrity officers are scarcely studied (1.4%), with most investigations concentrating solely on their responsibilities for teaching research integrity rather than their roles in investigating misconduct.
-
Q5: Where are the empirical studies on RI conducted? Which countries publish the most research on RI? (Fig. 6)
Fig. 6 Top 15 of national affiliation of the author. Legend. * The “European Union” striped line is the addition of lines present elsewhere in the figure (“Netherlands”, “Germany”, “Spain”, “Belgium” and “Croatia”) and of other countries not included in the figure. The 43 countries not in the top 15 (totaling 133 studies) are not represented but can be found in the Additional file 3. The total of articles (100%) is 660
The United States is by far the single country with the most publication on RI, representing 32.1% of the total. Northern Europe countries like the United Kingdom, Netherlands and Germany takes the rest of the podium. European Union as a whole comes second with having published 26.7% of articles on RI (Fig. 7).
Percentage of the RI articles in the national scientific and technical journal articles. Legend. The average of the countries with at least one paper is 0,03%. The 43 countries with less than 4 studies are not represented but can be found in the Additional file 3. The data for the scientific and technical journal articles by countries come from the World Bank Data [17]. The total of articles (100%) is 660. RI = research integrity
After normalizing for research output, the United States drops to the 14 th rank and Croatia takes the top spot by a significant margin. The Netherlands, recognized within the European research integrity community for its leadership in institutionalizing RI, takes second place. Jordan is third place and stands out as the only Middle Eastern country above the average.
-
Q6: Where is the empirical literature on RI published? Which journals publish the most empirical research on RI? Which disciplines contribute the most to RI literature? (Fig. 8)
Fig. 8 Top 10 journals by the number of empirical publications on RI. Legend. There was a total of 333 (50%) unique journal occurrences (i.g 50% of journal who have published on RI only published a single article on RI.) Only the top 10 journal that published the most on RI are named in the figure. The full list of journals can be found in the Additional file 3. The total of articles (100%) is 660
Most articles (53%) were published in journals that contributed only one article on RI. This is indicative of a highly dispersed field of research. Nevertheless, a clearly identifiable core of specialized journals exists. The top 10 most productive journals account for 33.6% of the total output. Four of the top five journals specialize in RI, while the fifth, Plos one is a mega journal (e.g. a journal with a very broad coverage. See the work of Lăzăroiu for more details on mega journal [18]). As the field continues to grow and establish itself as an independent domain of research, it will be interesting to observe whether the number of unique journals occurrences decrease, reflecting a trend toward more specialization (Fig. 9).
A clear characteristic of the empirical RI corpus is its diversity, encompassing nearly all scientific fields. The natural sciences represent a slight majority, accounting for 51.8%. The three dominant domains are then: medicine (30.8%), philosophy (16.5%), and psychology (10%).
-
Q7: To what degree is the general literature on RI grounded in empirical research? What is the proportion of empirical versus theoretical research? (Table 1)
Table 1 Empirical coverage
Discussion
By comparing the seven questions covered in the previous section, with the Aubert Bonn and Pinxten [4] study, we can find additional insights about the evolution of the RI field over the past decade. In doing so, we identified six noteworthy trends.
Our findings indicate that empirical studies constitute 33.5% of the total literature on RI.
A more systematic approach
Two intuitive hypotheses regarding the origin of scientific misconduct are prevalent in the RI literature. One attributes the issue to the occasional"Bad apple"and emphasizes education and training as potential solutions. The other attributes misconduct to the way science is organized, referring to it as a"Wicked system". This perspective favors structural or systemic reforms as their preferred solution. They are frequently referred to in practice, as seen among science journal editors [19] and RI research programs like “Beyond Bad Apples” [20]. In their study, Aubert Bonn & Pinxten [4] highlighted a"Mismatch Between What We Know and What We Propose,"revealing that while the determinant research focused on the “Wicked system” (systemic causes), the proposed solutions were using the “Bad apple” model (targeted individuals). Eight years later, our findings show that this discrepancy has significantly diminished. Most recent studies (52%) advocate for systemic solutions.. This marks an improved alignment between the types of causes investigated and the solutions proposed. The prominence of the “Bad Apple” hypothesis has declined from 54 to 30%, whereas the “Wicked System” hypothesis has risen from 46 to 52%. Moreover, we observed a sharp decline in research focused on"RI training, education, and mentorship,"which dropped from 18% to just 3,8% of the topics studied. This shift can be linked to the reduced focus on the"Bad apples"hypothesis, which traditionally emphasized education, and a growing emphasis on the"Wicked system"hypothesis, which advocates for organizational reforms.
Increased methodological diversity
Aubert Bonn & Pinxten [4] reported that, surveys, interviews, and focus groups constituted half (51%) of the methodologies used, while in the current study, they account for only 30%. This decrease can be interpreted as an improvement, as surveys on sensitive topics like research misconducts are known to be prone to biases that are difficult to control [21, 22]. The decreased reliance on these methods has allowed for greater diversification in research approaches, particularly benefiting meta-science methodologies. The use of the metascience methodology has increased significantly, rising from the third most employed approach (17%) to the most commonly used method (31.5%). Additionally, we were surprised by the experiment/simulation/quasi-experiment methodologies being now the third most represented (17.3%). This is higher than we anticipated given the inherent challenges associated with conducting experiments in this field. Most of these studies are kinds of modeling of science.
From description to solution
The focus of RI literature has shifted notably over the past decade. In the Aubert Bonn & Pinxten [4] study, the main focus was on"describing the problem"and"strengthening reporting standards,"which together represented 69% of the research at the time. This earlier emphasis was primarily on understanding the scope of RI issues. However, between 2015 and 2023, the focus has evolved toward solution-oriented research. The two most common themes in our recent period are"Assessing an approach's efficacy"and"Capacity building,"both of which aim to identify or create effective ways to improve RI. These themes now account for 56,3% of the research output. Descriptive studies, which once dominated, have decreased down to 43,7%.Footnote 6 This shift shows how the field has matured as it went from identifying problems to seeking solutions.
Increased globalization
In the Aubert Bonn & Pinxten [4] study, only six countries had produced ten or more articles on RI. By contrast, in the current landscape, this number has more than doubled, reaching fourteen countries. The United States previously accounted for over half of the literature on RI, has seen its dominance decrease. Although it remains the most productive country, its total share of RI publications has dropped to 32%. This shift suggests that research on RI is becoming more globally distributed, even though North America and Northern Europe continue to lead both in absolute term and when normalized by overall research output. Notably, China ranks only 53rd, significantly below the average. To put this in perspective, relative to its research output, Croatia has published 100 times more papers on RI than China. This slow research output on RI appears to be an Asian phenomenon overall rather than being unique to China. The only exception in the region is Singapore, an early proponent of RI and host of the Second World Conference on Research Integrity [23].
Empirical coverage of the field
This study found that empirical works account for 33.5% of the total, closely matching the 35% reported by Aubert Bonn & Pinxten. The proportion of empirical work remains relatively low for a field that deals with practical concerns. Aubert Bonn & Pinxten had already noted this discrepancy in their 2019 article and proposed an explanation that may still holds true today:
“Few of the authors of articles on RI are engaged full time in RI, and that collaborators and target audience sometimes spread through an array of distinct disciplines, it may still be challenging to engage in empirical works on the topic.” (pp 12).
Unresolved and newly identified gaps
We identified three points that could inspire future research in the field of RI. Two of these were already noted in the previous scoping review: the lack of studies on peer review and decision-makers. The last one, a lack of research on the private sector is newly identified.
Given the heavy criticism that peer review has faced (see [24, 25] for examples), along with the rise of various alternatives like registered reports, preprints, and platforms like PubPeer we expected peer review to receive more attention. Despite these factors, there is very little empirical research on peer review in relation to RI. Comparative studies examining the effectiveness of different peer review models would greatly benefit the cause of RI but are notably absent. While it would be methodologically challenging to systematically assess the effectiveness of different peer review approaches or their alternatives, such research would be both innovative and highly valuable for the scientific community.
The lack of studies on decision-makers and the private research sector may be attributed to the difficulty of accessing these groups. Unfortunately, this gap in the literature is increasingly problematic since both groups are becoming more important. Few studies have looked at the organizational or leadership level, despite the key role these individuals play in guiding the research community toward greater integrity, particularly in the context of the growing emphasis on the "Wicked systems" hypothesis. The lack of research on the private sector is likewise concerning. It is roughly three times larger than the public research sector. According to an OECD report [26], in 2020, the public sector accounted for only 25.5%Footnote 7 of research performed in OECD countries, while the private sector made up 74.5%.Footnote 8 The private sector is also not immune to research misconducts. One prominent example is the Industrial Bio-Test Laboratories (IBTL) case, where the company likely falsified 618 clinical trials and toxicity tests in the 70 s [27]. More recently, the highly publicized Theranos case involved faked blood-testing technologies, with irregularities surrounding falsified patents, as discussed by Vincent [28]. On the topic of patents, it is particularly concerning that we found no articles mentioning them, even though they are increasingly seen as equivalent to publications in terms of individual career advancement or institutional prestige.
RI literature should pay greater attention to decision-makers and private researchers. Addressing these gaps is essential for maintaining the relevance of the field in light of evolving research dynamics.
Study limitation
Our study did not adhere to a formal review protocol, meaning that no preregistration was done. Our approach evolved too rapidly at the beginning of the project. Although we initially planned to replicate the methodology of Aubert Bonn & Pinxten [4], it soon became clear that significant modificationsFootnote 9 were necessary to better address our research questions. These adjustments were made throughout the project, which made preregistration incompatible with our exploratory approach. The comparisons done in the discussion section with the Aubert Bonn & Pinxten [4] study should be interpreted with caution. Even if we started with their study as a template, we used different databases, methodologies, and classifications. See additional file 2 for a detailed comparison of our methodologies.
This study may have missed valid publications for four reasons.
1) It is important to acknowledge the issue of articles not being immediately indexed by databases, a phenomenon known as the"rolling wall."Since our study was conducted in early 2024 and includes studies published until late 2023, it's likely that some relevant studies may have not yet been indexed. This issue is particularly pronounced with Constellate, which has a slower rolling wall compared to Web of Science.
2) While including other databases such as Google Scholar, Scopus, and PubMed would have enhanced the comprehensiveness of our study, these were not used due to limited resources and the minimal expected gains, as detailed in the additional file 4.
3) We excluded grey literature (e.g., works that did not passed academic peer review). This is due to the fact that our databases (Web of science and Constellate) are unlikely to have systematically indexed grey literature which open the door to a selection bias.
4) Finally, we excluded articles that were in preprints or listed in the Predatory Journal List [16], although some of these might be valid. While we did not assess the quality of the included works, we aimed to ensure that only articles meeting minimal peer review standards were considered.
The process of determining how to categorize each study always involves a degree of subjectivity. This limitation is further compounded by the absence of an independent reviewer, a factor known to reduce bias [29]. While the study would have been stronger with an independent duplicate, the individual categorization of each of the 3,282 articles is organized and openly available in additional file 3. We made it so anyone can freely verify or replicate the categorization process.
Apart from a very basic correspondence check using the Predatory journal list [16] to eliminate articles from predatory journals, we did not assess the quality of the studies. Given that our analysis focuses on research topics and methodologies rather than on study outcomes, the quality of the studies will not affect our findings.
The empirical coverage was approximated rather than directly calculated. See the additional file 3 for the calculations and the additional file 4 for the methodology.
Limiting our search to English-language articles likely introduced a bias, as certain countries publish a significant portion of their research in their native languages. This is especially relevant for China. The Aubert Bonn & Pinxten [4] study faced the same limitation: “China, which is rapidly becoming an important player in scientific publishing worldwide, was scarcely represented in our sample. It is possible that the language limitations of our study (i.e., we only included articles in English) contributed to this disparity” (pp 12).
Conclusion
Our findings highlight the growing importance of meta-science methodologies, a shift in focus toward problem-solving, and a move away from the"Bad apple"in favor of the"Wicked system"hypothesis.
We also identified three research gaps likely to become increasingly important: research on decision-makers (such as scientists in leadership roles, policymakers, and institutions), the private research sector as well as patents, and the peer review system.
Data availability
The datasets supporting the conclusions of this article are included within the article and its additional files.
Notes
In hindsight, Constellate only had 8.2% more articles from the humanities and social sciences on research integrity than Web of Science.
“responsible innovation”, “responsible research”, “responsible conduct of research” and “metascience”.
It is important to note that we still exclude grey literature (e.g., works that did not passed academic peer review). This is due to the fact that both databases are unlikely to have systematically indexed grey literature which open the door to a selection bias.
The most recent ALLEA code (14) no longer use the QRP term but URP (unacceptable research practice) instead. We choose to stick with the QRP terminology as the literature has not yet adopted the new term.
Although it can be arduous to advocate for a clear-cut distinctions between RI, research ethics, and responsible science, this study requires it so as not to stretch itself beyond its means.
For more details about the comparisons see the additional file 2.
Percentage of Gross domestic expenditure on R&D (GERD) performed by the higher education.
sector 16.2% + performed by the government sector 9.3%
Percentage of Gross domestic expenditure on R&D (GERD) performed by the business enterprise.
sector 72.1% + performed by the private non-profit sector 2.4%
All of the modifications and their motivations can be found in the additional file 2.
Abbreviations
- RI:
-
Research Integrity
- FFP:
-
Fabrication, Falsification, Plagia
- QRP:
-
Questionable Research Practice
- URP:
-
Unacceptable Research Practice
References
Van Noorden R. More than 10,000 research papers were retracted in 2023 — a new record. Nature. 2023;624(7992):479–81. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/d41586-023-03974-8.
Oransky I. Retractions are increasing, but not enough. Nature. 2022;608(7921):9. https://doiorg.publicaciones.saludcastillayleon.es/10.1038/d41586-022-02071-6.
Munn Z, Peters MDJ, Stern C, et al. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. 2018;18:143. https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12874-018-0611-x.
Aubert Bonn N, Pinxten W. A decade of empirical research on research integrity: what have we (not) looked at? J Empir Res Hum Res Ethics. 2019;14(4):338–52. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/1556264619858534.
Neoh MJY, Carollo A, Lee A, Esposito G. Fifty years of research on questionable research practises in science: quantitative analysis of co-citation patterns. R Soc Open Sci. 2023;10(10). https://doiorg.publicaciones.saludcastillayleon.es/10.1098/rsos.230677.
Mansour A, Ruphy S. Tendances de la recherche sur l’intégrité scientifique. Lato Sensu. 2022;9(1):1‑16. https://doiorg.publicaciones.saludcastillayleon.es/10.20416/lsrsps.v9i1.1.
Ali I, Sultan P, Aboelmaged MG. A bibliometric analysis of academic misconduct research in higher education: Current status and future research opportunities. Account Res. 2020;28(6):372–93. https://doiorg.publicaciones.saludcastillayleon.es/10.1080/08989621.2020.1836620.
Peters MDJ, Godfrey C, McInerney P, Munn Z, Tricco AC, Khalil, H. Scoping Reviews (2020). JBI Manual for Evidence Synthesis. JBI; https://doiorg.publicaciones.saludcastillayleon.es/10.46658/JBIMES-24-09
Tricco AC, Lillie E, Zarin W, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and Explanation. Ann Intern Med. 2018;169(7):467–73. https://doiorg.publicaciones.saludcastillayleon.es/10.7326/m18-0850.
Martín-Martín A, Thelwall M, Orduna-Malea E, Delgado L-Cózar E. Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations’ COCI: A multidisciplinary comparison of coverage via citations. Scientometrics. 2021;126(1):871–906. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11192-020-03690-4.
Chavarro D, Ràfols I, Tang P. To what extent is inclusion in the Web of Science an indicator of journal ‘quality’? Res Eval. 2018;27(2):106–18. https://doiorg.publicaciones.saludcastillayleon.es/10.1093/reseval/rvy001.
Mongeon P, Paul-Hus A. The journal coverage of Web of Science and Scopus: A comparative analysis. Scientometrics. 2016;106(1):213–28. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11192-015-1765-5.
OFIS. What is research integrity? OFIS; 2023. Available from: https://www.ofis-france.fr/what-is-scientific-integrity/. Accessed Jan 2024.
ALLEA - All European Academies. The European Code of Conduct for Research Integrity. 2023. ALLEA. Available from: https://allea.org/code-of-conduct/. Accessed Jan 2024.
PRISMA 2020 flow diagram. 2020. https://www.prisma-statement.org/prisma-2020-flow-diagram Accessed Jan 2024. https://doiorg.publicaciones.saludcastillayleon.es/10.26356/ECoC. Accessed Mar 2024.
Predatory Journals. Predatory Journals List 2023. Predatory Journals. Available from: https://predatoryjournals.org/news/f/predatory-journals-list-2023. Accessed Feb 2025.
World Bank Open Data. Scientific and technical journal articles. World Bank; 2020. https://data.worldbank.org/indicator/IP.JRN.ARTC.SC. Accessed Jan 2024.
Lăzăroiu G. Do mega-journals constitute the future of scholarly communication? Educ Philos Theory. 2017;49(11):1047–50. https://doiorg.publicaciones.saludcastillayleon.es/10.1080/00131857.2017.1300022.
Not bad Apples but bad systems: AAAS session looks ahead to National Academies Report on Integrity in Scientific Research. Sci Ed. 2016 Aug 24. https://www.csescienceeditor.org/article/not-bad-apples-but-bad-systems-aaas-session-looks-ahead-to-national-academies-report-on-integrity-in-scientific-research/. Accessed March 2024.
CORDIS. Beyond Bad Apples: towards a behavioral and Evidence-Based approach to promote research ethics and research integrity in Europe. Horizon Europe. 2023. https://doiorg.publicaciones.saludcastillayleon.es/10.3030/101094714.
John LK, Loewenstein G, Prelec D. Measuring the Prevalence of Questionable Research Practices With Incentives for Truth Telling. Psychol Sci. 2012;23(5):524–32. https://doiorg.publicaciones.saludcastillayleon.es/10.1177/0956797611430953.
Y. Xie, K. Wang, Y. Kong, Prevalence of research misconduct and questionable research practices: a systematic review and meta-analysis. Science and Engineering Ethics 2021; 27:41. https://doiorg.publicaciones.saludcastillayleon.es/10.1007/s11948-021-00314-9.
Singapore Statement. WCRIF; 2010. https://www.wcrif.org/guidance/singapore-statement. Accessed March 2024.
Reardon J, Crawford J, Allen K, Walsh L. The peer review system is broken. We asked academics how to fix it. The Conversation. 2022. Available from: https://theconversation.com/the-peer-review-system-is-broken-we-asked-academics-how-to-fix-it-187034.
Kelly J, Sadeghieh T, Adeli K. Peer Review in Scientific Publications: Benefits, Critiques, & A Survival Guide. EJIFCC. 2014;25(3):227–243. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4975196/.
OECD, “International Comparisons”, in Main Science and Technology Indicators. OECD Publishing 2023. Volume 2022 Issue 2. https://doiorg.publicaciones.saludcastillayleon.es/10.1787/db4ee0f7-en.
Rosner D, Markowitz G. “Ashamed to Put My Name to It”: Monsanto, Industrial Bio-Test Laboratories, and the Use of Fraudulent Science, 1969–1985. Am J Public Health. 2023;113(6):661–6. https://doiorg.publicaciones.saludcastillayleon.es/10.2105/AJPH.2023.307247.
Look V. Theranos Patent for Point of Care System. The Patent Geek. Archived from the original on 2021–02–08. https://web.archive.org/web/20210208200533/https://thepatentgeek.com/2019/05/01/theranos-patent-for-point-of-care-system/. Accessed Mar 2024.
Rethlefsen ML, Farrell AM, Osterhaus Trzasko LC, Brigham TJ. Librarian co-authors correlated with higher quality reported search strategies in general internal medicine systematic reviews. J Clin Epidemiol. 2015;68(6):617–26. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/j.jclinepi.2014.11.025.
Acknowledgements
We wish to thank Nathalie Voarino, a member of the Ofis team who shared her experience in designing scoping reviews.
Funding
This work was supported by the French Agence Nationale de la Recherche under Grant ANR-20-CE27-0016.
Author information
Authors and Affiliations
Contributions
BV did the conceptualization, methodology, investigation and writing of the original draft. AB and SR provided supervision throughout the process and some writing review & editing. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
41073_2025_163_MOESM2_ESM.xlsx
Additional file 2. Correspondence table. The list of motivated modifications to the classification process from the Aubert Bonn & Pinxten methodology.
41073_2025_163_MOESM4_ESM.docx
Additional file 4. Detailed methodology. Detailed methodology, most notably the search queries and the validity analysis of the comparisons.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Vendé, B., Barberousse, A. & Ruphy, S. From 2015 to 2023, eight years of empirical research on research integrity: a scoping review. Res Integr Peer Rev 10, 5 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41073-025-00163-1
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s41073-025-00163-1