“If I have seen further it is because I have stood on the shoulders of giants”
(Isaac Newton)
Reproducible research practices include rigorously controlled and documented experiments using validated reagents. These practices are integral to the scientific method, and they enable acquisition of reliable and actionable research results. However, the art and practice of science is affected by challenges that go beyond the inherent complexity of the biology being explored. The pressures to publish, the focus on novel, positive, and impactful results, the use of suboptimal research practices, and the scarcity of research funding likely contribute to unacceptable levels of irreproducible scientific results (Freedman, Venugopalan, & Wisman, 2017).
A recent survey conducted by Springer (2016) reported that 90% of participants identified “more robust experimental design” as one of several key improvements needed for the conduct of better science, in addition to “better statistics” and “better mentorship”. More recently, Nature (2018) published the survey data which can be found on Figshare.
A manifesto for reproducible science made several recommendations and called attention to initiatives such as the Transparency and Openness Promotion (TOP) guidelines created by the Center for Open Science to improve research planning and reporting (Munafò, 2017), which was supported by over 5000 journals and research organizations.
On June 9, 2015, the U.S. National Institutes of Health (NIH) published a notice1 that identified four areas for improvement that are now required to be addressed by scientists in grant applications, as follows:
scientific premise forming the basis of the proposed research
rigorous experimental design for robust and unbiased results
consideration of sex and other relevant biologic variables
authentication of key biologic and chemical resources
The Association of Biomolecular Resource Facilities (ABRF) Committee on Core Rigor and Reproducibility (CCoRRe) recently conducted a survey to assess how shared resource facilities are currently assisting investigators with their need to demonstrate transparency and rigor in their research. In addition, the survey captured information from the shared resource personnel related to the challenges they face and the resources they need to support scientific transparency, rigor, and reproducibility.
The CCoRRe committee developed an 18-question online survey and shared it using SurveyMonkey. The survey was announced on the ABRF listservs and blogs and was open from February to April 2017. All survey participants remained anonymous.
The survey contained both multiple-choice and open-ended text questions. The open-ended text questions were categorized and coded following an inductive coding approach with at least two independent coders. Using the sampling error formula:
\[ e = { Zp(1-p) \over \sqrt{n} } \]
we compute that at a 95% confidence level (i.e., \(Z=1.96\)), with base probability \(p=1/2\) and sample size \(n=242\), the margin of error is ±3%.
A total of 242 individuals from 21 countries completed this section. The majority of the survey participants are core facility directors or managers (69%) and work in an academic setting (72%) in the United States (79%).
Optimal research core services require a full commitment to rigorous methods as an obligation and not as a choice. However, more than half of the participants identified lack of funding and technical staff training as primary deterrents to the implementation and maintenance of R&R initiatives (See Table 3.1). This illustrates the difficulties that fee-for-service core facilities face when considering the costs associated with establishing new standardized procedures, methods, or technologies, or improving documentation, transparency, and quality control.
Participants were asked to suggest solutions to mitigate or eliminate challenges and provide a clear path to improved R&R in cores. Some of the responders did emphasize the need for the development of universal guidelines and SOPs that would facilitate the consistent adoption across a technology or application and would incentivize investigators to comply with such published R&R guidelines (See Figure 3.2).
About 40% proposed that funding mechanisms should be available to cores from either discretionary institutional funds or federal agencies to promote and support R&R initiatives. It is clear that survey respondents believe that it is important to identify funding mechanisms to help core service providers become more visible as scientific experts, partners, and educators with the ability to directly influence research quality.
About one quarter of respondents noted that a radical cultural change at the highest institutional levels is necessary to support and foster research rigor and transparency. Research institutions, journals, and funding agencies must be willing to establish clear requirements as well as mandate and “provide incentives to support and monitor research rigor throughout the research life cycle” These “cultural” observations related to research culture and incentives were frequently noted in previously published reviews of the research reproducibility issue (Baker, 2016; Freedman et al., 2017).
Scientific shared resources support research laboratories to generate critical data across many disciplines. Core personnel maintain considerable expertise that is important for the quality of their work and for sharing with research scientists in their important role as research mentors. They ensure continuous improvement through professional and educational development and through their systematic approach to research methods.
1Through these four elements, the NIH intends to “enhance the reproducibility of research findings through increased scientific rigor and transparency” (https://ori.hhs.gov/images/ddblock/ORI%20Data%20Graphs%202006-2015.pdf)
Baker, M. (2016). 1,500 scientists lift the lid on reproducibility. Nature News, 533(7604), 452. Retrieved November 13, 2020, from http://www.nature.com/news/1-500-scientists-lift-the-lid-on-reproducibility-1.19970
Bustin, S. A. (2014). The reproducibility of biomedical research: Sleepers awake! Biomolecular Detection and Quantification, 2, 35–42. Retrieved November 13, 2020, from http://www.sciencedirect.com/science/article/pii/S2214753515000030
Freedman, P., Venugopalan, G., & Wisman, R. (2017). Reproducibility2020: Progress and priorities. F1000Research, 6. Retrieved November 13, 2020, from https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5461896/
Munafò, M. R. et a. (2017). A manifesto for reproducible science. Nature Human Behaviour, 1(1), 1–9. Retrieved November 13, 2020, from https://www.nature.com/articles/s41562-016-0021
Nature. (2018). Nature Reproducibility survey 2017. Figshare. Repository,. Retrieved from 10.6084/m9.figshare.6139937.v4
Springer, N. (2016). Reality check on reproducibility. Nature, 533(7604), 437.