Comparability and reproducibility of biomedical data
Open Access
- 27 November 2012
- journal article
- research article
- Published by Oxford University Press (OUP) in Briefings in Bioinformatics
- Vol. 14 (4), 391-401
- https://doi.org/10.1093/bib/bbs078
Abstract
With the development of novel assay technologies, biomedical experiments and analyses have gone through substantial evolution. Today, a typical experiment can simultaneously measure hundreds to thousands of individual features (e.g. genes) in dozens of biological conditions, resulting in gigabytes of data that need to be processed and analyzed. Because of the multiple steps involved in the data generation and analysis and the lack of details provided, it can be difficult for independent researchers to try to reproduce a published study. With the recent outrage following the halt of a cancer clinical trial due to the lack of reproducibility of the published study, researchers are now facing heavy pressure to ensure that their results are reproducible. Despite the global demand, too many published studies remain non-reproducible mainly due to the lack of availability of experimental protocol, data and/or computer code. Scientific discovery is an iterative process, where a published study generates new knowledge and data, resulting in new follow-up studies or clinical trials based on these results. As such, it is important for the results of a study to be quickly confirmed or discarded to avoid wasting time and money on novel projects. The availability of high-quality, reproducible data will also lead to more powerful analyses (or meta-analyses) where multiple data sets are combined to generate new knowledge. In this article, we review some of the recent developments regarding biomedical reproducibility and comparability and discuss some of the areas where the overall field could be improved.Keywords
This publication has 50 references indexed in Scilit:
- Wisdom of crowds for robust gene network inferenceNature Methods, 2012
- T Cell Assays and MIATA: The Essential Minimum for Maximum ImpactImmunity, 2012
- Standardizing immunophenotyping for the Human Immunology ProjectNature Reviews Immunology, 2012
- Development and implementation of an international proficiency testing program for a neutralizing antibody assay for HIV-1 in TZM-bl cellsJournal of Immunological Methods, 2012
- LabKey Server: An open source platform for scientific data integration, analysis and collaborationBMC Bioinformatics, 2011
- Cancer Biomarkers: Can We Turn Recent Failures into Success?JNCI Journal of the National Cancer Institute, 2010
- Accessible Reproducible ResearchScience, 2010
- Per‐channel basis normalization methods for flow cytometry dataCytometry Part A, 2009
- Model-based analysis of tiling-arrays for ChIP-chipProceedings of the National Academy of Sciences of the United States of America, 2006
- GenePattern 2.0Nature Genetics, 2006