Abstract
AbstractData management and publication are core components of the research process. An emerging challenge that has received limited attention in biology is managing, working with, and providing access to data under continual active collection. “Evolving data” present unique challenges in quality assurance and control, data publication, archiving, and reproducibility. We developed a evolving data workflow for a long-term ecological study that addresses many of the challenges associated with managing this type of data. We do this by leveraging existing tools to: 1) perform quality assurance and control; 2) import, restructure, version, and archive data; 3) rapidly publish new data in ways that ensure appropriate credit to all contributors; and 4) automate most steps in the data pipeline to reduce the time and effort required by researchers. The workflow uses two tools from software development, version control and continuous integration, to create a modern data management system that automates the pipeline.
Publisher
Cold Spring Harbor Laboratory
Reference36 articles.
1. Unmet needs for analyzing biological big data: A survey of 704 NSF principal investigators;PLOS Computational Biology,2017
2. Bergman, C. (2012, November 8). On the Preservation of Published Bioinformatics Code on Github. Retrieved June 1, 2018, from https://caseybergman.wordpress.com/2012/11/08/on-the-preservation-of-published-bioinformatics-code-on-github/
3. Brown, J. H. (1998). The Desert Granivory Experiments at Portal. In Experimental ecology: Issues and perspectives (pp. 71–95).
4. Early Warnings of Regime Shifts: A Whole-Ecosystem Experiment
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献