Main content

Home

Menu

Loading wiki pages...

View
Wiki Version:
## **RDA 2019** #### **Finland, Helsinki | 23 October 2019** --- #### **BoF: Curating for FAIR and Reproducibile Data and Code** ##### **Limor Peer, Institution for Social And Policy Studies (ISPS), Yale University** ##### **Thu-Mai Christian, Odum Institute, University of North Carolina at Chapel Hill** ##### **Florio Arguillas, Cornell Institute for Social and Economic Research (CISER), Cornell University** Scientific reproducibility provides a common purpose and language for data professionals and researchers. For data professionals, reproducibility can be a framework to hone and justify curation actions and decisions, and for researchers it offers a rationale for inserting best practices early into the research lifecycle. Curating for reproducibility (CURE) includes activities that ensure that statistical and analytic claims about given data can be reproduced with that data. Academic libraries and data archives have been stepping up to provide systems and standards for making research materials publicly accessible, but the datasets housed in repositories rarely meet the quality standards required by the scientific community. Even as data sharing becomes normative practice in the research community, there is growing awareness that access to data alone--even well-curated data --is not sufficient to guarantee the reproducibility of published research findings. Computational reproducibility, the ability to recreate computational results from the data and code used by the original researcher, is a key requirement to enable researchers to reap the benefits of data sharing ([Stodden et al., 2013][1]), but one that recent reports suggest is not being met. Verifying findings to confirm the integrity of the scientific record and to build upon previous work to discover and develop new innovations also requires access to the analysis code used to produce reported results. The exhaustive laundry list of tasks that characterize the traditional data curation workflow that enables data access--file review and normalization, metadata generation, assignment of persistent IDs, data cleaning, and assembly of contextual documentation--falls short when research reproducibility is the ultimate goal ([Peer et al., 2014][2]). In order to curate for reproducibility, activities must include a review of the computer code used to produce the analysis to verify that the code is executable and generates results identical to those presented in associated publications. CURE has been implementing practices and developing workflows and tools that support curating for reproducibility. [Presentation Slides][3] [Collaborative Session Notes][4] [1]: https://doi.org/10.1371/journal.pone.0067111 [2]: https://doi.org/10.2218/ijdc.v9i1.317 [3]: https://osf.io/brg6e/ [4]: https://docs.google.com/document/d/1IenuTqrcIK-ktBkohr1ypvGkTjXmhCcSOLHLhaAX8Qg/edit
OSF does not support the use of Internet Explorer. For optimal performance, please switch to another browser.
Accept
This website relies on cookies to help provide a better user experience. By clicking Accept or continuing to use the site, you agree. For more information, see our Privacy Policy and information on cookie use.
Accept
×

Start managing your projects on the OSF today.

Free and easy to use, the Open Science Framework supports the entire research lifecycle: planning, execution, reporting, archiving, and discovery.