Introduction Reproducing an experiment is essential for scientific development. Recent publications have shown a high level of concern about irreproducibility of results [1–4]. A survey published in Nature [1], involving more than 1,500 researchers, showed that more than 70% of them have failed to reproduce other scientists’ experiments and more than half have failed to replicate their own experiments. This has led to discussions on how researchers, journals, institutions, and funding agencies can increase transparency and reproducibility. In the FINEPRINT project we implement good practice guidelines and tools to ensure the reproducibility of our analyses and publications. We want to share our work with the world openly, as we believe that transparency is the best way to build trust in scientific findings. Furthermore, we hope to speed up scientific discoveries by allowing everybody to contribute, improve methods and answer new research questions. Ensuring the reproducibility of our research findings means keeping track of a variety of data, scripts, models and software versions. This includes handling several heterogeneous datasets derived from spatial and non-spatial data sources (e.g. global crop maps and commodity trade data) as well as versions of our scripts and models (e.g. FABIO). Furthermore, open distribution and proper documentation of research procedures are essential to achieve reproducibility. To improve our workflow, we have selected a set of open source tools and adopted best practice guidelines. In this Brief, we introduce our infrastructure, implemented to support our goals towards reproducibility and open science.

Infrastructure A large part of our research is done before performing actual data analysis or modelling, for example, when investigating data availability or different methodological approaches. This part of the scientific process is vital and worth proper documentation to facilitate the reproduction of findings. We adopted OpenProject as a web-based tool for task and project management and built a web service for research documentation using R Markdown. These tools facilitate collaboration and transparency across the team and always provide the most recent version of work and a complete version history. This setup ensures that our team always stays up-to-date regarding progress and future working steps (see Figure 1). We adopted Git and GitHub to facilitate collaborative code development, adequate version control as well as a platform to share our work with the scientific community. The FINEPRINT repositories are available online at github.com/fineprint-global. As we implement most of our work in R, we introduced best practice coding guidelines, along with the style guide proposed by [5]. Besides enforcing reproducibility, these guidelines help to improve the readability of the code and scripts - both for other researchers and ourselves The execution of the FINEPRINT project requires handling vast amounts of data, especially in geospatial formats. Therefore, we implemented data management systems that can properly handle data and metadata and provide easy access for processing and visualization. Our system combines Geoserver, for geospatial data management, and CKAN, an open data website tool. This infrastructure allows us to store, search, preview and access vast amounts of spatial and non-spatial data via web interface, web data services (e.g. using OGC Standards and through an application programming interface (API) from within in our scripts. Keeping track of our own development is not sufficient for ensuring reproducibility, due to frequent updates to software dependencies. We use Docker to provide a reproducible computational environment and facilitate the management of applications. Docker is a software platform to build and manage applications in an isolated computational environment - namely in containers, which carry all software dependencies and versions; i.e. akin to virtual machines but usually lighter and faster. The Docker engine provides the foundation for the dynamic and portable setup of several applications currently running on our server. Finally, we are developing interactive data visualizations using Shiny, in order to allow website users to customise graphical analyses and download corresponding data. We have already developed similar visualisation tools on other platforms - for example materialflows.net for visualising global material flow data.