What do we mean by ‘reproducible’?

Reproducibility is obtaining consistent results using the same input data; computational steps, methods, and code; and conditions of analysis. This definition is synonymous with “computational reproducibility,” and the terms are used interchangeably in this report.

Replicability is obtaining consistent results across studies aimed at answering the same scientific question, each of which has obtained its own data.

What are some bad things that can happen when we fail to focus on reproducibility?

Science is evidence-based by definition, and findings build on one another. Being able to reproduce the findings of other researchers is therefore nearly as important as contributing ‘new’ evidence. Researchers are fallable in their ability to write bug-free code, and by releasing code and data, these errors can be detected and corrected. Apart from this, reproducibiltiy and open-science sort of go hand-in-hand, in that code that is reproducible is also ideally open access, meaning that any researcher has the ability to take the code and data and reproduce the analyses.

What are the major barriers to reproducibility?

How can we improve reproducibility in our code?

There are a number of ways we can improve reproducibility in the code we write.

set.seed(1234); rbinom(10, 1, 0.25)
##  [1] 0 0 0 0 1 0 0 0 0 0
set.seed(1234); rbinom(10, 1, 0.25)
##  [1] 0 0 0 0 1 0 0 0 0 0
set.seed(1234);rbinom(10, 1, 0.5); rbinom(10,1,0.5); rbinom(10, 1, 0.25)
##  [1] 0 1 1 1 1 1 0 0 1 1
##  [1] 1 1 0 1 0 1 0 0 0 0
##  [1] 0 0 0 0 0 1 0 1 1 0
set.seed(123)
rbinom(10, 1, 0.25)
##  [1] 0 1 0 1 1 0 0 1 0 0
rbinom(10, 1, 0.25)
##  [1] 1 0 0 0 0 1 0 0 0 1

Let’s explore an example of this. I wrote a function that performs a “min-max standardization” on a given vector of data. This takes every entry of a vector and subtracts the minimum value, then divides by the maximum minus the minimum. The effect of this is that the data are now scaled between 0 and 1. Proper documentation following the format designed by the roxygen2 package developers and incorporated into the workhorse library devtools, both of which are maintained by Wickham and folks at RStudio/posit. The nice part of this form of documentation is that the documentation and the function are in the same file. Long ago, the documentation of functions in for distribution as an R package would require the developer to edit a separate document with all the details of each function.

#' Min-max standardization
#' 
#' @param x a vector of numeric data
#'
#' @return the same vector x, but standardized between 0 and 1
#' 
#' @examples 
#' getMinMax(x=1:10)

getMinMax <- function(x){
  (x-min(x)) / (max(x)-min(x))
}

Here we include information on what the function does, what arguments it takes, what the output will be, and provide a use case. This is most important for package developers, but it is good practice to provide some documentation of your code.

Let’s take the example of the function we wrote above. It’s well-documented, but let’s say I want to give it a vector that contains an NA value? What is going to happen?

test <- c(1:10, NA)
getMinMax(test)
##  [1] NA NA NA NA NA NA NA NA NA NA NA

That’s not great. Ideally, it would keep the NA values where they are, but still standardize the vector. Maybe even provide the user a warning about NA values being present? Let’s implement that now.

#' Min-max standardization
#' 
#' @param x a vector of numeric data
#'
#' @return the same vector x, but standardized between 0 and 1
#' 
#' @examples 
#' getMinMax(x=1:10)

getMinMax <- function(x){
  if(any(is.na(x))){
    warning('The vector x contains NA values. These will be ignored.')
  }
  (x-min(x, na.rm=TRUE)) / (max(x, na.rm=TRUE)-min(x, na.rm=TRUE))
}

So now, we warn the user about the input data having the NA values and have programmed our function in a way to account for those NA values. What other ways should we think about modifying this function for clarity and usability?

Work through example

Write a function to find and return the mean value of an array. Make this function flexible to:

getMean <- function(x){

}

Given an input of characters, write a function to calculate the sum of the input if each character had a numeric value corresponding to it’s position in the alphabet. Same as above with proper handling of errors/NAs/etc. documentation and proper code styling.

getCharSum <- function(x){

}

A note on linting

linting code refers to searching code for errors and style issues. This is not as comprehensive as unit tests, which test whether functions are working as intended, but is more about code style and catching potential missed commas, variables that are named and never used, etc. There are a number of linting packages in R that will essentially just do the thing for you. The most common at this point is probably styler, which focuses largely (entirely?) on code style. lintr is another one, but is less interactive (it’s more like a diff than just an adjustment to the code you already have written, which is what styler does).

Unit tests

Writing effective tests for your code can help provide sanity checks for you, testing if functions you write are working, exploring edge cases, etc. For instance, if I had code for something like the SIR model we discussed earlier, a good unit test would be to see if the summed classes of S, I, and R stayed consistent through time. That is, individuals are never gained or lost in the overall system in the basic case, so the overall size of the population should never change. Also, setting the initial number of infected individuals to 0 should result in no epidemic. These are base unit tests to provide us with a tiny bit of confidence that our functions are not giving us weird output. These should exist independent of your actual project code used for analysis and reproduction of results, and many people only do proper unit testing when writing R packages.

Let’s work through my first figshare repo from 2015 and see what we can improve

https://tinyurl.com/mkzty4sy

A final note

Put sessionInfo at end of document. Putting sessionInfo at the end the document will clearly list the R version, OS, and package versions that you are using. It’s a couple steps short of full reproducibility, but it at least shows the user (if you compile the code to html or pdf) exactly what set of conditions allowed the code to run all the way through.

A better approach would be to use a proper package manager, which R doesn’t really have, or to set up environments using other software within R. This would be similar to setting up a conda environment in python, and can be done with things like packrat (potentially deprecated) and renv (definitely not deprecated). This essentially stores a snapshot (like git!) of a workspace and all loaded packages, their versions, etc., which allows any user to hypothetically use the renv lockfile to recreate the exact same environment you used to run your code. Definitely worth exploring if you are writing code that you want to be reproducible over long time periods. The only downsides to this are that it still adds some OS-level dependencies that may not be fully captured.

The best solution to this issue is to use a containerized environment where your analysis runs. This could be done with something like Docker, which is software that essentially builds a very small virtual machine on your computer with the sole goal of running your analyses. Docker containers can be built from dockerfiles, which are instructions on how to construct the machine. There is a decent community of researchers working specifically with Docker containers in R (see https://rocker-project.org), so check out their image library (https://rocker-project.org/images/) if you are interested in learning more about this approach.

sessionInfo

sessionInfo()
## R version 4.5.1 (2025-06-13)
## Platform: x86_64-pc-linux-gnu
## Running under: Ubuntu 22.04.5 LTS
## 
## Matrix products: default
## BLAS:   /usr/lib/x86_64-linux-gnu/atlas/libblas.so.3.10.3 
## LAPACK: /usr/lib/x86_64-linux-gnu/atlas/liblapack.so.3.10.3;  LAPACK version 3.10.0
## 
## locale:
##  [1] LC_CTYPE=en_US.UTF-8       LC_NUMERIC=C              
##  [3] LC_TIME=en_US.UTF-8        LC_COLLATE=en_US.UTF-8    
##  [5] LC_MONETARY=en_US.UTF-8    LC_MESSAGES=en_US.UTF-8   
##  [7] LC_PAPER=en_US.UTF-8       LC_NAME=C                 
##  [9] LC_ADDRESS=C               LC_TELEPHONE=C            
## [11] LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C       
## 
## time zone: America/New_York
## tzcode source: system (glibc)
## 
## attached base packages:
## [1] stats     graphics  grDevices utils     datasets  methods   base     
## 
## loaded via a namespace (and not attached):
##  [1] digest_0.6.37     R6_2.6.1          fastmap_1.2.0     xfun_0.53        
##  [5] cachem_1.1.0      knitr_1.50        htmltools_0.5.8.1 rmarkdown_2.29   
##  [9] lifecycle_1.0.4   tinytex_0.57      cli_3.6.5         sass_0.4.10      
## [13] jquerylib_0.1.4   compiler_4.5.1    tools_4.5.1       evaluate_1.0.5   
## [17] bslib_0.9.0       yaml_2.3.10       rlang_1.1.6       jsonlite_2.0.0