or
or
By clicking below, you agree to our terms of service.
New to HackMD? Sign up
Syntax | Example | Reference | |
---|---|---|---|
# Header | Header | 基本排版 | |
- Unordered List |
|
||
1. Ordered List |
|
||
- [ ] Todo List |
|
||
> Blockquote | Blockquote |
||
**Bold font** | Bold font | ||
*Italics font* | Italics font | ||
~~Strikethrough~~ | |||
19^th^ | 19th | ||
H~2~O | H2O | ||
++Inserted text++ | Inserted text | ||
==Marked text== | Marked text | ||
[link text](https:// "title") | Link | ||
 | Image | ||
`Code` | Code |
在筆記中貼入程式碼 | |
```javascript var i = 0; ``` |
|
||
:smile: | ![]() |
Emoji list | |
{%youtube youtube_id %} | Externals | ||
$L^aT_eX$ | LaTeX | ||
:::info This is a alert area. ::: |
This is a alert area. |
On a scale of 0-10, how likely is it that you would recommend HackMD to your friends, family or business associates?
Please give us some advice and help us improve HackMD.
Do you want to remove this version name and description?
Syncing
xxxxxxxxxx
Repeatability in Practice
WHOAMI
Lab for Data Intensive Biology
Blog
Twitter
The basic idea
Goals of repeatability
In order of importance,
Some diagrams
Tools:
Fitting the tools together:
What many of us have converged on in the DIB Lab
git/GitHub + snakemake + (Jupyter Notebook or R) + conda
With this collection of tools, we can generally achieve full repeatability.
Some points :
The Journey
Overall strategy
For each paper we write, we try to provide a repeatable pipeline on GitHub. Then, for publication, we link the GitHub repo to Zenodo and cut a new version, which gives us a DOI to cite for exactly the version of the code used.
Large raw data files are hosted on specific archival services (e.g. the Sequence Read Archive). Zenodo or figshare or the Open Science Framework serve as nice places to dump intermediate files.
For figure generation, we tend to use Jupyter Notebooks and/or R scripts that are in the github repo. Where possible we put the summary files used to build the figures in the github repo; if they're bigger than 100MB, we put them on OSF; if they're bigger than 5 GB, we start looking at zenodo.
Note: to the best of our knowledge, no one has ever bothered to reproduce our papers from scratch. But we feel good about it anyway :).
The Bad Old Days
For Brown and Callan, 2004, I used a pile of scripts - shell scripts to execute Python scripts to do the heavy lifting. It worked well for the time, but was virtually impossible for anyone else to reproduce unless they were able to compile a specific Python package with C extensions (which at the time was extra hard).
Cool beans
Of note, this approach let me determine that a bug in the code didn't significantly affect the published results (I found the bug much later, when I was reusing the pipeline elsewhere).
The Days of Make
For many years, I switched to using make.
You can see a old-style repo here, 2016-metagenome-assembly-eval (see preprint). The repo contains the LaTeX paper source, some notebooks for graphing, and a
pipeline
directory with a big-arse Makefile in it.I'm actually updating this repo to resubmit in response to reviewers, and we're now using conda to manage software installs.
The brief middle: Docker! Docker! Docker!
When docker came out, I tried it out for a bunch of blog posts - see week of khmer. This ended up being used in a paper, Crossing the streams - see Docker instructions here.
I wasn't happy with this because it turns out that it's virtually impossible to get docker installed on clusters. But it sure did make it easy to run the software across a variety of platforms.
The last year or so
Now we've converged on an approach that uses conda and snakemake; see 2018 spacegraphcats paper - sorry, private link for now.
Other things to mention
The importance of libraries
We tend to encapsulate our reusable functionality in Python libraries that are well tested, versioned, etc. While this has some flaws, it means that the bits of our code that get reused in multiple (often shared) projects in the lab get progressively better tested. It also makes them available to others. The skillset required is pretty advanced though. (Examples: khmer, screed, sourmash).
On the flip side, we do not generally apply any kind of formal testing to our per-project scripts. At best, we try to identify small data sets that we can use to run the whole paper-specific pipeline from start to finish.
Binder
With relatively little effort, you can make your figure-generating Jupyter Notebooks / RMarkdown executable by anyone with a single click through mybinder.org - see my example RStudio repo, and try clicking "launch binder".
Lessons from Oslo, 2016
See this repo for some hands-on lessons in Jupyter Notebook, make, and git.
Workflows-as-applications
In this blog post, I talk about how we're using pydoit and snakemake to write command line applications. This is a pretty nifty way to tie a pile of code together into soemthing that looks like a pipeline but can take advantage of things like dependency graphs, cluster execution and Kubernetes.
Here are some links:
The downsides to this approach are that we now writing papers that have workflows in them that call out to workflows that use wrapped workflows… workflows all the way down!!!
Why snakemake?
Pros:
Cons:
Using argparse in Python
It's a small thing, but I strongly recommend using proper argument parsing - a main() function, with argparse. See top of this script for an example.
It makes setting default parameters and adding arguments really easy, and gives good error messages too.