OpenPBTA is now published:
Shapiro, J.A., Gaonkar, K.S., Spielman, S.J., Savonen, C.L., Bethell, C.J., Jin, R., Rathi, K.S., Zhu, Y., Egolf, L.E., Farrow, B.K., et al. (2023). OpenPBTA: The Open Pediatric Brain Tumor Atlas. Cell Genom., 100340. 10.1016/j.xgen.2023.100340.
We will no longer be accepting contributions to this repository.
Please check out d3b-center/OpenPedCan-analysis
for updates to PBTA data as well as other data sources.
Pediatric brain tumors are the most common solid tumors and the leading cause of cancer-related death in children. Our ability to understand and successfully treat these diseases is hindered by small sample sizes due to the overall rarity of unique molecular subtypes and tainted grouped analyses resulting from misclassification. In September of 2018, the Children's Brain Tumor Network (CBTN) released the Pediatric Brain Tumor Atlas (PBTA), a genomic dataset (whole genome sequencing, whole exome sequencing, RNA sequencing, proteomic, and clinical data) for nearly 1,000 tumors, available from the Gabriella Miller Kids First Portal.
The Open Pediatric Brain Tumor Atlas (OpenPBTA) Project is a global open science initiative to comprehensively define the molecular landscape of tumors of 943 patients from the CBTN and the PNOC003 DIPG clinical trial from the Pediatric Pacific Neuro-oncology Consortium through real-time, collaborative analyses and collaborative manuscript writing on GitHub.
The OpenPBTA operates on a pull request model to accept contributions from community participants. The maintainers have set up continuous integration software to confirm the reproducibility of analyses within the project’s Docker container. The collaborative manuscript is authored using Manubot software to provide an up-to-date public version of the manuscript. The project maintainers include scientists from Alex's Lemonade Stand Foundation's Childhood Cancer Data Lab and the Center for Data-Driven Discovery in Biomedicine at the Children's Hospital of Philadelphia. We invite researchers to join OpenPBTA to help rigorously characterize the genomic landscape of these diseases to enable more rapid discovery of additional mechanisms contributing to the pathogenesis of pediatric brain and spinal cord tumors and overall accelerate clinical translation on behalf of patients.
New to the project? Please be sure to read the following documentation before contributing:
- Learn about the fundamental data used for this project in
doc/data-formats.md
anddoc/data-files-description.md
- A history of data releases can be found in
doc/release-notes.md
- A history of data releases can be found in
- See what analyses are being performed in
analyses/README.md
- Read the remainder of this README document in full.
- Read our contributing guidelines in
CONTRIBUTING.md
in full.
Table of Contents generated with DocToc
- Data Description
- How to Obtain OpenPBTA Data
- How to Participate
- How to Add an Analysis
- Data release preparation
The OpenPBTA dataset includes gene expression, fusion, as well as somatic mutation, copy number, structural and variant results in combined tsv or matrix format.
Below is a summary of biospecimens by sequencing strategy:
Experimental Strategy | Normal | Tumor |
---|---|---|
Targeted DNA Panel | 1 | 1 |
RNA-Seq | 0 | 1036 |
WGS | 801 | 940 |
WXS | 31 | 31 |
All sequencing was performed on nucleic acids extracted from fresh-frozen tissues using paired-end strategies. The manuscript methods section has additional details.
Below is a detailed table of broad histologies for the 1028 RNA-Seq biospecimens:
Broad Histology | N |
---|---|
Benign tumor | 34 |
Choroid plexus tumor | 11 |
Diffuse astrocytic and oligodendroglial tumor | 191 |
Embryonal tumor | 184 |
Ependymal tumor | 93 |
Germ cell tumor | 13 |
Histiocytic tumor | 6 |
Low-grade astrocytic tumor | 303 |
Lymphoma | 1 |
Melanocytic tumor | 1 |
Meningioma | 29 |
Mesenchymal non-meningothelial tumor | 25 |
Metastatic secondary tumors | 5 |
Neuronal and mixed neuronal-glial tumor | 34 |
Non-CNS tumor | 1 |
Non-tumor | 3 |
Other astrocytic tumor | 3 |
Other tumor | 1 |
Pre-cancerous lesion | 14 |
Tumor of cranial and paraspinal nerves | 44 |
Tumor of pineal region | 5 |
Tumors of sellar region | 35 |
Below is a table of number of tumor biospecimens by phase of therapy (DNA and RNA):
Phase of Therapy | N |
---|---|
Initial CNS Tumor | 1520 |
Progressive | 302 |
Progressive Disease Post-Mortem | 13 |
Recurrence | 136 |
Second Malignancy | 35 |
Unavailable | 2 |
We are releasing this dataset on both CAVATICA and AWS S3.
Users performing analyses, should always refer to the symlinks in the data/
directory and not files within the release folder, as an updated release may be produced before a publication is prepared.
The data formats and caveats are described in more detail in doc/data-formats.md
.
For brief descriptions of the data files, see the data-files-description.md
file included in the download.
Use the data issue template to file issues if you have questions about or identify issues with OpenPBTA data.
We have created a shell script that will download the latest release from AWS S3.
macOS users must install md5sum
before running the download script the first time.
This can be installed with homebrew via the command brew install coreutils
or conda/miniconda via the command conda install -c conda-forge coreutils
.
Note: the download-data.sh
script now has the ability to skip downloads of unchanged files, but if you previously installed md5sum via brew you'll need to run brew unlink md5sha1sum && brew install coreutils
first to take advantage of this new feature.
Once this has been done, run bash download-data.sh
to acquire the latest release.
This will create symlinks in data/
to the latest files.
It's safe to re-run bash download-data.sh
to check that you have the most recent release of the data.
We will update the default release number whenever we produce a new release.
For any user registered on CAVATICA, the OpenPBTA data can be accessed from the CAVATICA public project below:
The release folder structure in CAVATICA mirrors that on AWS.
Users downloading via CAVATICA should place the data files within the data/release*
folder and then create symlinks to those files within /data
.
Have general questions or need help getting started using GitHub? You can join the Cancer Data Science Slack to connect with OpenPBTA organizers, other project participants, and the broader cancer data science community. Sign up and join the #open-pbta channel to get started!
There are certain analyses that we have planned or that others have proposed, but which nobody is currently in charge of completing. Check the existing issues to identify these. If you would like to take on a planned analysis, please comment on the issue noting your interest in tackling the issue in question. Ask clarifying questions to understand the current scope and goals. Then propose a potential solution. If the solution aligns with the goals, we will ask you to go ahead and start to implement the solution. You should provide updates to your progress in the issue. When you file a pull request with your solution, you should note that it closes the issue in question.
In addition to the planned analyses, we welcome contributors who wish to propose their own analyses of this dataset as part of the OpenPBTA project. Check the existing issues before proposing an analysis to see if something similar is already planned. If there is not a similar planned analysis, create a new issue. The ideal issue will describe the scientific goals of the analysis, the planned methods to address the scientific goals, the input data that is required for the planned methods, and a proposed timeline for the analysis. Project maintainers will interact on the issue to clarify any questions or raise any potential concerns.
This section describes the general workflow for implementing analytical code, and more details are described below. The first step is to identify an existing analysis or propose a new analysis, engage with the project maintainers to clarify the goals of the analysis, and then get the go ahead to move forward with the analysis.
You can perform your analyses via a script (R or Python) or via a notebook (R Markdown or Jupyter). Your analyses should produce one or more artifacts. Artifacts include both vector or high-resolution figures sufficient for inclusion in a manuscript as well as new summarizations of the data (tables, etc) that are intended for either use in subsequent analyses or distribution with the manuscript.
Analyses should be performed within the project's Docker container. We use a single monolithic container in these analyses for ease of use. If you need software that is not included, please edit the Dockerfile to install the relevant software or file a new issue on this repository requesting assistance.
Analyses are added to this repository via Pull Requests. Please read the Pull Request section of the contribution guidelines carefully. We are using continuous integration software applied to the supplied test datasets to confirm that the analysis can be carried out successfully within the Docker container.
Users performing analyses, should always refer to the symlinks in the data/
directory and not files within the release folder, as an updated release may be produced before a publication is prepared.
Our folder structure is designed to separate each analysis into its own set of notebooks that are independent of other analyses.
Within the analyses
directory, create a folder for your analysis.
Choose a name that is unique from other analyses and somewhat detailed.
For example, instead of gene-expression
, choose gene-expression-clustering
if you are clustering samples by their gene expression values.
You should assume that any data files are in the ../../data
directory and that their file names match what the download-data.sh
script produces.
These files should be read in at their relative path, so that we can re-run analyses if the underlying data change.
Files that are primarily graphic should be placed in a plots
subdirectory and should adhere to the color palette guide.
Files that are primarily tabular results files should be placed in a results
subdirectory.
Intermediate files that are useful within the processing steps but that do not represent final results should be placed in ../../scratch/
.
It is safe to assume that files placed in ../../scratch
will be available to all analyses within the same folder.
It is not safe to assume that files placed in ../../scratch
will be available from analyses in a different folder.
An example highlighting a new-analysis
directory is shown below.
The directory is placed alongside existing analyses within the analyses
directory.
In this case, the author of the analysis has run their workflows in R Markdown notebooks.
This is denoted with the .Rmd
suffix.
However, the author could have used Jupyter notebooks, R scripts, or another scriptable solution.
The author has produced their output figures as .pdf
files.
We have a preference for vector graphics as PDF files, though other forms of vector graphics are also appropriate.
The results folder contains a tabular summary as a comma separated values file.
We expect that the file suffix (.csv
, .tsv
) accurately denotes the format of the added files.
The author has also included a README.md
(see Documenting Your Analysis).
OpenPBTA-analysis
├── CONTRIBUTING.md
├── README.md
├── analyses
│  ├── existing-analysis-1
│  └── new-analysis
│  ├── 01-preprocess-data.Rmd
│  ├── 02-run-analyses.Rmd
│  ├── 03-make-figures.Rmd
│  ├── README.md
│  ├── plots
│  │  ├── figure1.pdf
│  │  └── figure2.pdf
│  ├── results
│  │ └── tabular_summary.csv
│  └── run-new-analysis.sh
├── data
└── scratch
A goal of the OpenPBTA project is to create a collection of workflows that are commonly used for atlas papers. As such, documenting your analytical code via comments and including information summarizing the purpose of your analysis is important.
When you file the first pull request creating a new analysis module, add your module to the Modules At A Glance table. This table contains fields for the directory name, what input files are required, a short description, and any files that you expect other analyses will rely on. As your analysis develops and input or output files change, please check this table remains up to date. This step is included in the pull request reproducibility checklist.
When an analysis module contains multiple steps or is nearing completion, add a README.md
file that summarizes the purpose of the module, any known limitations or required updates, and includes examples for how to run the analyses to the folder.
As shown above, analysis scripts within a folder should be numbered from 01
and are intended be run in order.
If the script produces any intermediate files, these files should be placed in ../../scratch
, which is used as described above.
A shell script that runs all analytical code in the intended order should be added to the analysis directory (e.g. run-new-analysis.sh
above).
See the continuous integration instructions for adding analyses with multiple steps for more information.
The CI system that we use will generate, as artifacts, the contents of the analyses
directory applied over a small test dataset.
Our goal is to capture all of the outputs that will be used for the OpenPBTA-manuscript as artifacts.
Files that are primarily graphic should be placed in a plots
subdirectory of the analysis's folder.
Plots should use the specified color palettes for this project.
See more specific instructions on how to use the color palette here.
Files that are primarily tabular results files should be placed in a results
subdirectory of the analysis's folder.
Files that are intermediate, which means that they are useful within an analysis but do not provide outputs intended for tables, figures, or supplementary tables or figures of the OpenPBTA-manuscript, should be placed in ../../scratch
.
We build our project Docker image from a versioned tidyverse
image from the Rocker Project (v3.6.0).
To add dependencies that are required for your analysis to the project Docker image, you must alter the project Dockerfile
.
- R packages installed on this image will be installed from an MRAN snapshot corresponding to the last day that R 3.6.0 was the most recent release (ref).
- Installing most packages, from CRAN or Bioconductor, should be done with our
install_bioc.R
script, which will ensure that the proper MRAN snapshot is used.BiocManager::install()
should not be used, as it will not install from MRAN. - R packages that are not available in the MRAN snapshot can be installed via GitHub using our
install_github.R
script, with the commit specified by the--ref
argument.- To avoid rate limits by GitHub when installing these packages, we use an access token which is passed to the build environment via Docker secrets.
To use this token, your installation step should start with
RUN --mount=type=secret,id=gh_pat
and you should pass the argument--pat_file /run/secrets/gh_pat
to theinstall_github.R
script
- To avoid rate limits by GitHub when installing these packages, we use an access token which is passed to the build environment via Docker secrets.
To use this token, your installation step should start with
- Installing most packages, from CRAN or Bioconductor, should be done with our
- Python packages should be installed with
pip3 install
with version numbers for all packages and dependencies specified.- As a secondary check, we maintain a
requirements.txt
file to check versions of all python packages and dependencies. - When adding a new package, make sure that all dependencies are also added; every package should appear with a specified version both in the
Dockerfile
andrequirements.txt
.
- As a secondary check, we maintain a
- Other software can be installed with
apt-get
, but this should never be used for R packages.
If you need assistance adding a dependency to the Dockerfile, file a new issue on this repository to request help.
The most recent version of the project Docker image, which is pushed to Docker Hub after a pull request gets merged into the master branch, can be obtained via the command line with:
docker pull ccdlopenpbta/open-pbta:latest
If you are a Mac or Windows user, the default limit for memory available to Docker is 2 GB. You will likely need to increase this limit for local development. [Mac documentation, Windows documentation]
Using rocker/tidyverse:3.6.0
as our base image allows for development via RStudio in the project Docker container.
If you'd like to develop in this manner, you may do so by running the following and changing <password>
to a password of you choosing at the command line:
docker run -e PASSWORD=<password> -p 8787:8787 ccdlopenpbta/open-pbta:latest
You can change the volume that the Docker container points to either via the Kitematic GUI or the --volume
flag to docker run
.
Once you've set the volume, you can navigate to localhost:8787
in your browser if you are a Linux or Mac OS X user.
The username will for login will be rstudio
and the password will be whatever password you set with the docker run
command above.
If you are a new user, you may find these instructions for a setting up a different Docker container or this guide from Andrew Heiss helpful.
While we encourage development within the Docker container, it is also possible to conduct analysis without Docker if that is desired. In this case, it is important to ensure that local or personal settings such as file paths or installed packages and libraries are not assumed in the analysis.
We have supplied an RStudio project (OpenPBTA-analysis.Rproj
) file at the root of the project to aid in organization and encourage reproducible defaults for analysis.
In particular, we do not source .Rprofile
files in new sessions or save/restore workspaces.
We use continuous integration (CI) to ensure that the project Docker image will build if there are any changes introduced to the Dockerfile
and that all analysis code will execute.
We have put together data files specifically for the purpose of CI that contain all of the features of the full data files for only a small subset of samples. You can see how this was done by viewing this module. We use the subset files to cut down on the computational resources and time required for testing.
Provided that your analytical code points to the symlinks in the data/
directory per the instructions above, adding the analysis to the CI (see below) will run your analysis on this subset of the data.
Do not hardcode sample names in your analytical code: there is no guarantee that those samples will be present in the subset files.
If you would like to work with the files used in CI locally, e.g., for debugging, you can obtain them from AWS by running the following in the root directory of the project:
bash scripts/download-ci-files.sh
Running this will change the symlinks in data
to point to the files in data/testing
.
For an analysis to be run in CI, it must be added to the Circle CI configuration file, .circleci/config.yml
.
A new analysis should be added as the last step of the run_analyses
section.
Here is an example analysis that simply lists the contents of the data directory that contains the files for the test:
- run:
name: List Data Directory Contents
command: ./scripts/run_in_ci.sh ls data/testing
Using ./scripts/run_in_ci.sh
allows you to run your analyses in the project Docker container.
If you wanted to add running an Rscript called cluster-samples.R
that was in an analysis folder called gene-expression-clustering
, you would add this script to continuous integration with:
- run:
name: Cluster Samples
command: ./scripts/run_in_ci.sh Rscript analyses/gene-expression-clustering/cluster-samples.R
This would run the cluster-samples.R
on the subset files that are specifically designed to be used for CI.
There is a different procedure for adding an analysis comprised of multiple scripts or notebooks to CI.
Per the contribution guidelines, each script or notebook should be added via a separate pull request.
For each of these pull requests, the individual script or notebook should be added as its own run in the .circleci/config.yml
file.
This validates that the code being added can be executed at the time of review.
Once all code for an analysis has been reviewed and merged, a final pull request for the analysis that is comprised of the following changes should be filed:
- A shell script that will run all script and/or notebooks in the analysis module.
- The multiple runs from the module that are in the
config.yml
file are replaced with a single run that runs the shell script.
If the gene-expression-clustering
analysis above instead required two scripts run sequentially (01-filter-samples.R
and 02-cluster-heatmap.R
), we would follow the procedure below.
In this pull request, we would add the following change to .circleci/config.yml
.
- run:
name: Filter Samples
command: ./scripts/run_in_ci.sh Rscript analyses/gene-expression-clustering/01-filter-samples.R
In this pull request, we would add the following change to .circleci/config.yml
.
This would be added below the Filter Samples
run.
- run:
name: Cluster Samples and Plot Heatmap
command: ./scripts/run_in_ci.sh Rscript analyses/gene-expression-clustering/02-cluster-heatmap.R
3. File and merge a pull request for the shell script that runs the entirety of gene-expression-clustering
.
In this pull request, we would add a shell script that runs 01-filter-samples.R
and 02-cluster-heatmap.R
.
Let's call this shell script run-gene-expression-clustering.sh
and place it in the analysis directory analyses/gene-expression-clustering
.
The contents of analyses/gene-expression-clustering/run-gene-expression-clustering.sh
may look like:
#!/bin/bash
# This script runs the gene-expression-clustering analysis
# Author's Name 2019
set -e
set -o pipefail
Rscript --vanilla analyses/gene-expression-clustering/01-filter-samples.R
Rscript --vanilla analyses/gene-expression-clustering/02-cluster-heatmap.R
We would remove the runs Filter Samples
and Cluster Samples and Plot Heatmap
from .circleci/config.yml
and instead replace them with a single run:
- run:
name: Cluster Samples and Plot Heatmap
command: ./scripts/run_in_ci.sh bash analyses/gene-expression-clustering/run-gene-expression-clustering.sh
The analyses run in CI use only a small portion of the data so that tests can be run quickly. For some analyses, there will not be enough samples to fully test code without altering certain parameters passed to methods. The preferred way to handle these is to run these analyses through a shell script that specifies default parameters using environment variables. The default parameters should be the ones that are most appropriate for the full set of data. In CI, these will be replaced.
We might decide that it makes the most sense to run an analysis using a more permissive statistical significance threshold in CI so that some "significant" pathways still appear and subsequent code that examines them can be tested.
We'd first write code capable of taking command line parameters.
In R, we could use optparse
to specify these in a script - imagine it's called pathway_sig.R
and it contains an option list:
option_list <- list(
optparse::make_option(
c("-a", "--alpha"),
type = "double",
help = "pathway significance threshold",
)
)
Then we would create a shell script (perhaps run_pathway_sig.sh
) that uses a default environment variable. If OPENPBTA_PATHSIG
is defined, it will be used. Otherwise, a value of 0.05 is used.
Note: the -
before the 0.05
below is necessary notation for a default parameter and not designating a negative 0.05.
PATHSIG=${OPENPBTA_PATHSIG:-0.05}
Rscript analyses/my-path/pathway_sig.R --alpha $PATHSIG
We can override this by passing environment variables in .circleci/config.yml
.
For testing, we might want to use an alpha level of 0.75 so that at least some "significant" pathways appear, which allows testing subsequent code that depends on them.
The run command in the .circleci/config.yml
is used to specify these parameters.
- run:
name: run pathway significance tests
command: OPENPBTA_PATHSIG=0.75 ./scripts/run_in_ci.sh bash analyses/my-path/run_pathway_sig.sh
In this example OPENPBTA_PATHSIG=0.75
species an environment variable OPENPBTA_PATHSIG
that is set to 0.75.
Any environment variables prefixed with OPENPBTA_
are passed to the specified shell script.
Environment variables without this prefix are not passed.
Some scripts in this analysis repository are required for preparing a data release. To learn more, please see these docs.