This repository contains the Python package stereoscope, which is the implementation of the method presented in the paper. In addition, scripts used to: preprocess, visualize, and compare results presented in the paper are also included.
Below examples illustrating how to use Stereoscope can be found. While some of these examples are cast as a guides to reproduce the results presented in the paper, they can easily be generalized and applied to any data set. The three examples are:
- Reproducing the mouse brain analysis: Conducting the complete analysis of the mouse brain presented in the paper, from downloading data to visualization of the results
- Reproducing the method comparison: Generation of synthetic data, running all three methods and comparing them
- Using pre-estimated parameters : Examining four additional breast cancer samples, applying already estimated single cell parameters to set of ST data.
Note: While the ST data will be publicly available upon publication, these files are currently password protected - access will be given to reviewers.
In order to install stereoscope
, we begin by cloning this very repository.
Open your terminal and go to your desired download directory - then enter
the following:
foo@bar:~$ git clone https://github.com/almaan/stereoscope
Now to actually install Stereoscope, just use the provided setup.py
file. If you are working
with conda or virtual environments, this is where you should create and source a
new environment.
the following commands
foo@bar:~$ cd stereoscope
foo@bar:~$ ./setup.py install --user
Assuming the installation was successful, you should now have access to both the stereoscope python package (stsc) and the stereoscope
command line interface
(CLI) meaning you can conduct the analysis from the terminal. To confirm that the installation was successful, we
will run two tests, which are:
foo@bar:~$ python -c "import stsc; print(stsc.__version__)"
stereoscope : 0.2.0
foo@bar:~$ stereoscope test
successfully installed stereoscope CLI
If you cannot access stereoscope from the command line, and receive a message stating something similar to:
foo@bar:~$ stereoscope test
bash stereoscope: command not found..
This means that your install location is not a part of the PATH variable.
Including the --user
flag during installation should place your packages
in your ~/.local
directory. Thus, this issue may be resolved by entering the
following into your terminal:
foo@bar:~$ export PATH=$PATH:/user/home/.local/bin
Having installed stereoscope
, we are now ready to start with the analysis.
Let us begin by reproducing the results presented for the mouse brain (Hippocampal region), to briefly remind you what
that means; we will analyze two (1k) ST-sections of the mouse brain (jointly) and map single cell data from the resource
mousebrain.org onto it. Here we will go through the whole workflow including downloading the
data, preprocessing, analysis and visualization. If you aren't that interested in the first parts and want to get
started with the analysis, you can skip step 1-2 and use the already processed data found in the
data/mousebrain
folder. Simply unzip these files into a folder named data/curated
.
As mentioned above, the data we used in the paper is downloaded from mousebrain.org. More specifically, we
used the loom
-file containing cells from the Hippocampus region. If you do not have loompy
installed, we refer you to
this link for further instructions. You can either download the data set via the web browser, or the
terminal using :
foo@bar:~$ cd data
foo@bar:~$ mkdir raw curated
foo@bar:~$ cd raw
foo@bar:~$ curl -O https://storage.googleapis.com/linnarsson-lab-loom/l1_hippocampus.loom
We will subsample the single cell data, this is not a requirement, but it allows us to run the analysis a bit faster.
First we create a modified loom
-file where a new set of labels is introduced, formed by joining the Cluster and Class labels
together, this adds (some) context to our labels compared to just using cluster indices. Enter the following in
the terminal:
foo@bar:~$ ../../preprocess/hippocampus/create-mod-loom.py l1_hippocampus.loom .
successfully created modifed loom-file >> mod_l1_hippocampus.loom
Next, we subsample our data set. We use a lower and upper bound (See Methods) of 25 repsectively 250 cells per cell type. Do this by entering:
foo@bar:~$ ../../preprocess/hippocampus/subsample-data.py -lf mod_l1_hippocampus.loom -o ../curated -lb 25 -ub 250 -cn
"Celltype_1"
Unique Identifier for set >> 20191029082957812352
Astrocytes_13 | was discarded due to insufficient number of cells
Astrocytes_14 | Used 250 cells
Astrocytes_38 | was discarded due to insufficient number of cells
Astrocytes_39 | was discarded due to insufficient number of cells
Astrocytes_40 | Used 250 cells
Astrocytes_41 | Used 31 cells
Astrocytes_42 | Used 250 cells
Astrocytes_44 | Used 41 cells
....
This will create three files in the data/curated folder - a count matrix of the cells included in the set, a meta-file containing their respective labels and a ''stats-file'' which displays the composition of the set. All files belonging to the same set are marked with a unique identifier which is time and date-based (here : 20191029082957812352), your identifier will therefore be different from the one above. As you may notice, not all ''cell types'' are listed; since we specified a lower bound, all types with fewer than 25 members (cells) are excluded.
We have included the two mouse brain sections presented in the paper as .tsv
files in the repo. You do not need to
preprocess these, but we have zipped them to reduce the file sizes. Unzip these files and place them in the data/curated
folder, you could do this
either interactively or for example by entering the following command into the terminal :
foo@bar:~$ 7z e mouse-st-data.zip
7-Zip [64] 16.02 : Copyright (c) 1999-2016 Igor Pavlov : 2016-05-21
Scanning the drive for archives:
1 file, 7254353 bytes (7085 KiB)
Extracting archive: mouse-st-data.zip
--
Path = mouse-st-data.zip
Type = zip
Physical Size = 7254353
Enter password (will not be echoed):
....
foo@bar:~$ mv st-hippo*tsv curated/
If all steps were successfull, this (below) should be representative of the content in your data/curated
folder
foo@bar:~$ ls -1 curated/*
20191029082957812352.cnt_data.tsv
20191029082957812352.mta_data.tsv
20191029082957812352.stats.tsv
st-hippo1.tsv
st-hippo2.tsv
The output from the subsampling and the ST-data we provided you with are given in the correct formats,
hence you will not have to do any additional work. However, we will briefly describe what type of files are required and
how they should be formatted in order to run stereoscope
:
- Single Cell Count Data File - a
.tsv
file with cells as rows and genes as columns, each cell (row) should have a unique label - Single Cell Annotation Data - a
.tsv
file with the same rownames as the count data file, either with one single column listing the annotations, or multiple columns where the column containing the labels should be named 'bio_celltype' - Spatial Transcriptomics (ST) Count Data - a tsv file containing with spots as rows and genes as columns
Note: We are planning to add support for h5ad
files, but are awaiting a somewhat more standardized format for spatial data
Some additional things to keep in mind are:
- Make sure that your ST and single cell data use the same gene identifiers. For example, having one set using ENSEMBL ids while
the other one has HGNC gene symbols will raise an error. However as long as the ids match,
stereoscope
is agnostic to the type of identifiers used. - Do not normalize your data - the model relies on using raw count data, your gene counts should thus always be integer numbers.
Now when the data has been prepared, we are ready to use steroscope
. We will run the complete analysis, estimating rates and
logits from the single cell data and then use these to infer the proportion values in our spatial data. We use the following
specs:
parameter | values |
---|---|
number of genes | 5000 |
sc epochs | 75000 |
sc batch size | 100 |
st epochs | 75000 |
st batch size | 100 |
learning rate | 0.01 |
gpu | True |
To run the analysis enter the following into your terminal :
foo@bar:~$ cd ../../res
foo@bar:~$ stereoscope run --sc_cnt ../data/curated/*cnt*.tsv --sc_labels ../data/curated/*mta*.tsv -sce 75000 -o hippo_1 -n 5000 --st_cnt ../data/curated/st-hippo*tsv -ste 75000 --gpu -stb 100 -scb 100
[2019-10-29 09:07:50,891 - stsc - INFO ] >> Using device cuda
[2019-10-29 09:07:50,891 - stsc - INFO ] >> fitting sc data | count file : ../data/curated/20191029082957812352.cnt_data.tsv | labels file : ../data/curated/20191029082957812352.mta_data.tsv
[2019-10-29 09:09:51,527 - stsc - INFO ] >> SC data GENES : 5000 SC data CELLS : 8449 SC data TYPES : 56
Epoch : 211 /75000 | Loss : 3.004263E+07 | [ ]
This will create a subfolder named hippo_1
in the res
folder, where all results, logs and data related to this analysis
will eventually be located.
For more information regarding which arguments and configurations you can make to your analysis use:
foo@bar:~$ stereoscope run -h
Even with GPU resources the analysis will take some time to complete. While the progress bar shows the current
status - it's also of interest to put this into context and track the progress
over time. We can do so by using the progress
module accordingly:
foo@bar:~$ stereoscope progress -lf hippo_1/sc_loss*txt & disown
This will open up a interactive window, where you can zoom and move around. The plot will be updated every 10
seconds, allowing you to see how the loss changes over time. Appending & disown
to the command, allows us to keep
using the same terminal window. Below is an example of what the progress-plot may look like.
Upon completeion of the analysis. The folder res/hippo_1
folder should contain the following set of files:
foo@bar:~$ ls -1
logits.2019-10-29090750.880065.tsv
R.2019-10-29090750.880065.tsv
sc_loss.2019-10-29090750.880065.txt
sc_model.2019-10-29090750.880065.pt
st-hippo1
st-hippo2
st_loss.2019-10-29090750.880065.txt
st_model.2019-10-29090750.880065.pt
stsc.2019-10-29090750.880065.log
With the subfolder content being:
foo@bar:~$ ls st-hippo*/
st-hippo1/:
W.2019-10-29090750.880065.tsv
st-hippo2/:
W.2019-10-29090750.880065.tsv
The W-files contain the proportion estimates, and thus what we are mainly interested in. Every section has its own ouput folder,
while the W-files share the same name. These files are given in a matrix format n_spots x n_types
where each
element represents the proportion of a cell type within a specific spot. To further illustrate :
foo@bar:~$ head st-hippo1/W*tsv -n 5 | cut -f 1-5 | column -t
Astrocytes_14 Astrocytes_40 Astrocytes_41 Astrocytes_42
4.83x31.08 0.017654551 0.019365933 0.021137744 0.019530216
19.98x24.93 0.020201972 0.017161706 0.016906356 0.017648073
15.87x9.01 0.017025681 0.017077405 0.016912553 0.016671246
5.83x27.97 0.019220736 0.018679986 0.017026993 0.017045338
We could justify the usage of a "threshold", for example setting all proportions lower than a certain value to zero. We will not to adjust our results and keep all values as they are.
4.2 Visualization
We include a tool (the look
module) for seamless visualization of the results generated by stereoscope
. To use look
do:
foo@bar:~$ stereoscope look -pp st-hippo*/W*tsv -o viz -sc i -sb s -nc 7 -c "umap" -g -ms 40
This will generate two types of images, saved to the folder viz
-
Separate visualizations : spots are plotted according to their array coordinates, and the intensity of their blue facecolor corresponds to the proportion value of each celltype. All of these plots are scaled internally (this highlights the spatial patterns, but is not suitable for comparisons), changing the argument
-sc i
to-sc s
will scale all values within each section. -
Joint visualizations : These are the type of "compressed" images described in the paper's Method. This type of visualization is found in the Supplementary. Regions of similar colors have a similar cell type compositions. The method used for the dimensionality reduction is
umap
, but you can also choose betweenpca
(slightly faster) andtsne
(much slower).
Below are some examples of the generated images:
You can customize the output of stereoscope look
in multiple ways by providing certain arguments, for more
information run:
foo@bar:~$ stereoscope look -h
One could also overlay the spots on the tissue image, to see how the spatial
patterns relates to the morphology, like we did in our figures. This type of
visualization is not a part of the stereoscope
package, but we do provide
scripts for this.
The material that you need for this would be:
- HE-image - image taken of the tissue, can be scaled but not cropped
- Transformation Matrix - matrix that maps spot coordinates to pixel coordinates
- Proportion Estimates - output from
stereoscope
analysis - Mask (optional)- mask to indicate which parts of the area to be included (transparent) resepectively excluded (black)
We actually use resized images (30% downscaled), since the original images are unnecessarily large.
Still being in the res
folder, all you have to do is run:
foo@bar:~$ ./map2he.py -i ../data/mouse/rsc/st-hippo1.jpg -t ..data/mouse/rsc/st-hippo1-tmat.txt -p st-hippo1/W*tsv -sf 0.3 -si -o he_overlay
using image file ../data/mouse/rsc/st-hippo1.jpg
using proportions file st-hippo1/W.2019-10-29090750.880065.tsv
Rendering type Astrocytes_14
Rendering type Astrocytes_40
....
Resulting in images like these (we rotated these images in our paper):
In the paper we compare stereoscope
with two other methods DWLS and
deconvSeq using synthetic data. This comparison is something we will reproduce
in this section.
Begin by unzipping the file data/comp/comp-data.zip
into the data/comp/
folder
foo@bar:~$ cd data/comp
foo@bar:~$ unzip comp-data.zip
foo@bar:~$ ls -1
comp-data.zip
real:
hippo-real-sc-cnt.tsv
hippo-real-sc-mta.tsv
synthetic:
counts.st-hippo-comp.tsv
generation.20190908194059502947.cnt_data.tsv
generation.20190908194059502947.mta_data.tsv
members.st-hippo-comp.tsv
proportions.st-hippo-comp.tsv
validation.20190908194059502947.cnt_data.tsv
validation.20190908194059502947.mta_data.tsv
As you can see, there is already a set of synthetic data available; you can use this data if you'd like to, if so, then move ahead to step 2. We will however go through the full procedure for generation of these files.
Begin by removing the prepared data, to make sure we don't mix thing up :
foo@bar:~$ rm synthetic/*
We first split our single cell data into a "generation" and ''validation'' set, preferably of equal size. The single cell data is the same Hippocampus set we used in the previous example. The only difference is that we subsampled w.r.t. to Subclass labels rather than Cluster labels.
foo@bar:~$ ../../comparison/synthetic_data_generation/make_sc_sets.py real/hippo-real-sc-cnt.tsv real/hippo-real-sc-mta.tsv synthetic
You should now have the following four files in your synthetic
folder :
foo@bar:~$ ls -1 synthetic/
generation.20190908194059502947.cnt_data.tsv
generation.20190908194059502947.mta_data.tsv
validation.20190908194059502947.cnt_data.tsv
validation.20190908194059502947.mta_data.tsv
We will use the "generation" set to generate synthetic ST data. As in the paper we use 1000 spots and 500 genes. To easily keep track of our files we tag them with st-hippo-comp. The validation set is set aside, to later be used as single cell data input to each respective method upon deconvolution.
foo@bar:~$ ../../comparison/synthetic_data_generation/make_st_set.py -c synthetic/generation.20190908194059502947.cnt_data.tsv -l synthetic/generation.20190908194059502947.mta_data.tsv -ns 1000 -ng 500 -o synthetic -t st-hippo-comp
A total of three files will be generated - ST expression data (counts), proportion values (proportions) and the number of cells (members). Since we know the actual proportions (ground truth) of cell types within each spot, this data is suitable to use for comparison of method performance.
foo@bar:~$ ls -1 synthetic/
counts.st-hippo-comp.tsv
generation.20190908194059502947.cnt_data.tsv
generation.20190908194059502947.mta_data.tsv
members.st-hippo-comp.tsv
proportions.st-hippo-comp.tsv
validation.20190908194059502947.cnt_data.tsv
validation.20190908194059502947.mta_data.tsv
We will run stereoscope
with the following arguments (see previous section for more details regarding the stereoscope
interface):
foo@bar:~$ stereoscope run --sc_cnt synthetic/valid*cnt* --sc_labels synthetic/valid*mta* -scb 256 \
-sce 50000 --st_cnt synthetic/counts.st-hippo-comp.tsv -ste 50000 -o ../res/comp-stereoscope \
--gpu -lr 0.1
Where the results will be saved to comp-stereoscope
in the res
folder
Since none of the other two methods were designed with ST data in mind, we have written wrappers for them. This
allows ST data to be used as input and render similar output files to stereoscope
.
We'd also like to state that it is not our intent to criticize these methods, but rather show that they are not optimal for usage with ST-data.
Disclaimer: While we've tried to add checks for dependencies and automatic installation of the required R
packages
for these two methods; deconvSeq
and DWLS
are not our tools and hence package dependencies may become a problem - we refer
to respective method's github page for further help in addressing such issues.
DWLS
Small modifications were made to the DWLS
code; some of the provided functions crashed upon execution, these changes
were however minor and do not affect the method itself. We provide a ''modded'' version of their file
Deconvolution_functions.R
. The two modifications we made are marked with the tag #MODIFICATION
within the code.
Using Seurat
for extraction of DE-genes was not successfull, thus we used MAST
for this purpose. Negative values were
occasionally given for some proportions, all of these were set to zero.
To run DWLS
go to the main directory of this repo and enter:
foo@bar:~$ comparison/alternative_methods/DWLS/DWLS-implementation.R -wd res/comp-DWLS \
-sc data/comp/synthetic/validation*cnt_data.tsv -mt data/comp/synthetic/validation.*.mta_data.tsv \
-st data/comp/synthetic/counts*.tsv
This will generate an output file formatted like that of stereoscope
, that is: a matrix with dimensions n_spots x n_types
,
where each element represents the proportion of a specific type within a given spot.
deconvSeq
We followed the instructions given in the HTML-Vignette
for deconvSeq
in order to estimate the proportions - no cell cycle filtering was performed.
To run deconvSeq
go to the main directory of the repo and enter:
foo@bar:~$ comparison/alternative_methods/deconvSeq/deonvSeq-implementation.R \
-sc data/comp/synthetic/validation.*.cnt_data.tsv \
-mt data/comp/synthetic/validation.*.mta_data.tsv \
-st data/comp/synthetic/counts.*.tsv -o res/comp-deconvSeq
NOTE : We noticed that the bioMart-dependent functions in deconvSeq do not always complete successfully, if such an error arise wait a minute and try again.
Having estimated the proportions using all three methods, we can now compare them. As a performance metric we compute
the RMSE between the actual proportion values within each spot and the estimated values. The results are visualizaed in a
boxplot, and a one-sided test (Wilcoxon signed-rank test) is also conducted to see whether stereoscope
performs better than the other methods.
foo@bar:~$ comparison/compare.py -rf res/comp-stereoscope/*/W*tsv res/deconvSeq-proportions.tsv res/comp-DWLS/proportions.tsv \
-tf data/comp/synthetic/proportions.hippo.tsv -o res/comp -mn stereoscope deconvseq DWLS
Which will generate the following image:
The dashed line represents the mean of the average performance upon randomly sampling propotrions from a homogenous Dirichlet distribution (concentration 1), repeated 1000 times.
Occasionaly you might face a scenario where you have run stereoscope
in a previous analysis and now want to use the
same single cell data set but different (or more) ST-data. The way we have implemented our method and built stereoscope
allows you to this in a seamless manner.
To exemplify, imagine you've been working with lung cancer ST and single cell data, and mapped the latter onto the former. Now assume your collaborators tell you that they have a nice set of breast cancer data that they want you to analyze for them. While a breast cancer data set would be ideal to work with, one initial analysis you could do is to see how the lung cancer single cell data maps onto the breast cancer spatial data - potenitally removing some of the lung-specific types.
We have downloaded and curated the data set presented in the publication _"Single-Cell Transcriptomics of Human and Mouse Lung Cancers Reveals Conserved Myeloid Populations across Individuals and Species", meaning we removed cell types which we considered as lung-specific such as club cells and ciliated cells. For more infomration regarding the data set see Rapolas. et. al.. We have also estimated single cell parameters for this data, which we provide you with (to emulate the scenario described above).
As for the ST data - in the original ST publication 4 Breast Cancer sections were presented - we will analyze theses breast cancer samples using the Lung Cancer single cell data set. Fortunately half of the work is already done.
Above : parts of Figure 4 in the original ST publication, displaying some of the breast cancer tissue.
The data is available at
this link, download the 4
breast cancer samples and put them in data/add-bc
either interactively (make sure to rename them as below) or using the terminal :
foo@bar:~$ cd data
foo@bar:~$ mkdir add-bc
foo@bar:~$ cd add-bc
foo@bar:~$ for ii in 1 2 3 4; do curl https://www.spatialresearch.org/wp-content/uploads/2016/07/Layer${ii}_BC_count_matrix-1.tsv --output st-bc${ii}.tsv; done
Once the download have completed you should have the following content in your data/add-bc
folder:
foo@bar:~$ ls -1
st-bc1.tsv
st-bc2.tsv
st-bc3.tsv
st-bc4.tsv
These files are already processed, only having spots under the tissue and uses HGNC gene symbols - hence there is no need to pre-process our data.
You will find the estimated single cell parameters (rates and logits) for the Lung Cancer data set in the
data/params-lc.zip
file, which when unzipped will create a folder data/params-lc
. To use these in
conjuction with the newly downloaded breast cancer ST-data simply do:
foo@bar:~$ stereoscope run --sc_fit ../params-lc/R*.tsv ../params/logits*.tsv --st_cnt st-bc*.tsv \
--ste 50000 -stb 256 -lr 0.01 --gpu -o ../../res/bc-add
[2019-10-30 05:44:00,255 - stsc - INFO ] >> Using device cuda
[2019-10-30 05:44:00,256 - stsc - INFO ] >> load sc parameter | rates (R) : ../params-lc/R.lc.tsv | logodds (logits) : ../params-lc/logits.lc.tsv
[2019-10-30 05:44:00,312 - stsc - INFO ] >> fit st data section(s) :['st-bc1.tsv', 'st-bc2.tsv', 'st-bc3.tsv','st-bc4.tsv']
[2019-10-30 05:44:13,668 - stsc - INFO ] >> ST data GENES : 16385 ST data SPOTS : 1031
The flag --sc_fit
expects two positional arguments, the first being the path to the rates and the second being
the path to the logits.
We can visualize the resullts by using the look
module of stereoscope
(see above for more detailed description).
This is simply done by :
foo@bar:~$ cd ../../res/bc-add
foo@bar:~$ stereoscope look -pp bc*/W*tsv -ms 80 -nc 5 -sc i -sb s -c "umap" -g -y -o viz
Which will generate a set of images like these:
We do not have a good reference like the Allen Brain Atlas to assess the quality of our results here, and we have no
ambition of presenting these as new findings in a biological context. Rather, this is to be considered as partly a
demonstration of how to use stereoscope
in different settings, but also how data sets from different tissues can be
combined if one desires.
- Implement
.h5ad
support - Add cell type priors based on single cell data
- implement subsampling as module
- cartal - for identifying small bugs in the
look
module and providing suggestions for additional features to enhance visualization.