Skip to main content

Compositional Perturbation Autoencoder (CPA)

Project description

CPA - Compositional Perturbation Autoencoder

PyPI version Documentation Status Downloads

What is CPA?

Alt text

CPA is a framework to learn the effects of perturbations at the single-cell level. CPA encodes and learns phenotypic drug responses across different cell types, doses, and combinations. CPA allows:

  • Out-of-distribution predictions of unseen drug and gene combinations at various doses and among different cell types.
  • Learn interpretable drug and cell-type latent spaces.
  • Estimate the dose-response curve for each perturbation and their combinations.
  • Transfer pertubration effects from on cell-type to an unseen cell-type.
  • Enable batch effect removal on a latent space and also gene expression space.

Installation

Installing CPA

You can install CPA using pip and also directly from the github to access latest development version. See detailed instructions here.

How to use CPA

Several tutorials are available here to get you started with CPA. The following table contains the list of tutorials:

Description Link
Predicting combinatorial drug perturbations Open In Colab - Open In Documentation
Predicting unseen perturbations uisng external embeddings enabling the model to predict unseen reponses to unseen drugs Open In Colab - Open In Documentation
Predicting combinatorial CRISPR perturbations Open In Colab - Open In Documentation
Context transfer (i.e. predict the effect of a perturbation (e.g. disease) on unseen cell types or transfer perturbation effects from one context to another) demo on IFN-β scRNA perturbation dataset Open In Colab - Open In Documentation
Batch effect removal in gene expression and latent space Open In Colab - Open In Documentation

How to optmize CPA hyperparamters for your data

We provide an example script to use the built-in hyperparameter optimization function in CPA (based on scvi-tools hyperparam optimizer). You can find the script at examples/tune_script.py.

After the hyperparameter optimization using tune_script.py is done, result_grid.pkl is saved in your current directory using the pickle library. You can load the results using the following code:

import pickle
with open('result_grid.pkl', 'rb') as f:
    result_grid = pickle.load(f)

From here, you can follow the instructions in the Ray Documentations to analyze the run, and choose the best hyperparameters for your data.

You can also use the integration with wandb to log the hyperparameter optimization results. You can find the script at examples/tune_script_wandb.py. --> use_wandb=True

Everything is based on Ray Tune. You can find more information about the hyperparameter optimization in the Ray Tune Documentations.

The tuner is adapted and adjusted from scvi-tools v1.2.0 (unreleased) release notes

Datasets and Pre-trained models

Datasets and pre-trained models are available here.

Recepie for Pre-processing a custom scRNAseq perturbation dataset

If you have access to you raw data, you can do the following steps to pre-process your dataset. A raw dataset should be a scanpy object containing raw counts and available required metadata (i.e. perturbation, dosage, etc.).

Pre-processing steps

  1. Check for required information in cell metadata: a) Perturbation information should be in adata.obs. b) Dosage information should be in adata.obs. In cases like CRISPR gene knockouts, disease states, time perturbations, etc, you can create & add a dummy dosage in your adata.obs. For example:

        adata.obs['dosage'] = adata.obs['perturbation'].astype(str).apply(lambda x: '+'.join(['1.0' for _ in x.split('+')])).values
    

    c) [If available] Cell type information should be in adata.obs. d) [Multi-batch integration] Batch information should be in adata.obs.

  2. Filter out cells with low number of counts (sc.pp.filter_cells). For example:

    sc.pp.filter_cells(adata, min_counts=100)
    

    [optional]

    sc.pp.filter_genes(adata, min_counts=5)
    
  3. Save the raw counts in adata.layers['counts'].

    adata.layers['counts'] = adata.X.copy()
    
  4. Normalize the counts (sc.pp.normalize_total).

    sc.pp.normalize_total(adata, target_sum=1e4, exclude_highly_expressed=True)
    
  5. Log transform the normalized counts (sc.pp.log1p).

    sc.pp.log1p(adata)
    
  6. Highly variable genes selection: There are two options: 1. Use the sc.pp.highly_variable_genes function to select highly variable genes. python sc.pp.highly_variable_genes(adata, n_top_genes=5000, subset=True) 2. (Highly Recommended specially for Multi-batch integration scenarios) Use scIB's highly variable genes selection function to select highly variable genes. This function is more robust to batch effects and can be used to select highly variable genes across multiple datasets. python import scIB adata_hvg = scIB.pp.hvg_batch(adata, batch_key='batch', n_top_genes=5000, copy=True)

Congrats! Now you're dataset is ready to be used with CPA. Don't forget to save your pre-processed dataset using adata.write_h5ad function.

Support and contribute

If you have a question or new architecture or a model that could be integrated into our pipeline, you can post an issue

Reference

If CPA is helpful in your research, please consider citing the Lotfollahi et al. 2023

@article{lotfollahi2023predicting,
    title={Predicting cellular responses to complex perturbations in high-throughput screens},
    author={Lotfollahi, Mohammad and Klimovskaia Susmelj, Anna and De Donno, Carlo and Hetzel, Leon and Ji, Yuge and Ibarra, Ignacio L and Srivatsan, Sanjay R and Naghipourfar, Mohsen and Daza, Riza M and 
    Martin, Beth and others},
    journal={Molecular Systems Biology},
    pages={e11517},
    year={2023}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cpa_tools-0.8.7.tar.gz (51.8 kB view details)

Uploaded Source

Built Distribution

cpa_tools-0.8.7-py3-none-any.whl (51.9 kB view details)

Uploaded Python 3

File details

Details for the file cpa_tools-0.8.7.tar.gz.

File metadata

  • Download URL: cpa_tools-0.8.7.tar.gz
  • Upload date:
  • Size: 51.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.9.18 Linux/6.5.0-26-generic

File hashes

Hashes for cpa_tools-0.8.7.tar.gz
Algorithm Hash digest
SHA256 b38040739b587fe278b2ef40e02f8f7f889ebd90ffc56ee8a9a03aece4f094f5
MD5 504981b978ac0884c59eccc2b3aea899
BLAKE2b-256 3798b3ee91b1ce2c72a231f25ebe8ad694e5bfe6074c4a3c8db0d9716f82e701

See more details on using hashes here.

File details

Details for the file cpa_tools-0.8.7-py3-none-any.whl.

File metadata

  • Download URL: cpa_tools-0.8.7-py3-none-any.whl
  • Upload date:
  • Size: 51.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.6.1 CPython/3.9.18 Linux/6.5.0-26-generic

File hashes

Hashes for cpa_tools-0.8.7-py3-none-any.whl
Algorithm Hash digest
SHA256 e51d6ee0efe95c93d7a7afecd502e3569bc043dfa0d61d8ffe7669c5faca00e0
MD5 1457c3db94b406cc6d74b202c12f065f
BLAKE2b-256 f61b23e790e8c4a8e33337fc9b88a7e56499cd70640686e5b8058d73fcc328d2

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page