Skip to main content

ABFE Workflow is a pipeline variant of our prior work.

Project description

ABFE_workflow

WARNING: The repo is currently under development - No Guarantees at the moment. But will be soon at a first version. This goes with a manuscript, that is currently submitted. :)

A snakemake based workflow for ABFE calculations using GMX. The workflow can be scaled on Slurm Queuing systems. The here provided Cyclophilin D Test systems and experimental values originate from:

Here a visualization of the triggered process:

New Features:

We are currently improving the user experience. The input was simplified to simply only the need of providing a pdb file for the receptor and .sdf files for the ligand.

>cli-abfe -h

usage: cli-abfe [-h] -p PROTEIN_PDB_PATH -l LIGAND_SDF_DIR -o OUTPUT_DIR_PATH [-c COFACTOR_SDF_PATH] [-nc NUMBER_OF_CPUS_PER_JOB] [-nj NUMBER_OF_PARALLEL_JOBS] [-nr NUMBER_OF_REPLICATES] [-submit]
                       [-gpu] [-hybrid]

optional arguments:
  -h, --help            show this help message and exit
  -p PROTEIN_PDB_PATH, --protein_pdb_path PROTEIN_PDB_PATH
                        Input protein pdb file path
  -l LIGAND_SDF_DIR, --ligand_sdf_dir LIGAND_SDF_DIR
                        Input ligand(s) sdf file path
  -o OUTPUT_DIR_PATH, --output_dir_path OUTPUT_DIR_PATH
                        Output approach folder
  -c COFACTOR_SDF_PATH, --cofactor_sdf_path COFACTOR_SDF_PATH
                        Input cofactor(s) sdf file path
  -nc NUMBER_OF_CPUS_PER_JOB, --number_of_cpus_per_job NUMBER_OF_CPUS_PER_JOB
                        Number of cpus per job
  -nj NUMBER_OF_PARALLEL_JOBS, --number_of_parallel_jobs NUMBER_OF_PARALLEL_JOBS
                        Number of jobs in parallel
  -nr NUMBER_OF_REPLICATES, --number_of_replicates NUMBER_OF_REPLICATES
                        Number of replicates
  -nosubmit             Will not automatically submit the ABFE calculations
  -nogpu                don't use gpus for the submissions?
  -nohybrid             don't do hybrid execution (complex jobs on gpu and ligand jobs on cpu (requires gpu flag))

Usage:

An example usage is provided with the examples/example_execution.sh, that uses the ABFE_Calculator.py script. If you remove the submit flag, you can a start a run, that only sets up the folder structure. (checkout our example folder)

Additional script information is provided via:

  conda activate abfe

  cli-abfe -h
  # or
  cli-abfe-gmx -h

Running an ABFE Campaign from Bash:

  conda activate abfe
  cli-abfe -p <path>/receptor.pdb \
           -l <path>/myligands \
           -o <path>/Out  \
           -nogpu -nohybrid -nc 8

Input

The input is suggested to be structured as follows for the commandline option:

  • <ligands>
    • ligand1.sdf
    • ligand2.sdf
    • ligand3.sdf
    • ...
  • receptor.pdb

For the python call:

  • ligand_sdfs:List[str] - paths to sdf files
  • protein_pdb_path: str - path to pdb file

Alternativley you can provide gromacs input files and use the command line tool ABFE_GMX_CLI. Please make sure your ligand is called LIG in the gmx files. The input sturucture should look like this:

  • <rooot_dir>
    • <ligand-1>
      • solvent
        • solvent.gro
        • solvent.top
      • complex
        • complex.gro
        • complex.top
    • <ligand-2>
      • solvent
        • solvent.gro
        • solvent.top
      • complex
        • complex.gro
        • complex.top ...

Install:

The package can be installed like the following script:

  cd ABFE_workflow
  conda env create --file ./environment.yml
  conda activate abfe
  pip install .

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ABFE_Workflow-0.5.0.tar.gz (51.8 kB view details)

Uploaded Source

Built Distribution

ABFE_Workflow-0.5.0-py3-none-any.whl (113.6 kB view details)

Uploaded Python 3

File details

Details for the file ABFE_Workflow-0.5.0.tar.gz.

File metadata

  • Download URL: ABFE_Workflow-0.5.0.tar.gz
  • Upload date:
  • Size: 51.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.10.12

File hashes

Hashes for ABFE_Workflow-0.5.0.tar.gz
Algorithm Hash digest
SHA256 c0a4c0d1d55b7c912e0ad7e52f1fcbe9167e619de28f7aa41d8f4cc0f36becc4
MD5 c8e85125d5f57a98b66c6e2249e3b246
BLAKE2b-256 6fcfeb93926fc52b626ccf7a69a3c353bd439af56b6cdb42de1eb2d4cb52ca9b

See more details on using hashes here.

Provenance

File details

Details for the file ABFE_Workflow-0.5.0-py3-none-any.whl.

File metadata

File hashes

Hashes for ABFE_Workflow-0.5.0-py3-none-any.whl
Algorithm Hash digest
SHA256 971679f7565c6b5e661827dbba1a862a3affaee21c2e0c18e333260ca5dde4ce
MD5 f3183e1515b7ba3a3c30dffb1ec26dd8
BLAKE2b-256 3b29056cddea43b159dc7b42a4b1b5138b5be18c82addd3a8b5b5785e65598b0

See more details on using hashes here.

Provenance

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page