Skip to main content

Utilities for expanding dask-jobqueue with appropriate settings for NCAR's clusters

Project description

ncar-jobqueue

Badges

CI GitHub Workflow Status GitHub Workflow Status Code Coverage Status
Docs Documentation Status
Package Conda PyPI
License License

Utilities for expanding dask-jobqueue with appropriate settings for NCAR's clusters.

Supported clusters:

  • Cheyenne
  • Casper (DAV)
  • CGD's Hobart
  • CGD's Izumi

Installation

NCAR-jobqueue can be installed from PyPI with pip:

python -m pip install ncar-jobqueue

(COMING SOON) NCAR-jobqueue will also be available from conda-forge for conda installations:

conda install -c conda-forge ncar-jobqueue

Usage

Casper

>>> from ncar_jobqueue import NCARCluster
>>> from dask.distributed import Client
>>> cluster = NCARCluster()
>>> cluster
NCARCluster(cores=0, memory=0 B, workers=0/0, jobs=0/0)
>>> cluster.scale(jobs=2)
>>> cluster
NCARCluster(cores=2, memory=50.00 GB, workers=2/2, jobs=2/2)
>>> client = Client(cluster)

Cheyenne

>>> from ncar_jobqueue import NCARCluster
>>> from dask.distributed import Client
>>> cluster = NCARCluster()
>>> cluster
NCARCluster(cores=0, memory=0 B, workers=0/0, jobs=0/0)
>>> cluster.scale(jobs=2)
>>> cluster
NCARCluster(cores=72, memory=218.00 GB, workers=2/2, jobs=2/2)
>>> client = Client(cluster)

Hobart

>>> from ncar_jobqueue import NCARCluster
>>> from dask.distributed import Client
>>> cluster = NCARCluster()
>>> cluster
NCARCluster(cores=0, memory=0 B, workers=0/0, jobs=0/0)
>>> cluster.scale(jobs=2)
>>> cluster
NCARCluster(cores=96, memory=192.00 GB, workers=2/2, jobs=2/2)
>>> client = Client(cluster)

Izumi

>>> from ncar_jobqueue import NCARCluster
>>> from dask.distributed import Client
>>> cluster = NCARCluster()
>>> cluster
NCARCluster(cores=0, memory=0 B, workers=0/0, jobs=0/0)
>>> cluster.scale(jobs=2)
>>> cluster
NCARCluster(cores=96, memory=192.00 GB, workers=2/2, jobs=2/2)
>>> client = Client(cluster)

Non-NCAR machines

On non-NCAR machines, ncar-jobqueue will warn the user, and it will use distributed.LocalCluster:

>>> from ncar_jobqueue import NCARCluster
.../ncar_jobqueue/cluster.py:42: UserWarning: Unable to determine which NCAR cluster you are running on... Returning a `distributed.LocalCluster` class.
warn(message)
>>> from dask.distributed import Client
>>> cluster = NCARCluster()
>>> cluster
NCARCluster('tcp://127.0.0.1:49334', workers=4, threads=8, memory=17.18 GB)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ncar-jobqueue-2020.11.18.tar.gz (14.7 kB view details)

Uploaded Source

Built Distribution

ncar_jobqueue-2020.11.18-py3-none-any.whl (11.6 kB view details)

Uploaded Python 3

File details

Details for the file ncar-jobqueue-2020.11.18.tar.gz.

File metadata

  • Download URL: ncar-jobqueue-2020.11.18.tar.gz
  • Upload date:
  • Size: 14.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.0 setuptools/49.2.1 requests-toolbelt/0.9.1 tqdm/4.52.0 CPython/3.9.0

File hashes

Hashes for ncar-jobqueue-2020.11.18.tar.gz
Algorithm Hash digest
SHA256 f5960caca4d4d0be1ceb46a4c851eee9ea78b18e97496f6f9e953a7257da8ad9
MD5 a3ce7d1dcdb537a183ff065514d716f5
BLAKE2b-256 416e4c71dd41595643d3ce1197f99c39c771ac8cbbb7e6c20811a0d059950df4

See more details on using hashes here.

File details

Details for the file ncar_jobqueue-2020.11.18-py3-none-any.whl.

File metadata

  • Download URL: ncar_jobqueue-2020.11.18-py3-none-any.whl
  • Upload date:
  • Size: 11.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.0 setuptools/49.2.1 requests-toolbelt/0.9.1 tqdm/4.52.0 CPython/3.9.0

File hashes

Hashes for ncar_jobqueue-2020.11.18-py3-none-any.whl
Algorithm Hash digest
SHA256 8cbe35894eab4c96cb1d565ac7b72a71276f9688074b37727205a191af550508
MD5 1a836bf5e4372a78e9e27cdc2bf78146
BLAKE2b-256 4159f1c52c314d7b1ea3c043dd7fed60c4ba0ed33d2bbb80483ac25839be26a1

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page