Skip to main content

Scale serial and MPI-parallel python functions over hundreds of compute nodes all from within a jupyter notebook or serial python process.

Project description

pympipool - up-scale python functions for high performance computing

Unittests Coverage Status Binder

Up-scaling python functions for high performance computing (HPC) can be challenging. While the python standard library provides interfaces for multiprocessing and asynchronous task execution, namely multiprocessing and concurrent.futures both are limited to the execution on a single compute node. So a series of python libraries have been developed to address the up-scaling of python functions for HPC. Starting in the datascience and machine learning community with solutions like dask over more HPC focused solutions like fireworks and parsl up to Python bindings for the message passing interface (MPI) named mpi4py. Each of these solutions has their advantages and disadvantages, in particular scaling beyond serial python functions, including thread based parallelism, MPI parallel python application or assignment of GPUs to individual python function remains challenging.

To address these challenges pympipool is developed with three goals in mind:

  • Extend the standard python library concurrent.futures.Executor interface, to minimize the barrier of up-scaling an existing workflow to be used on HPC resources.
  • Integrate thread based parallelism, MPI parallel python functions based on mpi4py and GPU assignment. This allows the users to accelerate their workflows one function at a time.
  • Embrace Jupyter notebooks for the interactive development of HPC workflows, as they allow the users to document their though process right next to the python code and their results all within one document.

HPC Context

In contrast to frameworks like dask, fireworks and parsl which can be used to submit a number of worker processes directly the the HPC queuing system and then transfer tasks from either the login node or an interactive allocation to these worker processes to accelerate the execution, mpi4py and pympipool follow a different approach. Here the user creates their HPC allocation first and then mpi4py or pympipool can be used to distribute the tasks within this allocation. The advantage of this approach is that no central data storage is required as the workers and the scheduling task can communicate directly.

Examples

The following examples illustrates how pympipool can be used to distribute a series of MPI parallel function calls within a queuing system allocation. example.py:

from pympipool import Executor

def calc(i):
    from mpi4py import MPI
    size = MPI.COMM_WORLD.Get_size()
    rank = MPI.COMM_WORLD.Get_rank()
    return i, size, rank

with Executor(max_workers=2, cores_per_worker=2) as exe:
    fs_0 = exe.submit(calc, 0)
    fs_1 = exe.submit(calc, 1)
    print(fs_0.result(), fs_1.result())

This example can be executed using::

python example.py

Which returns::

>>> [(0, 2, 0), (0, 2, 1)], [(1, 2, 0), (1, 2, 1)]

The important part in this example is that mpi4py is only used in the calc() function, not in the python script, consequently it is not necessary to call the script with mpiexec but instead a call with the regular python interpreter is sufficient. This highlights how pympipool allows the users to parallelize one function at a time and not having to convert their whole workflow to use mpi4py. The same code can also be executed inside a jupyter notebook directly which enables an interactive development process.

The standard concurrent.futures.Executor interface is extended by adding the option cores_per_worker=2 to assign multiple MPI ranks to each function call. To create two workers max_workers=2 each with two cores each requires a total of four CPU cores to be available. After submitting the function calc() with the corresponding parameter to the executor exe.submit(calc, 0) a python concurrent.futures.Future is returned. Consequently, the pympipool.Executor can be used as a drop-in replacement for the concurrent.futures.Executor which allows the user to add parallelism to their workflow one function at a time.

Backends

Depending on the availability of different resource schedulers in your HPC environment the pympipool.Executor uses a different backend, with the pympipool.flux.PyFluxExecutor being the preferred backend:

  • pympipool.mpi.PyMpiExecutor: The simplest executor of the three uses mpi4py as a backend. This simplifies the installation on all operating systems including Windows. Still at the same time it limits the up-scaling to a single compute node and serial or MPI parallel python functions. There is no support for thread based parallelism or GPU assignment. This interface is primarily used for testing and developing or as a fall-back solution. It is not recommended to use this interface in production.
  • pympipool.slurm.PySlurmExecutor: The SLURM workload manager is commonly used on HPC systems to schedule and distribute tasks. pympipool provides a python interface for scheduling the execution of python functions as SLURM job steps which are typically created using the srun command. This executor supports serial python functions, thread based parallelism, MPI based parallelism and the assignment of GPUs to individual python functions. When the SLURM workload manager is installed on your HPC cluster this interface can be a reasonable choice, still depending on the SLURM workload manager configuration in can be limited in terms of the fine-grained scheduling or the responsiveness when working with hundreds of compute nodes in an individual allocation.
  • pympipool.flux.PyFluxExecutor: The flux framework is the preferred backend for pympipool. Just like the pympipool.slurm.PySlurmExecutor it supports serial python functions, thread based parallelism, MPI based parallelism and the assignment of GPUs to individual python functions. Still the advantages of using the flux framework as a backend are the easy installation, the faster allocation of resources as the resources are managed within the allocation and no central databases is used and the superior level of fine-grained resource assignment which is typically not available on HPC resource schedulers.

Each of these backends consists of two parts a broker and a worker. When a new tasks is submitted from the user it is received by the broker and the broker identifies the first available worker. The worker then executes a task and returns it to the broker, who returns it to the user. While there is only one broker per pympipool.Executor the number of workers can be specified with the max_workers parameter.

Disclaimer

While we try to develop a stable and reliable software library, the development remains a opensource project under the BSD 3-Clause License without any warranties::

BSD 3-Clause License

Copyright (c) 2022, Jan Janssen
All rights reserved.

Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:

* Redistributions of source code must retain the above copyright notice, this
  list of conditions and the following disclaimer.

* Redistributions in binary form must reproduce the above copyright notice,
  this list of conditions and the following disclaimer in the documentation
  and/or other materials provided with the distribution.

* Neither the name of the copyright holder nor the names of its
  contributors may be used to endorse or promote products derived from
  this software without specific prior written permission.

THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.

Documentation

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pympipool-0.7.15.tar.gz (29.7 kB view details)

Uploaded Source

Built Distribution

pympipool-0.7.15-py3-none-any.whl (33.7 kB view details)

Uploaded Python 3

File details

Details for the file pympipool-0.7.15.tar.gz.

File metadata

  • Download URL: pympipool-0.7.15.tar.gz
  • Upload date:
  • Size: 29.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/4.0.2 CPython/3.11.8

File hashes

Hashes for pympipool-0.7.15.tar.gz
Algorithm Hash digest
SHA256 77d7ad38cf914e26913091a6ed6461c4bf378f213f35121bd6b9235f29f38349
MD5 50a5798868897db9c6ff5a6685424bb1
BLAKE2b-256 69a3914aab8416f96bc9e0596306c1eb206ab3c796e0a22125c1c2b86f9791f9

See more details on using hashes here.

File details

Details for the file pympipool-0.7.15-py3-none-any.whl.

File metadata

  • Download URL: pympipool-0.7.15-py3-none-any.whl
  • Upload date:
  • Size: 33.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/4.0.2 CPython/3.11.8

File hashes

Hashes for pympipool-0.7.15-py3-none-any.whl
Algorithm Hash digest
SHA256 6036f00ce96d58399f6ef79665891028f5931e04dba43a4d3043fa9dd48915ec
MD5 1e5b2fd86cd9ba1e6cdeb68c988b8997
BLAKE2b-256 f8c0efa79af442870ee6bae1f0a2d0cbc088351ea30281124d34102de0057f3d

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page