Skip to main content

A simple wrapper process around cloud service providers to run tools for the RAPIDS Accelerator for Apache Spark.

Project description

spark-rapids-user-tools

User tools to help with the adoption, installation, execution, and tuning of RAPIDS Accelerator for Apache Spark.

The wrapper improves end-user experience within the following dimensions:

  1. Qualification: Educate the CPU customer on the cost savings and acceleration potential of RAPIDS Accelerator for Apache Spark. The output shows a list of apps recommended for RAPIDS Accelerator for Apache Spark with estimated savings and speed-up.
  2. Bootstrap: Provide optimized RAPIDS Accelerator for Apache Spark configs based on GPU cluster shape. The output shows updated Spark config settings on driver node.
  3. Tuning: Tune RAPIDS Accelerator for Apache Spark configs based on initial job run leveraging Spark event logs. The output shows recommended per-app RAPIDS Accelerator for Apache Spark config settings.
  4. Diagnostics: Run diagnostic functions to validate the Dataproc with RAPIDS Accelerator for Apache Spark environment to make sure the cluster is healthy and ready for Spark jobs.

Getting started

Set up a Python environment with a version between 3.8 and 3.10

  1. Run the project in a virtual environment.

    $ python -m venv .venv
    $ source .venv/bin/activate
    
  2. Install spark-rapids-user-tools

    • Using released package.

      $ pip install spark-rapids-user-tools
      
    • Install from source.

      $ pip install -e .
      

      Note that you can also use optional test to install dependencies required to run the unit-tests pip install -e '.[test]'

    • Using wheel package built from the repo (see the build steps below).

      $ pip install <wheel-file>
      
  3. Make sure to install CSP SDK if you plan to run the tool wrapper.

Building from source

Set up a Python environment similar to the steps above.

  1. Run the provided build script to compile the project.

    $> ./build.sh
    
  2. Fat Mode: Similar to fat jar in Java, this mode solves the problem when web access is not available to download resources having Url-paths (http/https).
    The command builds the tools jar file and downloads the necessary dependencies and packages them with the source code into a single 'wheel' file.

    $> ./build.sh fat
    

Usage and supported platforms

Please refer to spark-rapids-user-tools guide for details on how to use the tools and the platform.

What's new

Please refer to CHANGELOG.md for our latest changes.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

File details

Details for the file spark_rapids_user_tools-23.8.2-188_65bdbf8-py3-none-any.whl.

File metadata

File hashes

Hashes for spark_rapids_user_tools-23.8.2-188_65bdbf8-py3-none-any.whl
Algorithm Hash digest
SHA256 c1dfd46f8cee9e9a043fbb0fc81d611e37883793096db89b26b1ab756ea08f60
MD5 6af5397a5a8d42c18673df52345ac5f8
BLAKE2b-256 6c10cbb01c0a03d105f7d7f8132bbb47b794799899d71f1a7eb3916714c7f420

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page