Unified Conda and Pip requirements management.
Project description
🚀 UniDep - Unified Conda and Pip Dependency Management 🚀
UniDep streamlines Python project dependency management by unifying Conda and Pip packages in a single system. Learn when to use UniDep in our FAQ.
Handling dependencies in Python projects can be challenging, especially when juggling Python and non-Python packages. This often leads to confusion and inefficiency, as developers juggle between multiple dependency files.
- 📝 Unified Dependency File: Use either
requirements.yaml
orpyproject.toml
to manage both Conda and Pip dependencies in one place. - ⚙️ Build System Integration: Integrates with Setuptools and Hatchling for automatic dependency handling during
pip install ./your-package
. - 💻 One-Command Installation:
unidep install
handles Conda, Pip, and local dependencies effortlessly. - 🏢 Monorepo-Friendly: Render (multiple)
requirements.yaml
orpyproject.toml
files into one Condaenvironment.yaml
file and maintain fully consistent global and per sub packageconda-lock
files. - 🌍 Platform-Specific Support: Specify dependencies for different operating systems or architectures.
- 🔧
pip-compile
Integration: Generate fully pinnedrequirements.txt
files fromrequirements.yaml
orpyproject.toml
files usingpip-compile
. - 🔒 Integration with
conda-lock
: Generate fully pinnedconda-lock.yml
files from (multiple)requirements.yaml
orpyproject.toml
file(s), leveragingconda-lock
. - 🤓 Nerd stats: written in Python, >99% test coverage, fully-typed, all Ruff's rules enabled, easily extensible, and minimal dependencies
unidep
is designed to make dependency management in Python projects as simple and efficient as possible.
Try it now and streamline your development process!
[!TIP] Check out the example
requirements.yaml
andpyproject.toml
below.
:books: Table of Contents
- :package: Installation
- :memo:
requirements.yaml
andpyproject.toml
structure - :jigsaw: Build System Integration
- :desktop_computer: As a CLI
- ❓ FAQ
- Q: When to use UniDep?
- Q: Just show me a full example!
- Q: Uses of UniDep in the wild?
- Q: How is this different from conda/mamba/pip?
- Q: I found a project using unidep, now what?
- Q: How to handle local dependencies that do not use UniDep?
- Q: Can't Conda already do this?
- Q: What is the difference between
conda-lock
andunidep conda-lock
? - Q: What is the difference between
hatch-conda
/pdm-conda
andunidep
?
- :hammer_and_wrench: Troubleshooting
- :warning: Limitations
:package: Installation
To install unidep
, run the following command:
pip install "unidep[all]"
or
conda install -c conda-forge unidep
:memo: requirements.yaml
and pyproject.toml
structure
unidep
allows either using a
requirements.yaml
file with a specific format (similar but not the same as a Condaenvironment.yaml
file) orpyproject.toml
file with a[tool.unidep]
section.
Both files contain the following keys:
- name (Optional): For documentation, not used in the output.
- channels: List of conda channels for packages, such as
conda-forge
. - dependencies: Mix of Conda and Pip packages.
- local_dependencies (Optional): List of paths to other
requirements.yaml
orpyproject.toml
files to include. - platforms (Optional): List of platforms that are supported (used in
conda-lock
).
Whether you use a requirements.yaml
or pyproject.toml
file, the same information can be specified in either.
Choose the format that works best for your project.
Example
Example requirements.yaml
Example of a requirements.yaml
file:
name: example_environment
channels:
- conda-forge
dependencies:
- numpy # same name on conda and pip
- conda: python-graphviz # When names differ between Conda and Pip
pip: graphviz
- pip: slurm-usage >=1.1.0,<2 # pip-only
- conda: mumps # conda-only
# Use platform selectors
- conda: cuda-toolkit =11.8 # [linux64]
local_dependencies:
- ../other-project-using-unidep # include other projects that use unidep
- ../common-requirements.yaml # include other requirements.yaml files
- ../project-not-managed-by-unidep # 🚨 Skips its dependencies!
platforms: # (Optional) specify platforms that are supported (used in conda-lock)
- linux-64
- osx-arm64
[!IMPORTANT]
unidep
can process this duringpip install
and create a Conda installableenvironment.yaml
orconda-lock.yml
file, and more!
[!NOTE] For a more in-depth example containing multiple installable projects, see the
example
directory.
Example pyproject.toml
Alternatively, one can fully configure the dependencies in the pyproject.toml
file in the [tool.unidep]
section:
[tool.unidep]
channels = ["conda-forge"]
dependencies = [
"numpy", # same name on conda and pip
{ conda = "python-graphviz", pip = "graphviz" }, # When names differ between Conda and Pip
{ pip = "slurm-usage >=1.1.0,<2" }, # pip-only
{ conda = "mumps" }, # conda-only
{ conda = "cuda-toolkit =11.8:linux64" } # Use platform selectors by appending `:linux64`
]
local_dependencies = [
"../other-project-using-unidep", # include other projects that use unidep
"../common-requirements.yaml" # include other requirements.yaml files
"../project-not-managed-by-unidep" # 🚨 Skips its dependencies!
]
platforms = [ # (Optional) specify platforms that are supported (used in conda-lock)
"linux-64",
"osx-arm64"
]
This data structure is identical to the requirements.yaml
format, with the exception of the name
field and the platform selectors.
In the requirements.yaml
file, one can use e.g., # [linux64]
, which in the pyproject.toml
file is :linux64
at the end of the package name.
See Build System Integration for more information on how to set up unidep
with different build systems (Setuptools or Hatchling).
[!IMPORTANT] In these docs, we often mention the
requirements.yaml
format for simplicity, but the same information can be specified inpyproject.toml
as well. Everything that is possible inrequirements.yaml
is also possible inpyproject.toml
!
Key Points
- Standard names (e.g.,
- numpy
) are assumed to be the same for Conda and Pip. - Use a dictionary with
conda: <package>
andpip: <package>
to specify different names across platforms. - Use
pip:
to specify packages that are only available through Pip. - Use
conda:
to specify packages that are only available through Conda. - Use
# [selector]
(YAML only) orpackage:selector
to specify platform-specific dependencies. - Use
platforms:
to specify the platforms that are supported. - Use
local_dependencies:
to include otherrequirements.yaml
orpyproject.toml
files and merge them into one. Also allows projects that are not managed byunidep
to be included, but be aware that this skips their dependencies!
We use the YAML notation here, but the same information can be specified in
pyproject.toml
as well.
Supported Version Pinnings
UniDep supports a range of version pinning operators (the same as Conda):
-
Standard Version Constraints: Specify exact versions or ranges with standard operators like
=
,>
,<
,>=
,<=
.- Example:
=1.0.0
,>1.0.0, <2.0.0
.
- Example:
-
Version Exclusions: Exclude specific versions using
!=
.- Example:
!=1.5.0
.
- Example:
-
Redundant Pinning Resolution: Automatically resolves redundant version specifications.
- Example:
>1.0.0, >0.5.0
simplifies to>1.0.0
.
- Example:
-
Contradictory Version Detection: Errors are raised for contradictory pinnings to maintain dependency integrity. See the Conflict Resolution section for more information.
- Example: Specifying
>2.0.0, <1.5.0
triggers aVersionConflictError
.
- Example: Specifying
-
Invalid Pinning Detection: Detects and raises errors for unrecognized or improperly formatted version specifications.
-
Conda Build Pinning: UniDep also supports Conda's build pinning, allowing you to specify builds in your pinning patterns.
- Example: Conda supports pinning builds like
qsimcirq * cuda*
orvtk * *egl*
. - Limitation: While UniDep allows such build pinning, it requires that there be a single pin per package. UniDep cannot resolve conflicts where multiple build pinnings are specified for the same package.
- Example: UniDep can handle
qsimcirq * cuda*
, but it cannot resolve a scenario with bothqsimcirq * cuda*
andqsimcirq * cpu*
.
- Example: UniDep can handle
- Example: Conda supports pinning builds like
-
Other Special Cases: In addition to Conda build pins, UniDep supports all special pinning formats, such as VCS (Version Control System) URLs or local file paths. This includes formats like
package @ git+https://git/repo/here
orpackage @ file:///path/to/package
. However, UniDep has a limitation: it can handle only one special pin per package. These special pins can be combined with an unpinned version specification, but not with multiple special pin formats for the same package.- Example: UniDep can manage dependencies specified as
package @ git+https://git/repo/here
andpackage
in the samerequirements.yaml
. However, it cannot resolve scenarios where bothpackage @ git+https://git/repo/here
andpackage @ file:///path/to/package
are specified for the same package.
- Example: UniDep can manage dependencies specified as
[!WARNING] Pinning Validation and Combination: UniDep actively validates and/or combines pinnings only when multiple different pinnings are specified for the same package. This means if your
requirements.yaml
files include multiple pinnings for a single package, UniDep will attempt to resolve them into a single, coherent specification. However, if the pinnings are contradictory or incompatible, UniDep will raise an error to alert you of the conflict.
Conflict Resolution
unidep
features a conflict resolution mechanism to manage version conflicts and platform-specific dependencies in requirements.yaml
or pyproject.toml
files.
How It Works
-
Version Pinning Priority:
unidep
gives priority to version-pinned packages when the same package is specified multiple times. For instance, if bothfoo
andfoo <1
are listed,foo <1
is selected due to its specific version pin. -
Platform-Specific Version Pinning:
unidep
resolves platform-specific dependency conflicts by preferring the version with the narrowest platform scope. For instance, givenfoo <3 # [linux64]
andfoo >1
, it installsfoo >1,<3
exclusively on Linux-64 andfoo >1
on all other platforms. -
Intractable Conflicts: When conflicts are irreconcilable (e.g.,
foo >1
vs.foo <1
),unidep
raises an exception.
Platform Selectors
This tool supports a range of platform selectors that allow for specific handling of dependencies based on the user's operating system and architecture. This feature is particularly useful for managing conditional dependencies in diverse environments.
Supported Selectors
The following selectors are supported:
linux
: For all Linux-based systems.linux64
: Specifically for 64-bit Linux systems.aarch64
: For Linux systems on ARM64 architectures.ppc64le
: For Linux on PowerPC 64-bit Little Endian architectures.osx
: For all macOS systems.osx64
: Specifically for 64-bit macOS systems.arm64
: For macOS systems on ARM64 architectures (Apple Silicon).macos
: An alternative toosx
for macOS systems.unix
: A general selector for all UNIX-like systems (includes Linux and macOS).win
: For all Windows systems.win64
: Specifically for 64-bit Windows systems.
Usage
Selectors are used in requirements.yaml
files to conditionally include dependencies based on the platform:
dependencies:
- some-package >=1 # [unix]
- another-package # [win]
- special-package # [osx64]
- pip: cirq # [macos win]
conda: cirq # [linux]
Or when using pyproject.toml
instead of requirements.yaml
:
[tool.unidep]
dependencies = [
"some-package >=1:unix",
"another-package:win",
"special-package:osx64",
{ pip = "cirq:macos win", conda = "cirq:linux" },
]
In this example:
some-package
is included only in UNIX-like environments (Linux and macOS).another-package
is specific to Windows.special-package
is included only for 64-bit macOS systems.cirq
is managed bypip
on macOS and Windows, and byconda
on Linux. This demonstrates how you can specify different package managers for the same package based on the platform.
Note that the package-name:unix
syntax can also be used in the requirements.yaml
file, but the package-name # [unix]
syntax is not supported in pyproject.toml
.
Implementation
unidep
parses these selectors and filters dependencies according to the platform where it's being installed.
It is also used for creating environment and lock files that are portable across different platforms, ensuring that each environment has the appropriate dependencies installed.
:jigsaw: Build System Integration
[!TIP] See
example/
for working examples of usingunidep
with different build systems.
unidep
seamlessly integrates with popular Python build systems to simplify dependency management in your projects.
Example packages
Explore these installable example packages to understand how unidep
integrates with different build tools and configurations:
Project | Build Tool | pyproject.toml |
requirements.yaml |
setup.py |
---|---|---|---|---|
setup_py_project |
setuptools |
✅ | ✅ | ✅ |
setuptools_project |
setuptools |
✅ | ✅ | ❌ |
pyproject_toml_project |
setuptools |
✅ | ❌ | ❌ |
hatch_project |
hatch |
✅ | ✅ | ❌ |
hatch2_project |
hatch |
✅ | ❌ | ❌ |
Setuptools Integration
For projects using setuptools
, configure unidep
in pyproject.toml
and either specify dependencies in a requirements.yaml
file or include them in pyproject.toml
too.
- Using
pyproject.toml
only: The[project.dependencies]
field inpyproject.toml
gets automatically populated fromrequirements.yaml
or from the[tool.unidep]
section inpyproject.toml
. - Using
setup.py
: Theinstall_requires
field insetup.py
automatically reflects dependencies specified inrequirements.yaml
orpyproject.toml
.
Example pyproject.toml
Configuration:
[build-system]
build-backend = "setuptools.build_meta"
requires = ["setuptools", "unidep"]
[project]
dynamic = ["dependencies"]
Hatchling Integration
For projects managed with Hatch, unidep
can be configured in pyproject.toml
to automatically process the dependencies from requirements.yaml
or from the [tool.unidep]
section in pyproject.toml
.
Example Configuration for Hatch:
[build-system]
requires = ["hatchling", "unidep"]
build-backend = "hatchling.build"
[project]
dynamic = ["dependencies"]
# Additional project configurations
[tool.hatch]
# Additional Hatch configurations
[tool.hatch.metadata.hooks.unidep]
:desktop_computer: As a CLI
See example for more information or check the output of unidep -h
for the available sub commands:
usage: unidep [-h]
{merge,install,install-all,conda-lock,pip-compile,pip,conda,version}
...
Unified Conda and Pip requirements management.
positional arguments:
{merge,install,install-all,conda-lock,pip-compile,pip,conda,version}
Subcommands
merge Combine multiple (or a single) `requirements.yaml` or
`pyproject.toml` files into a single Conda installable
`environment.yaml` file.
install Automatically install all dependencies from one or
more `requirements.yaml` or `pyproject.toml` files.
This command first installs dependencies with Conda,
then with Pip. Finally, it installs local packages
(those containing the `requirements.yaml` or
`pyproject.toml` files) using `pip install [-e]
./project`.
install-all Install dependencies from all `requirements.yaml` or
`pyproject.toml` files found in the current directory
or specified directory. This command first installs
dependencies using Conda, then Pip, and finally the
local packages.
conda-lock Generate a global `conda-lock.yml` file for a
collection of `requirements.yaml` or `pyproject.toml`
files. Additionally, create individual `conda-
lock.yml` files for each `requirements.yaml` or
`pyproject.toml` file consistent with the global lock
file.
pip-compile Generate a fully pinned `requirements.txt` file from
one or more `requirements.yaml` or `pyproject.toml`
files using `pip-compile` from `pip-tools`. This
command consolidates all pip dependencies defined in
the `requirements.yaml` or `pyproject.toml` files and
compiles them into a single `requirements.txt` file,
taking into account the specific versions and
dependencies of each package.
pip Get the pip requirements for the current platform
only.
conda Get the conda requirements for the current platform
only.
version Print version information of unidep.
options:
-h, --help show this help message and exit
unidep merge
Use unidep merge
to scan directories for requirements.yaml
file(s) and combine them into an environment.yaml
file.
See unidep merge -h
for more information:
usage: unidep merge [-h] [-o OUTPUT] [-n NAME] [--stdout]
[--selector {sel,comment}] [-d DIRECTORY] [-v]
[--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}]
[--depth DEPTH] [--skip-dependency SKIP_DEPENDENCY]
[--ignore-pin IGNORE_PIN] [--overwrite-pin OVERWRITE_PIN]
Combine multiple (or a single) `requirements.yaml` or `pyproject.toml` files
into a single Conda installable `environment.yaml` file. Example usage:
`unidep merge --directory . --depth 1 --output environment.yaml` to search for
`requirements.yaml` or `pyproject.toml` files in the current directory and its
subdirectories and create `environment.yaml`. These are the defaults, so you
can also just run `unidep merge`.
options:
-h, --help show this help message and exit
-o OUTPUT, --output OUTPUT
Output file for the conda environment, by default
`environment.yaml`
-n NAME, --name NAME Name of the conda environment, by default `myenv`
--stdout Output to stdout instead of a file
--selector {sel,comment}
The selector to use for the environment markers, if
`sel` then `- numpy # [linux]` becomes `sel(linux):
numpy`, if `comment` then it remains `- numpy #
[linux]`, by default `sel`
-d DIRECTORY, --directory DIRECTORY
Base directory to scan for `requirements.yaml` or
`pyproject.toml` file(s), by default `.`
-v, --verbose Print verbose output
--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}, -p {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}
The platform(s) to get the requirements for. Multiple
platforms can be specified. By default, the current
platform (`linux-64`) is used.
--depth DEPTH Maximum depth to scan for `requirements.yaml` or
`pyproject.toml` files, by default 1
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
unidep install
Use unidep install
on one or more requirements.yaml
files and install the dependencies on the current platform using conda, then install the remaining dependencies with pip, and finally install the current package with pip install [-e] .
.
See unidep install -h
for more information:
usage: unidep install [-h] [-v] [-e] [--skip-local] [--skip-pip]
[--skip-conda] [--skip-dependency SKIP_DEPENDENCY]
[--no-dependencies]
[--conda-executable {conda,mamba,micromamba}]
[--dry-run] [--ignore-pin IGNORE_PIN]
[--overwrite-pin OVERWRITE_PIN]
files [files ...]
Automatically install all dependencies from one or more `requirements.yaml` or
`pyproject.toml` files. This command first installs dependencies with Conda,
then with Pip. Finally, it installs local packages (those containing the
`requirements.yaml` or `pyproject.toml` files) using `pip install [-e]
./project`. Example usage: `unidep install .` for a single project. For
multiple projects: `unidep install ./project1 ./project2`. The command accepts
both file paths and directories containing a `requirements.yaml` or
`pyproject.toml` file. Use `--editable` or `-e` to install the local packages
in editable mode. See `unidep install-all` to install all `requirements.yaml`
or `pyproject.toml` files in and below the current folder.
positional arguments:
files The `requirements.yaml` or `pyproject.toml` file(s) to
parse or folder(s) that contain those file(s), by
default `.`
options:
-h, --help show this help message and exit
-v, --verbose Print verbose output
-e, --editable Install the project in editable mode
--skip-local Skip installing local dependencies
--skip-pip Skip installing pip dependencies from
`requirements.yaml` or `pyproject.toml`
--skip-conda Skip installing conda dependencies from
`requirements.yaml` or `pyproject.toml`
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--no-dependencies Skip installing dependencies from `requirements.yaml`
or `pyproject.toml` file(s) and only install local
package(s). Useful after installing a `conda-lock.yml`
file because then all dependencies have already been
installed.
--conda-executable {conda,mamba,micromamba}
The conda executable to use
--dry-run, --dry Only print the commands that would be run
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
unidep install-all
Use unidep install-all
on a folder with packages that contain requirements.yaml
files and install the dependencies on the current platform using conda, then install the remaining dependencies with pip, and finally install the current package with pip install [-e] ./package1 ./package2
.
See unidep install-all -h
for more information:
usage: unidep install [-h] [-v] [-e] [--skip-local] [--skip-pip]
[--skip-conda] [--skip-dependency SKIP_DEPENDENCY]
[--no-dependencies]
[--conda-executable {conda,mamba,micromamba}]
[--dry-run] [--ignore-pin IGNORE_PIN]
[--overwrite-pin OVERWRITE_PIN]
files [files ...]
Automatically install all dependencies from one or more `requirements.yaml` or
`pyproject.toml` files. This command first installs dependencies with Conda,
then with Pip. Finally, it installs local packages (those containing the
`requirements.yaml` or `pyproject.toml` files) using `pip install [-e]
./project`. Example usage: `unidep install .` for a single project. For
multiple projects: `unidep install ./project1 ./project2`. The command accepts
both file paths and directories containing a `requirements.yaml` or
`pyproject.toml` file. Use `--editable` or `-e` to install the local packages
in editable mode. See `unidep install-all` to install all `requirements.yaml`
or `pyproject.toml` files in and below the current folder.
positional arguments:
files The `requirements.yaml` or `pyproject.toml` file(s) to
parse or folder(s) that contain those file(s), by
default `.`
options:
-h, --help show this help message and exit
-v, --verbose Print verbose output
-e, --editable Install the project in editable mode
--skip-local Skip installing local dependencies
--skip-pip Skip installing pip dependencies from
`requirements.yaml` or `pyproject.toml`
--skip-conda Skip installing conda dependencies from
`requirements.yaml` or `pyproject.toml`
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--no-dependencies Skip installing dependencies from `requirements.yaml`
or `pyproject.toml` file(s) and only install local
package(s). Useful after installing a `conda-lock.yml`
file because then all dependencies have already been
installed.
--conda-executable {conda,mamba,micromamba}
The conda executable to use
--dry-run, --dry Only print the commands that would be run
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
unidep conda-lock
Use unidep conda-lock
on one or multiple requirements.yaml
files and output the conda-lock file.
Optionally, when using a monorepo with multiple subpackages (with their own requirements.yaml
files), generate a lock file for each subpackage.
See unidep conda-lock -h
for more information:
usage: unidep conda-lock [-h] [--only-global] [--lockfile LOCKFILE]
[--check-input-hash] [-d DIRECTORY] [-v]
[--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}]
[--depth DEPTH] [--skip-dependency SKIP_DEPENDENCY]
[--ignore-pin IGNORE_PIN]
[--overwrite-pin OVERWRITE_PIN]
Generate a global `conda-lock.yml` file for a collection of
`requirements.yaml` or `pyproject.toml` files. Additionally, create individual
`conda-lock.yml` files for each `requirements.yaml` or `pyproject.toml` file
consistent with the global lock file. Example usage: `unidep conda-lock
--directory ./projects` to generate conda-lock files for all
`requirements.yaml` or `pyproject.toml` files in the `./projects` directory.
Use `--only-global` to generate only the global lock file. The `--check-input-
hash` option can be used to avoid regenerating lock files if the input hasn't
changed.
options:
-h, --help show this help message and exit
--only-global Only generate the global lock file
--lockfile LOCKFILE Specify a path for the global lockfile (default:
`conda-lock.yml` in current directory). Path should be
relative, e.g., `--lockfile ./locks/example.conda-
lock.yml`.
--check-input-hash Check existing input hashes in lockfiles before
regenerating lock files. This flag is directly passed
to `conda-lock`.
-d DIRECTORY, --directory DIRECTORY
Base directory to scan for `requirements.yaml` or
`pyproject.toml` file(s), by default `.`
-v, --verbose Print verbose output
--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}, -p {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}
The platform(s) to get the requirements for. Multiple
platforms can be specified. By default, the current
platform (`linux-64`) is used.
--depth DEPTH Maximum depth to scan for `requirements.yaml` or
`pyproject.toml` files, by default 1
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
unidep pip-compile
Use unidep pip-compile
on one or multiple requirements.yaml
files and output a fully locked requirements.txt
file using pip-compile
from pip-tools
.
See unidep pip-compile -h
for more information:
usage: unidep pip-compile [-h] [-o OUTPUT_FILE] [-d DIRECTORY] [-v]
[--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}]
[--depth DEPTH] [--skip-dependency SKIP_DEPENDENCY]
[--ignore-pin IGNORE_PIN]
[--overwrite-pin OVERWRITE_PIN]
...
Generate a fully pinned `requirements.txt` file from one or more
`requirements.yaml` or `pyproject.toml` files using `pip-compile` from `pip-
tools`. This command consolidates all pip dependencies defined in the
`requirements.yaml` or `pyproject.toml` files and compiles them into a single
`requirements.txt` file, taking into account the specific versions and
dependencies of each package. Example usage: `unidep pip-compile --directory
./projects` to generate a `requirements.txt` file for all `requirements.yaml`
or `pyproject.toml` files in the `./projects` directory. Use `--output-file
requirements.txt` to specify a different output file.
positional arguments:
extra_flags Extra flags to pass to `pip-compile`. These flags are
passed directly and should be provided in the format
expected by `pip-compile`. For example, `unidep pip-
compile -- --generate-hashes --allow-unsafe`. Note
that the `--` is required to separate the flags for
`unidep` from the flags for `pip-compile`.
options:
-h, --help show this help message and exit
-o OUTPUT_FILE, --output-file OUTPUT_FILE
Output file for the pip requirements, by default
`requirements.txt`
-d DIRECTORY, --directory DIRECTORY
Base directory to scan for `requirements.yaml` or
`pyproject.toml` file(s), by default `.`
-v, --verbose Print verbose output
--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}, -p {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}
The platform(s) to get the requirements for. Multiple
platforms can be specified. By default, the current
platform (`linux-64`) is used.
--depth DEPTH Maximum depth to scan for `requirements.yaml` or
`pyproject.toml` files, by default 1
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
unidep pip
Use unidep pip
on a requirements.yaml
file and output the pip installable dependencies on the current platform (default).
See unidep pip -h
for more information:
usage: unidep pip [-h] [-f FILE] [-v]
[--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}]
[--skip-dependency SKIP_DEPENDENCY]
[--ignore-pin IGNORE_PIN] [--overwrite-pin OVERWRITE_PIN]
[--separator SEPARATOR]
Get the pip requirements for the current platform only. Example usage: `unidep
pip --file folder1 --file folder2/requirements.yaml --seperator ' ' --platform
linux-64` to extract all the pip dependencies specific to the linux-64
platform. Note that the `--file` argument can be used multiple times to
specify multiple `requirements.yaml` or `pyproject.toml` files and that --file
can also be a folder that contains a `requirements.yaml` or `pyproject.toml`
file.
options:
-h, --help show this help message and exit
-f FILE, --file FILE The `requirements.yaml` or `pyproject.toml` file to
parse, or folder that contains that file, by default
`.`
-v, --verbose Print verbose output
--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}, -p {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}
The platform(s) to get the requirements for. Multiple
platforms can be specified. By default, the current
platform (`linux-64`) is used.
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
--separator SEPARATOR
The separator between the dependencies, by default ` `
unidep conda
Use unidep conda
on a requirements.yaml
file and output the conda installable dependencies on the current platform (default).
See unidep conda -h
for more information:
usage: unidep conda [-h] [-f FILE] [-v]
[--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}]
[--skip-dependency SKIP_DEPENDENCY]
[--ignore-pin IGNORE_PIN] [--overwrite-pin OVERWRITE_PIN]
[--separator SEPARATOR]
Get the conda requirements for the current platform only. Example usage:
`unidep conda --file folder1 --file folder2/requirements.yaml --seperator ' '
--platform linux-64` to extract all the conda dependencies specific to the
linux-64 platform. Note that the `--file` argument can be used multiple times
to specify multiple `requirements.yaml` or `pyproject.toml` files and that
--file can also be a folder that contains a `requirements.yaml` or
`pyproject.toml` file.
options:
-h, --help show this help message and exit
-f FILE, --file FILE The `requirements.yaml` or `pyproject.toml` file to
parse, or folder that contains that file, by default
`.`
-v, --verbose Print verbose output
--platform {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}, -p {linux-64,linux-aarch64,linux-ppc64le,osx-64,osx-arm64,win-64}
The platform(s) to get the requirements for. Multiple
platforms can be specified. By default, the current
platform (`linux-64`) is used.
--skip-dependency SKIP_DEPENDENCY
Skip installing a specific dependency that is in one
of the `requirements.yaml` or `pyproject.toml` files.
This option can be used multiple times, each time
specifying a different package to skip. For example,
use `--skip-dependency pandas` to skip installing
pandas.
--ignore-pin IGNORE_PIN
Ignore the version pin for a specific package, e.g.,
`--ignore-pin numpy`. This option can be repeated to
ignore multiple packages.
--overwrite-pin OVERWRITE_PIN
Overwrite the version pin for a specific package,
e.g., `--overwrite-pin 'numpy==1.19.2'`. This option
can be repeated to overwrite the pins of multiple
packages.
--separator SEPARATOR
The separator between the dependencies, by default ` `
❓ FAQ
Here is a list of questions we have either been asked by users or potential pitfalls we hope to help users avoid:
Q: When to use UniDep?
A: UniDep is particularly useful for setting up full development environments that require both Python and non-Python dependencies (e.g., CUDA, compilers, etc.) with a single command.
In fields like research, data science, robotics, AI, and ML projects, it is common to work from a locally cloned Git repository.
Setting up a full development environment can be a pain, especially if you need to install non Python dependencies like compilers, low-level numerical libraries, or CUDA (luckily Conda has all of them).
Typically, instructions are different for each OS and their corresponding package managers (apt
, brew
, yum
, winget
, etc.).
With UniDep, you can specify all your Pip and Conda dependencies in a single file.
To get set up on a new machine, you just need to install Conda (we recommend micromamba) and run pip install unidep; unidep install-all -e
in your project directory, to install all dependencies and local packages in editable mode in the current Conda environment.
For fully reproducible environments, you can run unidep conda-lock
to generate a conda-lock.yml
file.
Then, run conda env create -f conda-lock.yml -n myenv
to create a new Conda environment with all the third-party dependencies.
Finally, run unidep install-all -e --no-dependencies
to install all your local packages in editable mode.
For those who prefer not to use Conda, you can simply run pip install -e .
on a project using UniDep.
You'll need to install the non-Python dependencies yourself, but you'll have a list of them in the requirements.yaml
file.
In summary, use UniDep if you:
- Prefer installing packages with conda but still want your package to be pip installable.
- Are tired of synchronizing your Pip requirements (
requirements.txt
) and Conda requirements (environment.yaml
). - Want a low-effort, comprehensive development environment setup.
Q: Just show me a full example!
A: Check out the example
folder.
Q: Uses of UniDep in the wild?
A: UniDep really shines when used in a monorepo with multiple dependent projects, however, since these are typically private, we cannot share them.
However, an example of a single package that is public is home-assistant-streamdeck-yaml
.
This is a Python package that allows to interact with Home Assistant from an Elgato Stream Deck connected via USB to e.g., a Raspberry Pi.
It requires a couple of system dependencies (e.g., libusb
and hidapi
), which are typically installed with apt
or brew
.
The README.md
shows different installation instructions on Linux, MacOS, and Windows for non-Conda installs, however, with UniDep, we can just use unidep install .
on all platforms.
It is fully configured via pyproject.toml
.
The 2 Dockerfile
s show 2 different ways of using UniDep:
Dockerfile.locked
: Installingconda-lock.yml
(generated withunidep conda-lock
) and thenpip install .
the local package.Dockerfile.latest
: Usingunidep install .
to install all dependencies, first with conda, then pip, then the local package.
Q: How is this different from conda/mamba/pip?
A: UniDep uses pip and conda under the hood to install dependencies, but it is not a replacement for them. UniDep will print the commands it runs, so you can see exactly what it is doing.
Q: I found a project using unidep, now what?
A: You can install it like any other Python package using pip install
.
However, to take full advantage of UniDep's functionality, clone the repository and run unidep install-all -e
in the project directory.
This installs all dependencies in editable mode in the current Conda environment.
Q: How to handle local dependencies that do not use UniDep?
A: You can use the local_dependencies
field in the requirements.yaml
or pyproject.toml
file to specify local dependencies.
However, if a local dependency is not managed by UniDep, it will skip installing its dependencies!
To include all its dependencies, either convert the package to use UniDep (🏆), or maintain a separate requirements.yaml
file, e.g., for a package called foo
create, foo-requirements.yaml
:
dependencies:
# List the dependencies of foo here
- numpy
- scipy
- matplotlib
- bar
local_dependencies:
- ./path/to/foo # This is the path to the package
Then, in the requirements.yaml
or pyproject.toml
file of the package that uses foo
, list foo-requirements.yaml
as a local dependency:
local_dependencies:
- ./path/to/foo-requirements.yaml
Q: Can't Conda already do this?
A: Not quite. Conda can indeed install both Conda and Pip dependencies via an environment.yaml
file, however, it does not work the other way around.
Pip cannot install the pip
dependencies from an environment.yaml
file.
This means, that if you want your package to be installable with pip install -e .
and support Conda, you need to maintain two separate files: environment.yaml
and requirements.txt
(or specify these dependencies in pyproject.toml
or setup.py
).
Q: What is the difference between conda-lock
and unidep conda-lock
?
A: conda-lock
is a standalone tool that creates a conda-lock.yml
file from a environment.yaml
file.
On the other hand, unidep conda-lock
is a command within the UniDep tool that also generates a conda-lock.yml
file (leveraging conda-lock
), but it does so from one or more requirements.yaml
or pyproject.toml
files.
When managing multiple dependent projects (e.g., in a monorepo), a unique feature of unidep conda-lock
is its ability to create consistent individual conda-lock.yml
files for each requirements.yaml
or pyproject.toml
file, ensuring consistency with a global conda-lock.yml
file.
This feature is not available in the standalone conda-lock
tool.
Q: What is the difference between hatch-conda
/ pdm-conda
and unidep
?
A: hatch-conda
is a plugin for hatch
that integrates Conda environments into hatch
.
A key difference is that hatch-conda
keeps Conda and Pip dependencies separate, choosing to install packages with either Conda or Pip.
This results in Conda being a hard requirement, for example, if numba
is specified for Conda, it cannot be installed with Pip despite its availability on PyPI.
In contrast, UniDep does not require Conda.
Without Conda, it can still install any dependency that is available on PyPI (e.g., numba
is both Conda and Pip installable).
However, without Conda, UniDep will not install dependencies exclusive to Conda.
These Conda-specific dependencies can often be installed through alternative package managers like apt
, brew
, yum
, or by building them from source.
Another key difference is that hatch-conda
is managing Hatch environments whereas unidep
can install Pip dependencies in the current Python environment (venv, Conda, Hatch, etc.), however, to optimally use UniDep, we recommend using Conda environments to additionally install non-Python dependencies.
Similar to hatch-conda
, unidep
also integrates with Hatchling, but it works in a slightly different way.
A: pdm-conda
is a plugin for pdm
designed to facilitate the use of Conda environments in conjunction with pdm
.
Like hatch-conda
, pdm-conda
opts to install packages either with Conda or Pip.
It is closely integrated with pdm
, primarily enabling the inclusion of Conda packages in pdm
's lock file (pdm.lock
).
However, pdm-conda
lacks extensive cross-platform support.
For instance, when adding a package like Numba using pdm-conda
, it gets locked to the current platform (e.g., osx-arm64) without the flexibility to specify compatibility for other platforms such as linux64.
In contrast, UniDep allows for cross-platform compatibility, enabling the user to specify dependencies for multiple platforms.
UniDep currently does not support pdm
, but it does support Hatchling and Setuptools.
UniDep stands out from both pdm-conda
and hatch-conda
with its additional functionalities, particularly beneficial for monorepos and projects spanning multiple operating systems. For instance:
- Conda Lock Files: Create
conda-lock.yml
files for all packages with consistent sub-lock files per package. - CLI tools: Provides tools like
unidep install-all -e
which will install multiple local projects (e.g., in monorepo) and all its dependencies first with Conda, then remaining ones with Pip, and finally the local dependencies in editable mode with Pip. - Conda Environment Files: Can create standard Conda
environment.yaml
files by combining the dependencies from manyrequirements.yaml
orpyproject.toml
files. - Platform-Specific Dependencies: Allows specifying dependencies for certain platforms (e.g., linux64, osx-arm64), enhancing cross-platform compatibility.
:hammer_and_wrench: Troubleshooting
pip install
fails with FileNotFoundError
When using a project that uses local_dependencies: [../not/current/dir]
in the requirements.yaml
file:
local_dependencies:
# File in a different directory than the pyproject.toml file
- ../common-requirements.yaml
You might get an error like this when using a pip
version older than 22.0
:
$ pip install /path/to/your/project/using/unidep
...
File "/usr/lib/python3.8/pathlib.py", line 1222, in open
return io.open(self, mode, buffering, encoding, errors, newline,
File "/usr/lib/python3.8/pathlib.py", line 1078, in _opener
return self._accessor.open(self, flags, mode)
FileNotFoundError: [Errno 2] No such file or directory: '/tmp/common-requirements.yaml'
The solution is to upgrade pip
to version 22.0
or newer:
pip install --upgrade pip
:warning: Limitations
- Conda-Focused: Best suited for Conda environments. However, note that having
conda
is not a requirement to install packages that use UniDep. - Setuptools and Hatchling only: Currently only works with setuptools and Hatchling, not flit, poetry, or other build systems. Open an issue if you'd like to see support for other build systems.
- No logic operators in platform selectors and no Python selectors.
Try unidep
today for a streamlined approach to managing your Conda environment dependencies across multiple projects! 🎉👏
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file unidep-0.47.0.tar.gz
.
File metadata
- Download URL: unidep-0.47.0.tar.gz
- Upload date:
- Size: 81.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.12.1
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | b779830e992b6cf5f66733da8dc51228408aea4499b53ef739851fc1034b2b6e |
|
MD5 | 3cfb38e3cd20fe106d186ede14fc8011 |
|
BLAKE2b-256 | 5440e571464f8ea5d78ac74be857ff1104120473073c1effc592e807a28361b1 |
File details
Details for the file unidep-0.47.0-py3-none-any.whl
.
File metadata
- Download URL: unidep-0.47.0-py3-none-any.whl
- Upload date:
- Size: 49.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.12.1
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4105c670c6681fbf5100fbd53478d02cd485b596c4d1f16e47a73bec89cc3ee7 |
|
MD5 | 2a614e75e56d36df8d2cfee1e6df1fb8 |
|
BLAKE2b-256 | c4f8aa96d844c073fb3b5a9c228b64b52d84b68ae6cfa4b2f34b501cad05bb19 |