Experimental tools for converting PyTorch models to ONNX
Project description
PyTorch to ONNX Exporter
Experimental torch ONNX exporter.
[!WARNING] This is an experimental project and is not designed for production use. Use
torch.onnx.export
for these purposes.
Installation
pip install --upgrade torch-onnx
Usage
import torch
import torch_onnx
from onnxscript import ir
import onnx
# Get an exported program with torch.export
exported = torch.export.export(...)
model = torch_onnx.exported_program_to_ir(exported)
proto = ir.to_proto(model)
# This will give you an ATen dialect graph (un-lowered ONNX graph with ATen ops)
onnx.save(proto, "model.onnx")
# Or patch the torch.onnx export API
# Set error_report=True to get a detailed error report if the export fails
torch_onnx.patch_torch(error_report=True, profile=True)
torch.onnx.export(...)
# Use the analysis API to print an analysis report for unsupported ops
torch_onnx.analyze(exported)
Design
{dynamo/jit} -> {ExportedProgram} -> {torchlib} -> {ONNX IR} -> {ONNX}
- Use ExportedProgram
- Rely on robustness of the torch.export implementation
- Reduce complexity in the exporter
- This does not solve dynamo limitations, but it avoids introducing additional breakage by running fx passes
- Flat graph; Scope info as metadata, not functions
- Because existing tools are not good at handling them
- Eager optimization where appropriate
- Because exsiting tools are not good at optimizing
- Drop in replacement for torch.onnx.export
- Minimum migration effort
- Build graph eagerly in the exporter
- Give the exporter full control over the graph being built
Why is this doable?
- We need to verify torch.export coverage on Huggingface Optimum https://github.com/huggingface/optimum/tree/main/optimum/exporters/onnx; and they are not patching torch.onnx itself.
- Path torch.onnx.export such that packages do not need to change a single line to use dynamo
- We have all operators implemented and portable
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
torch_onnx-0.0.15.tar.gz
(36.2 kB
view details)
Built Distribution
File details
Details for the file torch_onnx-0.0.15.tar.gz
.
File metadata
- Download URL: torch_onnx-0.0.15.tar.gz
- Upload date:
- Size: 36.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/4.0.1 CPython/3.11.9
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 8807252d1286340eff178b41dc6f8db6c15e43161db0e326dcf95fe54cfe1ea9 |
|
MD5 | d7fcfcfba4f2d30337bb806881146613 |
|
BLAKE2b-256 | 3f42120e6d7bd2a683932d0e386f04202af4a570ead1874c99ad979c5400b4ba |
File details
Details for the file torch_onnx-0.0.15-py3-none-any.whl
.
File metadata
- Download URL: torch_onnx-0.0.15-py3-none-any.whl
- Upload date:
- Size: 40.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/4.0.1 CPython/3.11.9
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 74c5d7c1ecb5fadb3457dbe18670d448c6384e42a8692edb62c7ab84893c37b0 |
|
MD5 | b10ceaedf3e31ec837a074ba8bafe17d |
|
BLAKE2b-256 | 08d21c6c62fbfc3dab1163ae5bb9b254f95fd4eaedc6c78c888fd0058c340be7 |