# burn-onnx **Repository Path**: daoos_admin/burn-onnx ## Basic Information - **Project Name**: burn-onnx - **Description**: No description available - **Primary Language**: Rust - **License**: Apache-2.0 - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2026-03-03 - **Last Updated**: 2026-03-03 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README
# Burn ONNX [![Current Crates.io Version](https://img.shields.io/crates/v/burn-onnx.svg)](https://crates.io/crates/burn-onnx) [![Documentation](https://img.shields.io/badge/docs-latest-blue)](https://docs.rs/burn-onnx) [![Test Status](https://github.com/tracel-ai/burn-onnx/actions/workflows/test.yml/badge.svg)](https://github.com/tracel-ai/burn-onnx/actions/workflows/test.yml) [![license](https://shields.io/badge/license-MIT%2FApache--2.0-blue)](LICENSE-MIT) [![Ask DeepWiki](https://deepwiki.com/badge.svg)](https://deepwiki.com/tracel-ai/burn-onnx) **Import ONNX models into the [Burn](https://burn.dev) deep learning framework.** [Repository](https://github.com/tracel-ai/burn-onnx) | [Burn Repository](https://github.com/tracel-ai/burn)
## Overview `burn-onnx` converts ONNX models to native Burn Rust code, allowing you to run models from PyTorch, TensorFlow, and other frameworks on any Burn backend - from WebAssembly to CUDA. **Key features:** - Generates readable, modifiable Rust source code from ONNX models - Produces `burnpack` weight files for efficient loading - Works with any Burn backend (CPU, GPU, WebGPU, embedded) - Supports both `std` and `no_std` environments - Full opset compliance: all supported operators work across ONNX opset versions 1 through 24 - Graph simplification (enabled by default): attention coalescing, constant folding, constant shape propagation, idempotent-op elimination, identity-element elimination, CSE, dead code elimination, and permute-reshape detection ## Quick Start Add to your `Cargo.toml`: ```toml [build-dependencies] burn-onnx = "0.21" ``` In your `build.rs`: ```rust use burn_onnx::ModelGen; fn main() { ModelGen::new() .input("src/model/my_model.onnx") .out_dir("model/") .run_from_script(); } ``` Include the generated code in `src/model/mod.rs`: ```rust pub mod my_model { include!(concat!(env!("OUT_DIR"), "/model/my_model.rs")); } ``` Then use the model: ```rust use burn::backend::NdArray; use crate::model::my_model::Model; let model: Model> = Model::default(); let output = model.forward(input_tensor); ``` For detailed usage instructions, see the [ONNX Import Guide](https://burn.dev/books/burn/onnx-import.html) in the Burn Book. ## Examples | Example | Description | | ------------------------------------------------------------- | ----------------------------------- | | [onnx-inference](examples/onnx-inference) | Basic ONNX model inference | | [image-classification-web](examples/image-classification-web) | WebAssembly/WebGPU image classifier | ## Supported Operators See the [Supported ONNX Operators](SUPPORTED-ONNX-OPS.md) table for the complete list of supported operators. ## Contributing We welcome contributions! Please read the [Contributing Guidelines](CONTRIBUTING.md) before opening a PR, and the [Development Guide](DEVELOPMENT-GUIDE.md) for architecture and implementation details. For questions and discussions, join us on [Discord](https://discord.gg/uPEBbYYDB6). ## License Licensed under either of [Apache License, Version 2.0](LICENSE-APACHE) or [MIT license](LICENSE-MIT) at your option.