High-performance ONNX runtime written entirely in pure Rust. Zero C/C++ dependencies, 147 operators fully supported, wgpu GPU acceleration, SIMD (AVX2/NEON), WASM + no_std ready, graph optimizer, async execution, model encryption. 30k+ SLoC, 590+ tests. The sovereign ONNX inference layer for SciRS2 and the entire COOLJAPAN ecosystem (now 21M+ SLoC total).
The ONNX inference foundation of the COOLJAPAN ecosystem just went fully sovereign and production-ready.
Today we released OxiONNX 0.1.0 — a complete, high-performance ONNX inference engine written entirely in pure Rust.
No C. No C++. No ONNX Runtime binaries. No external protobuf or CUDA dependencies.
No unsafe code in hot paths. No build hell.
Just clean, memory-safe, blazing-fast ONNX model execution that compiles to a single static binary (or <1 MB WASM) and runs everywhere — from laptops to browsers to edge devices to cloud GPUs.
For years, ONNX inference meant depending on the heavy C++/CUDA-based ONNX Runtime or vendor-specific runtimes.
These tools are powerful but suffer from:
OxiONNX 0.1.0 ends all of that.
It delivers high-performance inference while being 100% memory-safe and fully auditable.
Notable results:
The architecture uses a clean multi-crate design, radically optimized for modern Rust:
Core Layer (oxionnx-core)
Tensor, DType, Graph, Operator trait, arena allocator, buffer pooling, strided views.
Operators (oxionnx-ops)
147 fully implemented operators with automatic type promotion and mixed precision (f16/f32).
Proto & Graph (oxionnx-proto)
Pure-Rust ONNX protobuf parser + graph optimizer (constant folding, operator fusion, CSE, dead-code elimination).
GPU Backend (oxionnx-gpu)
wgpu compute shaders (optional, zero-overhead fallback to CPU).
Key Rust advantages:
no_std + alloc supportwasm-bindgen for browser-native AIOxiONNX is now the official ONNX inference backend for the entire COOLJAPAN stack (total ecosystem: 21M+ SLoC Rust, 597 crates, 40+ production-grade libraries):
Repository: https://github.com/cool-japan/oxionnx
Star the repo if you want fast, safe, sovereign ONNX inference without C/C++ or vendor lock-in.
The era of “just use onnxruntime” with all its native dependencies is over.
Pure Rust ONNX inference is here — fast, portable, memory-safe, and sovereign.
— KitaSan at COOLJAPAN OÜ March 26, 2026