COOLJAPAN
2026-03-26

OxiONNX 0.1.0 Released — Pure Rust ONNX Inference Engine with 147 Operators

High-performance ONNX runtime written entirely in pure Rust. Zero C/C++ dependencies, 147 operators fully supported, wgpu GPU acceleration, SIMD (AVX2/NEON), WASM + no_std ready, graph optimizer, async execution, model encryption. 30k+ SLoC, 590+ tests. The sovereign ONNX inference layer for SciRS2 and the entire COOLJAPAN ecosystem (now 21M+ SLoC total).

The ONNX inference foundation of the COOLJAPAN ecosystem just went fully sovereign and production-ready.

Today we released OxiONNX 0.1.0 — a complete, high-performance ONNX inference engine written entirely in pure Rust.

No C. No C++. No ONNX Runtime binaries. No external protobuf or CUDA dependencies.
No unsafe code in hot paths. No build hell.
Just clean, memory-safe, blazing-fast ONNX model execution that compiles to a single static binary (or <1 MB WASM) and runs everywhere — from laptops to browsers to edge devices to cloud GPUs.

Why OxiONNX 0.1.0 is a game changer

For years, ONNX inference meant depending on the heavy C++/CUDA-based ONNX Runtime or vendor-specific runtimes.

These tools are powerful but suffer from:

OxiONNX 0.1.0 ends all of that.

It delivers high-performance inference while being 100% memory-safe and fully auditable.
Notable results:

Technical Deep Dive: How We Built a Production-Grade ONNX Runtime in Pure Rust

The architecture uses a clean multi-crate design, radically optimized for modern Rust:

  1. Core Layer (oxionnx-core)
    Tensor, DType, Graph, Operator trait, arena allocator, buffer pooling, strided views.

  2. Operators (oxionnx-ops)
    147 fully implemented operators with automatic type promotion and mixed precision (f16/f32).

  3. Proto & Graph (oxionnx-proto)
    Pure-Rust ONNX protobuf parser + graph optimizer (constant folding, operator fusion, CSE, dead-code elimination).

  4. GPU Backend (oxionnx-gpu)
    wgpu compute shaders (optional, zero-overhead fallback to CPU).

Key Rust advantages:

What’s inside 0.1.0 (released March 26)

This is the foundation

OxiONNX is now the official ONNX inference backend for the entire COOLJAPAN stack (total ecosystem: 21M+ SLoC Rust, 597 crates, 40+ production-grade libraries):

Repository: https://github.com/cool-japan/oxionnx

Star the repo if you want fast, safe, sovereign ONNX inference without C/C++ or vendor lock-in.

The era of “just use onnxruntime” with all its native dependencies is over.

Pure Rust ONNX inference is here — fast, portable, memory-safe, and sovereign.

KitaSan at COOLJAPAN OÜ March 26, 2026