High-performance LLM infrastructure — built entirely in Rust.
Tensorust is building the next generation of AI infrastructure, fast, safe, and uncompromising, using the power of Rust.
Where others glue together Python and C++, we build from scratch. Every layer of the stack, authored in systems-grade Rust. Why? Because the future of AI requires speed, reliability, and control at the metal.
We're not here to optimize someone else's mess. We're here to redefine the stack.
These aren’t side repos — they’re Tensorust’s foundational layers.
| Project | Description | Link |
|---|---|---|
everyother-token |
A custom Rust tokenizer engine optimized for transformer-based models | ‣ |
rust-perf-bench |
Performance benchmarking suite for LLM layers written in pure Rust | [Private/in development] |
gpt-json-parser |
Streaming JSON parser for LLM-generated content in Rust | [Private/in development] |
fastllama-rs (upcoming) |
Lightweight llama2/llama3 inference engine in Rust with Metal + CUDA backends | [Private/in development] |