Skip to content

dimforge/inferi

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

inferi − on-device GPU inference on every platform


inferi is a set of Rust libraries for local LLM and vision model inference on the GPU. Shaders are written in Rust via rust-gpu, targeting Vulkan/WebGPU through wgpu. It runs natively on desktop, mobile, and in the browser.

Warning: inferi is under active development and may be missing features. Contributions welcome!

Getting started

Install cargo-gpu (required to compile shaders):

cargo install --git https://github.com/Rust-GPU/cargo-gpu cargo-gpu
cargo gpu install

Run the chat app natively:

cd crates/inferi-chat
dx run --release --features desktop

Or the CLI version:

cd crates/inferi-chat
cargo run --release --features desktop -- --headless --inspect '/path/to/model.gguf'

About

Cross-platform GPU LLM inference in Rust

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages