mistralrs-community/gemma-4-E4B-it-UQFF
Updated
None defined yet.
Community hub for pre-quantized models in UQFF format, ready to run with mistral.rs.
Fast, flexible LLM inference. A Rust-native inference engine supporting text, image, video, and audio — with built-in quantization, tool calling, web search, and MCP client support.
mistralrs quantize.mistralrs serve --ui gives you a web interface instantly.mistralrs tune benchmarks your system and picks optimal quantization + device mapping.Install (Linux/macOS):
curl --proto '=https' --tlsv1.2 -sSf https://raw.githubusercontent.com/EricLBuehler/mistral.rs/master/install.sh | sh
Install (Windows):
irm https://raw.githubusercontent.com/EricLBuehler/mistral.rs/master/install.ps1 | iex
Run a UQFF model:
mistralrs run -m mistralrs-community/gemma-4-E4B-it-UQFF --from-uqff 4
Or quantize any model on the fly with ISQ:
mistralrs run -m google/gemma-4-E4B-it --isq 4