AI & ML interests

Model Quantization

Recent Activity

tomkay  updated a collection about 8 hours ago
Qwen3.5-122B-A10B
tomkay  updated a collection about 8 hours ago
Qwen3.5-35B-A3B
View all activity

Organization Card

🐑 baa.ai

Smaller. Smarter. Sovereign.

Making frontier models run anywhere

We publish high-quality mixed-precision quantized models for Apple Silicon and GGUF. Our models use a proprietary optimisation method that delivers superior quality at your target memory budget

🌐 Website · 


🤗 Models

All models are published as MLX (Apple Silicon) and GGUF (cross-platform) formats.

Model Family Sizes Available Format
Qwen3.5-397B 220GB, 224GB MLX, GGUF
Qwen3.5-122B 52GB, 128GB, 154GB MLX, GGUF
Qwen3.5-35B 15–51GB (8 variants) MLX, GGUF
Llama-4-Maverick (402B) 407GB MLX, GGUF
Llama-4-Scout (109B) 117GB MLX, GGUF
Llama-3.1/3.3-70B 47GB MLX
MiniMax-M2.5 (229B) 179GB MLX, GGUF
Nemotron-120B MLX, GGUF
GLM-4.7-Flash 16GB MLX
Qwen3-30B/8B 16GB, 6GB MLX

baa.ai

datasets 0

None public yet