How to use from
Unsloth Studio
Install Unsloth Studio (macOS, Linux, WSL)
curl -fsSL https://unsloth.ai/install.sh | sh
# Run unsloth studio
unsloth studio -H 0.0.0.0 -p 8888
# Then open http://localhost:8888 in your browser
# Search for ibm-granite/granite-4.1-30b-GGUF to start chatting
Install Unsloth Studio (Windows)
irm https://unsloth.ai/install.ps1 | iex
# Run unsloth studio
unsloth studio -H 0.0.0.0 -p 8888
# Then open http://localhost:8888 in your browser
# Search for ibm-granite/granite-4.1-30b-GGUF to start chatting
Using HuggingFace Spaces for Unsloth
# No setup required
# Open https://huggingface.co/spaces/unsloth/studio in your browser
# Search for ibm-granite/granite-4.1-30b-GGUF to start chatting
Quick Links

This repository contains models that have been converted to the GGUF format with various quantizations from an IBM Granite .safetensors model.

Please reference the base model's full model card here: https://huggingface.co/ibm-granite/granite-4.1-30b

Merging the .bf16 model

The bf16 model had to be split into multiple files to accommodate single file size restrictions using the llama-gguf-split tool, with its default --split settings, which can be built from the ggml-org/llama.cpp project.

Use the following command to merge the split files which points to the first file in the sequence:

llama-gguf-split --merge granite-4.1-30b-bf16-00001-of-00005.gguf granite-4.1-30b-bf16.gguf

The remaining split filenames are inferred by the tool based upon the 00001-of-0000x naming convention.

Downloads last month
232
GGUF
Model size
29B params
Architecture
granite
Hardware compatibility
Log In to add your hardware

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ibm-granite/granite-4.1-30b-GGUF

Quantized
(25)
this model

Collection including ibm-granite/granite-4.1-30b-GGUF