Gemma-3-270M-Aurora-ML v3

LoRA fine-tune of unsloth/gemma-3-270m-it specialized for the ALCF Aurora supercomputer (Intel Xeon Sapphire Rapids + Intel GPU Max 1550 / Ponte Vecchio, oneAPI / SYCL, PBS Pro).

Off-the-shelf code-LLMs hallucinate Aurora specifics โ€” they suggest nvcc instead of icpx -fsycl, srun / aprun instead of mpiexec, NERSC's /global/cfs instead of /lus/flare, and CUDA device strings instead of xpu. This adapter teaches the base model the actual Aurora toolchain, file system layout, scheduler conventions, and recommended PyTorch/TensorFlow/SYCL idioms.

Model summary

Base model unsloth/gemma-3-270m-it
Format GGUF, f16 โ€” single file, llama.cpp / Ollama / LM Studio compatible
Fine-tuning LoRA (PEFT) โ€” r=32, ฮฑ=64, dropout 0.0, 2 epochs
Optimizer AdamW fused, lr 2e-4 cosine, warmup 3%, batch 1 ร— grad-accum 8
Precision / seq-len bf16, 1,536 tokens
Training data aurora-docs-distill-v2-datascience โ€” 1,117 ChatML rows
Train loss (final) 1.2462
Hardware 1 Aurora PVC tile (1/12 of a node, 64 GB HBM), IPEX + PyTorch 2.10 XPU backend
Eval (53-Q Aurora, 0โ€“5) pending

Quick start

On Aurora (PVC GPU, SYCL llama.cpp build) โ€” interactive PBS session:

# 1. Grab a debug node
qsub -I -A <project> -q debug -l select=1,walltime=01:00:00,filesystems=home:flare

# 2. Load the toolchain
module load frameworks
source /lus/flare/projects/<project>/scripts/env.sh   # or your own oneAPI setup
export ONEAPI_DEVICE_SELECTOR=level_zero:gpu

# 3. Download to flare (NOT $HOME โ€” quota is small)
hf download shazzadulimun/gemma3-270m-aurora-ml-v3-gguf --local-dir /lus/flare/projects/<project>/models/aurora-chat-v3

# 4. Run on a single PVC tile
/path/to/llama.cpp/build_sycl/bin/llama-cli \
    -m /lus/flare/projects/<project>/models/aurora-chat-v3/*.gguf \
    -ngl 999 -sm none --temp 0.0 -cnv \
    -p "How do I launch one MPI rank per GPU tile on Aurora?"

Anywhere else (laptop, workstation, any GPU):

hf download shazzadulimun/gemma3-270m-aurora-ml-v3-gguf --local-dir ./model
./llama-cli -m ./model/*.gguf -ngl 999 --temp 0.0 -cnv

Or Ollama / LM Studio: ollama run hf.co/shazzadulimun/gemma3-270m-aurora-ml-v3-gguf

Training data

Distilled from openai/gpt-oss-120b on ALCF Sophia (vLLM) over 130 cleaned chunks of docs.alcf.anl.gov/aurora. 1,117 training rows + 139 validation rows in ChatML format with embedded chain-of-thought (**Reasoning:** / **Answer:**).

Topic specialist โ€” Data Science / AI. Subset filtered to PyTorch/XPU, TensorFlow, vLLM, DeepSpeed, Megatron, JAX, Jupyter, and ML framework setup on Aurora. Trains a model to know torch.xpu, module load frameworks, IPEX patterns, etc.

Full corpus + reproduction scripts: SIslamMun/Generator @ aurora-datasets-2026-04-30.

Limitations

  • Synthetic-data biases. Teacher (gpt-oss-120b) can confabulate plausible-looking but incorrect commands. Treat outputs as a verifiable first draft, not authoritative.
  • Doc snapshot is fixed at 2026-04-29. Module versions, queue names, and APIs change โ€” anything published after that date isn't reflected here.
  • Aurora-only. Specifics (/lus/flare, xpu, PBS queues) won't transfer to Frontier, Polaris, or other systems.
  • Use temperature โ‰ค 0.1 for technical answers; higher temps invite invented flag names and paths.

Citation

@misc{aurora-llms-2026,
  title  = { Gemma-3-270M-Aurora-ML v3 },
  author = { Islam Mun, Shazzadul },
  year   = { 2026 },
  url    = { https://huggingface.co/shazzadulimun/gemma3-270m-aurora-ml-v3-gguf },
  note   = { LoRA fine-tune of gemma-3-270m-it; data distilled from gpt-oss-120b on docs.alcf.anl.gov/aurora }
}

License

Apache-2.0 for the adapter weights and synthetic training data. Source corpus is public ALCF user documentation. Base model retains its own license โ€” see unsloth/gemma-3-270m-it.

Downloads last month
240
GGUF
Model size
0.3B params
Architecture
gemma3
Hardware compatibility
Log In to add your hardware

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for shazzadulimun/gemma3-270m-aurora-ml-v3-gguf

Quantized
(28)
this model