nm-testing 's Collections Models in CI
updated
nm-testing/Meta-Llama-3-8B-Instruct-W8A8-FP8-Channelwise-compressed-tensors
Text Generation
• 8B • Updated • 8
• 2
nm-testing/Meta-Llama-3-8B-Instruct-FBGEMM-nonuniform
Text Generation
• 8B • Updated • 15
nm-testing/Meta-Llama-3-8B-FP8-compressed-tensors-test
Text Generation
• 8B • Updated • 15.2k
nm-testing/Meta-Llama-3-8B-Instruct-W8-Channel-A8-Dynamic-Asym-Per-Token-Test
8B • Updated • 6.34k
• 1
nm-testing/Meta-Llama-3-8B-Instruct-W8-Channel-A8-Dynamic-Per-Token-Test
Text Generation
• 8B • Updated • 14
nm-testing/Meta-Llama-3-8B-Instruct-nonuniform-test
Text Generation
• 8B • Updated • 23.3k
nm-testing/Meta-Llama-3-70B-Instruct-FBGEMM-nonuniform
Text Generation
• 71B • Updated • 767
• 1
nm-testing/Qwen1.5-MoE-A2.7B-Chat-quantized.w4a16
14B • Updated • 106k
• 1
nm-testing/Qwen2-1.5B-Instruct-FP8W8
Text Generation
• 2B • Updated • 13
nm-testing/SparseLlama-3.1-8B-gsm8k-pruned.2of4-chnl_wts_per_tok_dyn_act_fp8-BitM
5B • Updated • 2
nm-testing/tinyllama-oneshot-w8w8-test-static-shape-change
Text Generation
• 1B • Updated • 59.9k
nm-testing/pixtral-12b-FP8-dynamic
Image-Text-to-Text
• Updated • 201
• 1
RedHatAI/Mistral-Small-3.1-24B-Instruct-2503-FP8-dynamic
Image-Text-to-Text
• 24B • Updated • 4.81k
• 9
nm-testing/Llama-3.2-1B-Instruct-FP8-KV
1B • Updated • 11.1k
nm-testing/tinyllama-oneshot-w8a8-channel-dynamic-token-v2
Text Generation
• 1B • Updated • 20.2k
nm-testing/tinyllama-oneshot-w8-channel-a8-tensor
Text Generation
• 1B • Updated • 822
RedHatAI/Llama-3.2-1B-quantized.w8a8
1B • Updated • 59.1k
• 1
nm-testing/tinyllama-oneshot-w8a8-dynamic-token-v2
Text Generation
• 1B • Updated • 15.3k
nm-testing/asym-w8w8-int8-static-per-tensor-tiny-llama
1B • Updated • 7.96k
nm-testing/Meta-Llama-3-8B-Instruct-W8A8-Static-Per-Tensor-Sym
8B • Updated • 37
nm-testing/Meta-Llama-3-8B-Instruct-W8A8-Static-Per-Tensor-Asym
8B • Updated • 38
nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_per_tok_dyn_act_int8-BitM
0.7B • Updated • 11
nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-chnl_wts_tensor_act_int8-BitM
0.7B • Updated • 8
nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_per_tok_dyn_act_int8-BitM
0.7B • Updated • 11
nm-testing/TinyLlama-1.1B-Chat-v1.0-gsm8k-pruned.2of4-tensor_wts_tensor_act_int8-BitM
0.7B • Updated • 8
nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Channel-Weight-testing
1B • Updated • 9
nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Static-testing
1B • Updated • 8
nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Tensor-Weight-testing
1B • Updated • 8
nm-testing/TinyLlama-1.1B-Chat-v1.0-2of4-Sparse-Dense-Compressor
1B • Updated • 11
nm-testing/llama2.c-stories42M-pruned2.4-compressed
48.6M • Updated • 8
nm-testing/TinyLlama-1.1B-Chat-v1.0-NVFP4
0.7B • Updated • 24.3k
nm-testing/Llama-3.2-1B-Instruct-spinquantR1R2R4-w4a16
0.7B • Updated • 7.69k
nm-testing/Llama-3.2-1B-Instruct-quip-w4a16
0.8B • Updated • 7.67k
nm-testing/tinyllama-oneshot-w4a16-channel-v2
Text Generation
• 0.3B • Updated • 17k
• 1
nm-testing/test-w4a16-mixtral-actorder-group
6B • Updated • 1.29k
nm-testing/TinyLlama-1.1B-Chat-v1.0-kvcache-fp8-attn_head
nm-testing/TinyLlama-1.1B-Chat-v1.0-kvcache-fp8-tensor
1B • Updated • 7.69k
nm-testing/Qwen3-30B-A3B-MXFP4A16
17B • Updated • 10.5k
nm-testing/Qwen3-0.6B-MXFP8
0.8B • Updated • 3
nm-testing/TinyLlama-1.1B-Chat-v1.0-MXFP8
nm-testing/dflash-qwen3-8b-speculators
2B • Updated • 12.4k
nm-testing/TinyLlama-1.1B-Chat-v1.0-MXFP4
0.6B • Updated • 86