RedHatAI/Mistral-Nemo-Instruct-2407-FP8
Text Generation
• 12B • Updated • 568
• 18
RedHatAI/Qwen2-57B-A14B-Instruct-FP8
Text Generation
• 57B • Updated • 597
• 1
RedHatAI/Llama-2-7b-chat-hf-FP8
Text Generation
• 7B • Updated • 205
RedHatAI/Mistral-7B-Instruct-v0.3-FP8
Text Generation
• 7B • Updated • 1.48k
• 3
RedHatAI/Qwen2-0.5B-Instruct-FP8
Text Generation
• 0.5B • Updated • 367
• 3
RedHatAI/Qwen2-1.5B-Instruct-FP8
Text Generation
• 2B • Updated • 21.7k
RedHatAI/Qwen2-7B-Instruct-FP8
Text Generation
• 8B • Updated • 5.77k
• • 2
RedHatAI/Qwen2-72B-Instruct-FP8
Text Generation
• 73B • Updated • 658
• 15
RedHatAI/Mixtral-8x7B-Instruct-v0.1-AutoFP8
Text Generation
• 47B • Updated • 47
• 3
RedHatAI/Meta-Llama-3-70B-Instruct-FP8
Text Generation
• 71B • Updated • 4.38k
• • 13
RedHatAI/Meta-Llama-3-8B-Instruct-FP8
Text Generation
• Updated • 7.74k
• • 24
RedHatAI/DeepSeek-Coder-V2-Lite-Base-FP8
Text Generation
• 16B • Updated • 9
RedHatAI/DeepSeek-Coder-V2-Lite-Instruct-FP8
Text Generation
• 16B • Updated • 50.5k
• 11
RedHatAI/Qwen2-7B-Instruct-quantized.w4a16
Text Generation
• 8B • Updated • 96
RedHatAI/Qwen2-72B-Instruct-quantized.w4a16
Text Generation
• 73B • Updated • 130
• 4
RedHatAI/Qwen2-1.5B-Instruct-quantized.w4a16
Text Generation
• 2B • Updated • 15
RedHatAI/Qwen2-0.5B-Instruct-quantized.w4a16
Text Generation
• 0.6B • Updated • 430
RedHatAI/Qwen2-72B-Instruct-quantized.w8a16
Text Generation
• 73B • Updated • 5
• 1
RedHatAI/Qwen2-7B-Instruct-quantized.w8a16
Text Generation
• 8B • Updated • 1
RedHatAI/Qwen2-1.5B-Instruct-quantized.w8a16
Text Generation
• 2B • Updated • 2
RedHatAI/Qwen2-0.5B-Instruct-quantized.w8a16
Text Generation
• 0.5B • Updated • 2
RedHatAI/Llama-2-7b-chat-quantized.w4a16
Text Generation
• 7B • Updated • 124
RedHatAI/Meta-Llama-3-8B-Instruct-quantized.w4a16
Text Generation
• 8B • Updated • 92
• 2
RedHatAI/Llama-2-7b-chat-quantized.w8a16
Text Generation
• 7B • Updated • 3
RedHatAI/Mistral-7B-Instruct-v0.3-quantized.w8a16
Text Generation
• 7B • Updated • 98
RedHatAI/Meta-Llama-3-70B-Instruct-quantized.w8a16
Text Generation
• 71B • Updated • 9
• 6
RedHatAI/Meta-Llama-3-8B-Instruct-quantized.w8a16
Text Generation
• 8B • Updated • 506
• 3
RedHatAI/SparseLLama-2-7b-ultrachat_200k-pruned_50.2of4
Text Generation
• 7B • Updated • 2
RedHatAI/SparseLlama-2-7b-evolcodealpaca-pruned_50.2of4
Text Generation
• 7B • Updated • 3
RedHatAI/Meta-Llama-3-70B-Instruct-FP8-KV
Text Generation
• 71B • Updated • 9
• 3