Introduce
Quantizing the TIGER-Lab/StructLM-7B-Mistral to f16, q2, q3, q4, q5, q6 and q8 with Llama.cpp.
Prompt Template
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
### Instruction:
{instruction}
{input}
{question}
### Response:
IMPORTANT!! - For more details, check out StructLM-7B-Mistral.
- Downloads last month
- 32
Hardware compatibility
Log In to add your hardware
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
