NOTE: You will need a recent build of llama.cpp to run these quants (i.e. at least commit 494c870).

GGUF importance matrix (imatrix) quants for https://huggingface.co/ibm/labradorite-13b

Layers Context Template
40
4096
<|system|>
{sys_prompt}
<|user|>
{inputs}
<|assistant|>
{response}<|endoftext|>
Downloads last month
62
GGUF
Model size
13B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

3-bit

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for dranger003/labradorite-13b-iMat.GGUF

Quantized
(3)
this model