|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
- ru |
|
|
base_model: |
|
|
- h34v7/DXP-Zero-V1.0-24b-Small-Instruct |
|
|
--- |
|
|
|
|
|
# DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF |
|
|
|
|
|
BF16 available [here](https://huggingface.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct). |
|
|
|
|
|
### Recommended Settings |
|
|
``` |
|
|
"temperature": 0.8, |
|
|
"top_k": 40, |
|
|
"top_p": 0.95, |
|
|
"min_p": 0.05, |
|
|
"repeat_last_n": 40, |
|
|
"repeat_penalty": 1.2, |
|
|
``` |
|
|
|
|
|
### Run on Ollama |
|
|
These are non-imatrix. I'll release the imatrix version later. |
|
|
|
|
|
GGUF 3-bit Q3_K_M about 27 GB of vRAM/RAM: |
|
|
``` |
|
|
ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q3_K_M |
|
|
``` |
|
|
|
|
|
GGUF 4-bit Q4_K_M about 30 GB of vRAM/RAM: |
|
|
``` |
|
|
ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q4_K_M |
|
|
``` |
|
|
|
|
|
GGUF 5-bit Q5_K_M about 33 GB of vRAM/RAM: |
|
|
``` |
|
|
ollama run hf.co/h34v7/DXP-Zero-V1.0-24b-Small-Instruct-i1-GGUF:Q5_K_M |
|
|
``` |