Model info updated
Browse files
Qwen3-8B-Q3_K_S/README.md
CHANGED
|
@@ -21,7 +21,7 @@ Quantized version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) at **
|
|
| 21 |
## Model Info
|
| 22 |
|
| 23 |
- **Format**: GGUF (for llama.cpp and compatible runtimes)
|
| 24 |
-
- **Size**: 3.
|
| 25 |
- **Precision**: Q3_K_S
|
| 26 |
- **Base Model**: [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B)
|
| 27 |
- **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
@@ -122,7 +122,7 @@ Here’s how you can query this model via API using `curl` and `jq`. Replace the
|
|
| 122 |
|
| 123 |
```bash
|
| 124 |
curl http://localhost:11434/api/generate -s -N -d '{
|
| 125 |
-
"model": "hf.co/geoffmunn/Qwen3-8B:Q3_K_S
|
| 126 |
"prompt": "Repeat the following instruction exactly as given: Summarize what a neural network is in one sentence.",
|
| 127 |
"temperature": 0.5,
|
| 128 |
"top_p": 0.95,
|
|
|
|
| 21 |
## Model Info
|
| 22 |
|
| 23 |
- **Format**: GGUF (for llama.cpp and compatible runtimes)
|
| 24 |
+
- **Size**: 3.77 GB
|
| 25 |
- **Precision**: Q3_K_S
|
| 26 |
- **Base Model**: [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B)
|
| 27 |
- **Conversion Tool**: [llama.cpp](https://github.com/ggerganov/llama.cpp)
|
|
|
|
| 122 |
|
| 123 |
```bash
|
| 124 |
curl http://localhost:11434/api/generate -s -N -d '{
|
| 125 |
+
"model": "hf.co/geoffmunn/Qwen3-8B:Q3_K_S",
|
| 126 |
"prompt": "Repeat the following instruction exactly as given: Summarize what a neural network is in one sentence.",
|
| 127 |
"temperature": 0.5,
|
| 128 |
"top_p": 0.95,
|