Upload folder using huggingface_hub
Browse files- .gitattributes +1 -35
- Modelfile +11 -0
- README.md +52 -0
- llama3-thinkQ8.gguf +3 -0
.gitattributes
CHANGED
|
@@ -1,35 +1 @@
|
|
| 1 |
-
*.
|
| 2 |
-
*.arrow filter=lfs diff=lfs merge=lfs -text
|
| 3 |
-
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 4 |
-
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
| 5 |
-
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 6 |
-
*.ftz filter=lfs diff=lfs merge=lfs -text
|
| 7 |
-
*.gz filter=lfs diff=lfs merge=lfs -text
|
| 8 |
-
*.h5 filter=lfs diff=lfs merge=lfs -text
|
| 9 |
-
*.joblib filter=lfs diff=lfs merge=lfs -text
|
| 10 |
-
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
| 11 |
-
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
| 12 |
-
*.model filter=lfs diff=lfs merge=lfs -text
|
| 13 |
-
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
| 14 |
-
*.npy filter=lfs diff=lfs merge=lfs -text
|
| 15 |
-
*.npz filter=lfs diff=lfs merge=lfs -text
|
| 16 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
| 17 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
| 18 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
| 19 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
| 20 |
-
*.pickle filter=lfs diff=lfs merge=lfs -text
|
| 21 |
-
*.pkl filter=lfs diff=lfs merge=lfs -text
|
| 22 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
| 23 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
| 24 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
| 25 |
-
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
| 26 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
| 27 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
| 28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
| 29 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
| 30 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
| 31 |
-
*.wasm filter=lfs diff=lfs merge=lfs -text
|
| 32 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
| 33 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
-
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
| 1 |
+
*.gguf filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Modelfile
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
FROM llama3-thinkQ8.gguf
|
| 2 |
+
|
| 3 |
+
# Model parameters
|
| 4 |
+
PARAMETER temperature 0.8
|
| 5 |
+
PARAMETER top_p 0.9
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
# System prompt
|
| 9 |
+
SYSTEM """You are a helpful assistant. You will check the user request and you will think and generate brainstorming and self-thoughts in your mind and respond only in the following format:
|
| 10 |
+
<think> {your thoughts here} </think>
|
| 11 |
+
<answer> {your final answer here} </answer>. Use the tags once and place all your output inside them ONLY"""
|
README.md
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Llama3-ThinkQ8
|
| 2 |
+
|
| 3 |
+
A fine-tuned version of Llama 3 that shows explicit thinking using `<think>` and `<answer>` tags. This model is quantized to 8-bit (Q8) for efficient inference.
|
| 4 |
+
|
| 5 |
+
## Model Details
|
| 6 |
+
- **Base Model**: Llama 3
|
| 7 |
+
- **Quantization**: 8-bit (Q8)
|
| 8 |
+
- **Special Feature**: Explicit thinking process with tags
|
| 9 |
+
|
| 10 |
+
## How to Use with Ollama
|
| 11 |
+
|
| 12 |
+
### 1. Install Ollama
|
| 13 |
+
If you haven't already installed Ollama, follow the instructions at [ollama.ai](https://ollama.ai).
|
| 14 |
+
|
| 15 |
+
### 2. Download the model file
|
| 16 |
+
Download the GGUF file from this repository.
|
| 17 |
+
|
| 18 |
+
### 3. Create the Ollama model
|
| 19 |
+
Create a file named `Modelfile` with this content:
|
| 20 |
+
|
| 21 |
+
```
|
| 22 |
+
FROM llama3-thinkQ8.gguf
|
| 23 |
+
# Model parameters
|
| 24 |
+
PARAMETER temperature 0.8
|
| 25 |
+
PARAMETER top_p 0.9
|
| 26 |
+
# System prompt
|
| 27 |
+
SYSTEM """You are a helpful assistant. You will check the user request and you will think and generate brainstorming and self-thoughts in your mind and respond only in the following format:
|
| 28 |
+
<think> {your thoughts here} </think>
|
| 29 |
+
<answer> {your final answer here} </answer>. Use the tags once and place all your output inside them ONLY"""
|
| 30 |
+
```
|
| 31 |
+
|
| 32 |
+
Then run:
|
| 33 |
+
```bash
|
| 34 |
+
ollama create llama3-think -f Modelfile
|
| 35 |
+
```
|
| 36 |
+
|
| 37 |
+
### 4. Run the model
|
| 38 |
+
```bash
|
| 39 |
+
ollama run llama3-think
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
## Example Prompts
|
| 43 |
+
|
| 44 |
+
Try these examples:
|
| 45 |
+
|
| 46 |
+
```
|
| 47 |
+
Using each number in this tensor ONLY once (5, 8, 3) and any arithmetic operation like add, subtract, multiply, divide, create an equation that equals 19.
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
+
```
|
| 51 |
+
Explain the concept of quantum entanglement to a high school student.
|
| 52 |
+
```
|
llama3-thinkQ8.gguf
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e362b3bfe7ade389b5302bc626666b2295d720d5b575325aa622bca5edf2083a
|
| 3 |
+
size 8540775328
|