Update README.md
Browse files
README.md
CHANGED
|
@@ -11,7 +11,7 @@ tags:
|
|
| 11 |
|
| 12 |
---
|
| 13 |
|
| 14 |
-
#
|
| 15 |
|
| 16 |
## Model Description
|
| 17 |
|
|
@@ -37,7 +37,7 @@ The model supports fluent conversational responses and reasoning-style outputs,
|
|
| 37 |
```python
|
| 38 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 39 |
|
| 40 |
-
model_name = "dheeyantra/
|
| 41 |
|
| 42 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 43 |
model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
|
|
@@ -85,7 +85,7 @@ Example vLLM command:
|
|
| 85 |
|
| 86 |
```bash
|
| 87 |
vllm serve \
|
| 88 |
-
--model dheeyantra/
|
| 89 |
--host 0.0.0.0 \
|
| 90 |
--port 8000
|
| 91 |
```
|
|
|
|
| 11 |
|
| 12 |
---
|
| 13 |
|
| 14 |
+
# Dhee-Nxtgen-Qwen3-Telugu-v1
|
| 15 |
|
| 16 |
## Model Description
|
| 17 |
|
|
|
|
| 37 |
```python
|
| 38 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 39 |
|
| 40 |
+
model_name = "dheeyantra/dhee-nxtgen-qwen3-telugu-v1"
|
| 41 |
|
| 42 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 43 |
model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
|
|
|
|
| 85 |
|
| 86 |
```bash
|
| 87 |
vllm serve \
|
| 88 |
+
--model dheeyantra/dhee-nxtgen-qwen3-telugu-v1 \
|
| 89 |
--host 0.0.0.0 \
|
| 90 |
--port 8000
|
| 91 |
```
|