Update app.py
Browse files
app.py
CHANGED
|
@@ -14,8 +14,8 @@ clients = {
|
|
| 14 |
"meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
|
| 15 |
model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
|
| 16 |
),
|
| 17 |
-
"
|
| 18 |
-
model="
|
| 19 |
),
|
| 20 |
"meta-llama/Llama-2-7b-chat-hf": InferenceClient(
|
| 21 |
model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
|
|
@@ -27,35 +27,41 @@ clients = {
|
|
| 27 |
|
| 28 |
# Model presets
|
| 29 |
presets = {
|
| 30 |
-
"mistralai/Mistral-7B-Instruct-v0.3": {
|
| 31 |
-
"Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.8},
|
| 32 |
-
"Normal": {"max_new_tokens": 512, "temperature": 0.6, "top_p": 0.75},
|
| 33 |
-
"Quality": {"max_new_tokens": 1024, "temperature": 0.45, "top_p": 0.60},
|
| 34 |
-
"Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.7},
|
| 35 |
-
},
|
| 36 |
"meta-llama/Llama-3.2-3B-Instruct": {
|
| 37 |
-
"Fast": {"max_new_tokens":
|
| 38 |
-
"Normal": {"max_new_tokens":
|
| 39 |
-
"Quality": {"max_new_tokens":
|
| 40 |
-
"Unreal Performance": {"max_new_tokens":
|
| 41 |
},
|
| 42 |
-
"
|
| 43 |
-
"Fast": {"max_new_tokens":
|
| 44 |
-
"Normal": {"max_new_tokens":
|
| 45 |
-
"Quality": {"max_new_tokens":
|
| 46 |
-
"Unreal Performance": {"max_new_tokens":
|
| 47 |
},
|
| 48 |
"meta-llama/Llama-2-7b-chat-hf": {
|
| 49 |
-
"Fast": {"max_new_tokens":
|
| 50 |
-
"Normal": {"max_new_tokens":
|
| 51 |
-
"Quality": {"max_new_tokens":
|
| 52 |
-
"Unreal Performance": {"max_new_tokens":
|
| 53 |
},
|
| 54 |
"meta-llama/Llama-3.3-70B-Instruct": {
|
| 55 |
-
"Fast": {"max_new_tokens":
|
| 56 |
-
"Normal": {"max_new_tokens":
|
| 57 |
-
"Quality": {"max_new_tokens":
|
| 58 |
-
"Unreal Performance": {"max_new_tokens":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
}
|
| 60 |
}
|
| 61 |
|
|
@@ -204,7 +210,6 @@ system_messages = {
|
|
| 204 |
|
| 205 |
# Model registry
|
| 206 |
model_registry = [
|
| 207 |
-
("deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "Lake 1 Flash", "released"),
|
| 208 |
("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
|
| 209 |
("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
|
| 210 |
("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
|
|
@@ -215,15 +220,6 @@ model_registry = [
|
|
| 215 |
|
| 216 |
# Model information
|
| 217 |
model_info = {
|
| 218 |
-
"Lake 1 Flash": {
|
| 219 |
-
"description": "English only, fast and lightweight model optimized for quick responses",
|
| 220 |
-
"parameters": "1.5B",
|
| 221 |
-
"training_data": "Diverse internet text + curated datasets",
|
| 222 |
-
"developer": "BI Corp",
|
| 223 |
-
"best_for": "Quick queries, mobile applications",
|
| 224 |
-
"architecture": "Transformer-based",
|
| 225 |
-
"context_window": "4096 tokens"
|
| 226 |
-
},
|
| 227 |
"Lake 1 Base": {
|
| 228 |
"description": "Balanced model offering good performance across tasks",
|
| 229 |
"parameters": "7B",
|
|
@@ -259,6 +255,24 @@ model_info = {
|
|
| 259 |
"best_for": "Research & advanced applications",
|
| 260 |
"architecture": "Mixture of Experts",
|
| 261 |
"context_window": "16384 tokens"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
}
|
| 263 |
}
|
| 264 |
|
|
|
|
| 14 |
"meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
|
| 15 |
model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
|
| 16 |
),
|
| 17 |
+
"mistralai/Mistral-7B-Instruct-v0.3": InferenceClient(
|
| 18 |
+
model="mistralai/Mistral-7B-Instruct-v0.3", token=hf_token
|
| 19 |
),
|
| 20 |
"meta-llama/Llama-2-7b-chat-hf": InferenceClient(
|
| 21 |
model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
|
|
|
|
| 27 |
|
| 28 |
# Model presets
|
| 29 |
presets = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
"meta-llama/Llama-3.2-3B-Instruct": {
|
| 31 |
+
"Fast": {"max_new_tokens": 1024, "temperature": 1.0, "top_p": 0.9},
|
| 32 |
+
"Normal": {"max_new_tokens": 2048, "temperature": 0.7, "top_p": 0.95},
|
| 33 |
+
"Quality": {"max_new_tokens": 4096, "temperature": 0.5, "top_p": 0.90},
|
| 34 |
+
"Unreal Performance": {"max_new_tokens": 8192, "temperature": 0.6, "top_p": 0.75},
|
| 35 |
},
|
| 36 |
+
"mistralai/Mistral-7B-Instruct-v0.3": {
|
| 37 |
+
"Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
|
| 38 |
+
"Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
|
| 39 |
+
"Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
|
| 40 |
+
"Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
|
| 41 |
},
|
| 42 |
"meta-llama/Llama-2-7b-chat-hf": {
|
| 43 |
+
"Fast": {"max_new_tokens": 512, "temperature": 1.0, "top_p": 0.9},
|
| 44 |
+
"Normal": {"max_new_tokens": 1024, "temperature": 0.7, "top_p": 0.95},
|
| 45 |
+
"Quality": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.90},
|
| 46 |
+
"Unreal Performance": {"max_new_tokens": 4096, "temperature": 0.6, "top_p": 0.75},
|
| 47 |
},
|
| 48 |
"meta-llama/Llama-3.3-70B-Instruct": {
|
| 49 |
+
"Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
|
| 50 |
+
"Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
|
| 51 |
+
"Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
|
| 52 |
+
"Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
|
| 53 |
+
},
|
| 54 |
+
"meta-llama/Llama-2-13b-chat-hf": {
|
| 55 |
+
"Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
|
| 56 |
+
"Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
|
| 57 |
+
"Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
|
| 58 |
+
"Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
|
| 59 |
+
},
|
| 60 |
+
"deepseek-ai/deepseek-llm-67b-chat": {
|
| 61 |
+
"Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
|
| 62 |
+
"Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
|
| 63 |
+
"Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
|
| 64 |
+
"Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
|
| 65 |
}
|
| 66 |
}
|
| 67 |
|
|
|
|
| 210 |
|
| 211 |
# Model registry
|
| 212 |
model_registry = [
|
|
|
|
| 213 |
("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
|
| 214 |
("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
|
| 215 |
("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
|
|
|
|
| 220 |
|
| 221 |
# Model information
|
| 222 |
model_info = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 223 |
"Lake 1 Base": {
|
| 224 |
"description": "Balanced model offering good performance across tasks",
|
| 225 |
"parameters": "7B",
|
|
|
|
| 255 |
"best_for": "Research & advanced applications",
|
| 256 |
"architecture": "Mixture of Experts",
|
| 257 |
"context_window": "16384 tokens"
|
| 258 |
+
},
|
| 259 |
+
"Lake 2 Advanced": {
|
| 260 |
+
"description": "Enhanced performance model with advanced reasoning capabilities.",
|
| 261 |
+
"parameters": "13B",
|
| 262 |
+
"training_data": "Diverse datasets including tech, literature, and general knowledge.",
|
| 263 |
+
"developer": "BI Corp",
|
| 264 |
+
"best_for": "Handling large-scale queries and in-depth topics.",
|
| 265 |
+
"architecture": "Transformer-based with specialized learning capabilities.",
|
| 266 |
+
"context_window": "16384 tokens"
|
| 267 |
+
},
|
| 268 |
+
"Lake 2 Pro [Planned]": {
|
| 269 |
+
"description": "Pro-level model with deep learning architecture for high-end applications.",
|
| 270 |
+
"parameters": "67B",
|
| 271 |
+
"training_data": "Expert-level data across multiple industries.",
|
| 272 |
+
"developer": "BI Corp",
|
| 273 |
+
"best_for": "High-performance computing and enterprise-level tasks.",
|
| 274 |
+
"architecture": "Hybrid architecture leveraging the latest advances in deep learning.",
|
| 275 |
+
"context_window": "32768 tokens"
|
| 276 |
}
|
| 277 |
}
|
| 278 |
|