Update app.py
Browse files
app.py
CHANGED
@@ -14,8 +14,8 @@ clients = {
|
|
14 |
"meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
|
15 |
model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
|
16 |
),
|
17 |
-
"
|
18 |
-
model="
|
19 |
),
|
20 |
"meta-llama/Llama-2-7b-chat-hf": InferenceClient(
|
21 |
model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
|
@@ -27,35 +27,41 @@ clients = {
|
|
27 |
|
28 |
# Model presets
|
29 |
presets = {
|
30 |
-
"mistralai/Mistral-7B-Instruct-v0.3": {
|
31 |
-
"Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.8},
|
32 |
-
"Normal": {"max_new_tokens": 512, "temperature": 0.6, "top_p": 0.75},
|
33 |
-
"Quality": {"max_new_tokens": 1024, "temperature": 0.45, "top_p": 0.60},
|
34 |
-
"Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.7},
|
35 |
-
},
|
36 |
"meta-llama/Llama-3.2-3B-Instruct": {
|
37 |
-
"Fast": {"max_new_tokens":
|
38 |
-
"Normal": {"max_new_tokens":
|
39 |
-
"Quality": {"max_new_tokens":
|
40 |
-
"Unreal Performance": {"max_new_tokens":
|
41 |
},
|
42 |
-
"
|
43 |
-
"Fast": {"max_new_tokens":
|
44 |
-
"Normal": {"max_new_tokens":
|
45 |
-
"Quality": {"max_new_tokens":
|
46 |
-
"Unreal Performance": {"max_new_tokens":
|
47 |
},
|
48 |
"meta-llama/Llama-2-7b-chat-hf": {
|
49 |
-
"Fast": {"max_new_tokens":
|
50 |
-
"Normal": {"max_new_tokens":
|
51 |
-
"Quality": {"max_new_tokens":
|
52 |
-
"Unreal Performance": {"max_new_tokens":
|
53 |
},
|
54 |
"meta-llama/Llama-3.3-70B-Instruct": {
|
55 |
-
"Fast": {"max_new_tokens":
|
56 |
-
"Normal": {"max_new_tokens":
|
57 |
-
"Quality": {"max_new_tokens":
|
58 |
-
"Unreal Performance": {"max_new_tokens":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
}
|
60 |
}
|
61 |
|
@@ -204,7 +210,6 @@ system_messages = {
|
|
204 |
|
205 |
# Model registry
|
206 |
model_registry = [
|
207 |
-
("deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "Lake 1 Flash", "released"),
|
208 |
("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
|
209 |
("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
|
210 |
("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
|
@@ -215,15 +220,6 @@ model_registry = [
|
|
215 |
|
216 |
# Model information
|
217 |
model_info = {
|
218 |
-
"Lake 1 Flash": {
|
219 |
-
"description": "English only, fast and lightweight model optimized for quick responses",
|
220 |
-
"parameters": "1.5B",
|
221 |
-
"training_data": "Diverse internet text + curated datasets",
|
222 |
-
"developer": "BI Corp",
|
223 |
-
"best_for": "Quick queries, mobile applications",
|
224 |
-
"architecture": "Transformer-based",
|
225 |
-
"context_window": "4096 tokens"
|
226 |
-
},
|
227 |
"Lake 1 Base": {
|
228 |
"description": "Balanced model offering good performance across tasks",
|
229 |
"parameters": "7B",
|
@@ -259,6 +255,24 @@ model_info = {
|
|
259 |
"best_for": "Research & advanced applications",
|
260 |
"architecture": "Mixture of Experts",
|
261 |
"context_window": "16384 tokens"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
262 |
}
|
263 |
}
|
264 |
|
|
|
14 |
"meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
|
15 |
model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
|
16 |
),
|
17 |
+
"mistralai/Mistral-7B-Instruct-v0.3": InferenceClient(
|
18 |
+
model="mistralai/Mistral-7B-Instruct-v0.3", token=hf_token
|
19 |
),
|
20 |
"meta-llama/Llama-2-7b-chat-hf": InferenceClient(
|
21 |
model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
|
|
|
27 |
|
28 |
# Model presets
|
29 |
presets = {
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
"meta-llama/Llama-3.2-3B-Instruct": {
|
31 |
+
"Fast": {"max_new_tokens": 1024, "temperature": 1.0, "top_p": 0.9},
|
32 |
+
"Normal": {"max_new_tokens": 2048, "temperature": 0.7, "top_p": 0.95},
|
33 |
+
"Quality": {"max_new_tokens": 4096, "temperature": 0.5, "top_p": 0.90},
|
34 |
+
"Unreal Performance": {"max_new_tokens": 8192, "temperature": 0.6, "top_p": 0.75},
|
35 |
},
|
36 |
+
"mistralai/Mistral-7B-Instruct-v0.3": {
|
37 |
+
"Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
|
38 |
+
"Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
|
39 |
+
"Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
|
40 |
+
"Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
|
41 |
},
|
42 |
"meta-llama/Llama-2-7b-chat-hf": {
|
43 |
+
"Fast": {"max_new_tokens": 512, "temperature": 1.0, "top_p": 0.9},
|
44 |
+
"Normal": {"max_new_tokens": 1024, "temperature": 0.7, "top_p": 0.95},
|
45 |
+
"Quality": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.90},
|
46 |
+
"Unreal Performance": {"max_new_tokens": 4096, "temperature": 0.6, "top_p": 0.75},
|
47 |
},
|
48 |
"meta-llama/Llama-3.3-70B-Instruct": {
|
49 |
+
"Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
|
50 |
+
"Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
|
51 |
+
"Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
|
52 |
+
"Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
|
53 |
+
},
|
54 |
+
"meta-llama/Llama-2-13b-chat-hf": {
|
55 |
+
"Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
|
56 |
+
"Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
|
57 |
+
"Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
|
58 |
+
"Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
|
59 |
+
},
|
60 |
+
"deepseek-ai/deepseek-llm-67b-chat": {
|
61 |
+
"Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
|
62 |
+
"Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
|
63 |
+
"Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
|
64 |
+
"Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
|
65 |
}
|
66 |
}
|
67 |
|
|
|
210 |
|
211 |
# Model registry
|
212 |
model_registry = [
|
|
|
213 |
("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
|
214 |
("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
|
215 |
("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
|
|
|
220 |
|
221 |
# Model information
|
222 |
model_info = {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
223 |
"Lake 1 Base": {
|
224 |
"description": "Balanced model offering good performance across tasks",
|
225 |
"parameters": "7B",
|
|
|
255 |
"best_for": "Research & advanced applications",
|
256 |
"architecture": "Mixture of Experts",
|
257 |
"context_window": "16384 tokens"
|
258 |
+
},
|
259 |
+
"Lake 2 Advanced": {
|
260 |
+
"description": "Enhanced performance model with advanced reasoning capabilities.",
|
261 |
+
"parameters": "13B",
|
262 |
+
"training_data": "Diverse datasets including tech, literature, and general knowledge.",
|
263 |
+
"developer": "BI Corp",
|
264 |
+
"best_for": "Handling large-scale queries and in-depth topics.",
|
265 |
+
"architecture": "Transformer-based with specialized learning capabilities.",
|
266 |
+
"context_window": "16384 tokens"
|
267 |
+
},
|
268 |
+
"Lake 2 Pro [Planned]": {
|
269 |
+
"description": "Pro-level model with deep learning architecture for high-end applications.",
|
270 |
+
"parameters": "67B",
|
271 |
+
"training_data": "Expert-level data across multiple industries.",
|
272 |
+
"developer": "BI Corp",
|
273 |
+
"best_for": "High-performance computing and enterprise-level tasks.",
|
274 |
+
"architecture": "Hybrid architecture leveraging the latest advances in deep learning.",
|
275 |
+
"context_window": "32768 tokens"
|
276 |
}
|
277 |
}
|
278 |
|