BICORP commited on
Commit
4cc88f9
·
verified ·
1 Parent(s): c0ff298

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +49 -35
app.py CHANGED
@@ -14,8 +14,8 @@ clients = {
14
  "meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
15
  model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
16
  ),
17
- "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B": InferenceClient(
18
- model="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", token=hf_token
19
  ),
20
  "meta-llama/Llama-2-7b-chat-hf": InferenceClient(
21
  model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
@@ -27,35 +27,41 @@ clients = {
27
 
28
  # Model presets
29
  presets = {
30
- "mistralai/Mistral-7B-Instruct-v0.3": {
31
- "Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.8},
32
- "Normal": {"max_new_tokens": 512, "temperature": 0.6, "top_p": 0.75},
33
- "Quality": {"max_new_tokens": 1024, "temperature": 0.45, "top_p": 0.60},
34
- "Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.7},
35
- },
36
  "meta-llama/Llama-3.2-3B-Instruct": {
37
- "Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.9},
38
- "Normal": {"max_new_tokens": 512, "temperature": 0.7, "top_p": 0.95},
39
- "Quality": {"max_new_tokens": 1024, "temperature": 0.5, "top_p": 0.90},
40
- "Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.6, "top_p": 0.75},
41
  },
42
- "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B": {
43
- "Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.9},
44
- "Normal": {"max_new_tokens": 512, "temperature": 0.7, "top_p": 0.95},
45
- "Quality": {"max_new_tokens": 1024, "temperature": 0.5, "top_p": 0.90},
46
- "Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.6, "top_p": 0.75},
47
  },
48
  "meta-llama/Llama-2-7b-chat-hf": {
49
- "Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.9},
50
- "Normal": {"max_new_tokens": 512, "temperature": 0.7, "top_p": 0.95},
51
- "Quality": {"max_new_tokens": 1024, "temperature": 0.5, "top_p": 0.90},
52
- "Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.6, "top_p": 0.75},
53
  },
54
  "meta-llama/Llama-3.3-70B-Instruct": {
55
- "Fast": {"max_new_tokens": 256, "temperature": 1.0, "top_p": 0.9},
56
- "Normal": {"max_new_tokens": 512, "temperature": 0.7, "top_p": 0.95},
57
- "Quality": {"max_new_tokens": 1024, "temperature": 0.5, "top_p": 0.90},
58
- "Unreal Performance": {"max_new_tokens": 2048, "temperature": 0.6, "top_p": 0.75},
 
 
 
 
 
 
 
 
 
 
 
 
59
  }
60
  }
61
 
@@ -204,7 +210,6 @@ system_messages = {
204
 
205
  # Model registry
206
  model_registry = [
207
- ("deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "Lake 1 Flash", "released"),
208
  ("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
209
  ("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
210
  ("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
@@ -215,15 +220,6 @@ model_registry = [
215
 
216
  # Model information
217
  model_info = {
218
- "Lake 1 Flash": {
219
- "description": "English only, fast and lightweight model optimized for quick responses",
220
- "parameters": "1.5B",
221
- "training_data": "Diverse internet text + curated datasets",
222
- "developer": "BI Corp",
223
- "best_for": "Quick queries, mobile applications",
224
- "architecture": "Transformer-based",
225
- "context_window": "4096 tokens"
226
- },
227
  "Lake 1 Base": {
228
  "description": "Balanced model offering good performance across tasks",
229
  "parameters": "7B",
@@ -259,6 +255,24 @@ model_info = {
259
  "best_for": "Research & advanced applications",
260
  "architecture": "Mixture of Experts",
261
  "context_window": "16384 tokens"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
262
  }
263
  }
264
 
 
14
  "meta-llama/Llama-3.2-3B-Instruct": InferenceClient(
15
  model="meta-llama/Llama-3.2-3B-Instruct", token=hf_token
16
  ),
17
+ "mistralai/Mistral-7B-Instruct-v0.3": InferenceClient(
18
+ model="mistralai/Mistral-7B-Instruct-v0.3", token=hf_token
19
  ),
20
  "meta-llama/Llama-2-7b-chat-hf": InferenceClient(
21
  model="meta-llama/Llama-2-7b-chat-hf", token=hf_token
 
27
 
28
  # Model presets
29
  presets = {
 
 
 
 
 
 
30
  "meta-llama/Llama-3.2-3B-Instruct": {
31
+ "Fast": {"max_new_tokens": 1024, "temperature": 1.0, "top_p": 0.9},
32
+ "Normal": {"max_new_tokens": 2048, "temperature": 0.7, "top_p": 0.95},
33
+ "Quality": {"max_new_tokens": 4096, "temperature": 0.5, "top_p": 0.90},
34
+ "Unreal Performance": {"max_new_tokens": 8192, "temperature": 0.6, "top_p": 0.75},
35
  },
36
+ "mistralai/Mistral-7B-Instruct-v0.3": {
37
+ "Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
38
+ "Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
39
+ "Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
40
+ "Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
41
  },
42
  "meta-llama/Llama-2-7b-chat-hf": {
43
+ "Fast": {"max_new_tokens": 512, "temperature": 1.0, "top_p": 0.9},
44
+ "Normal": {"max_new_tokens": 1024, "temperature": 0.7, "top_p": 0.95},
45
+ "Quality": {"max_new_tokens": 2048, "temperature": 0.5, "top_p": 0.90},
46
+ "Unreal Performance": {"max_new_tokens": 4096, "temperature": 0.6, "top_p": 0.75},
47
  },
48
  "meta-llama/Llama-3.3-70B-Instruct": {
49
+ "Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
50
+ "Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
51
+ "Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
52
+ "Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
53
+ },
54
+ "meta-llama/Llama-2-13b-chat-hf": {
55
+ "Fast": {"max_new_tokens": 2048, "temperature": 1.0, "top_p": 0.9},
56
+ "Normal": {"max_new_tokens": 4096, "temperature": 0.7, "top_p": 0.95},
57
+ "Quality": {"max_new_tokens": 8192, "temperature": 0.5, "top_p": 0.90},
58
+ "Unreal Performance": {"max_new_tokens": 16384, "temperature": 0.6, "top_p": 0.75},
59
+ },
60
+ "deepseek-ai/deepseek-llm-67b-chat": {
61
+ "Fast": {"max_new_tokens": 4096, "temperature": 1.0, "top_p": 0.9},
62
+ "Normal": {"max_new_tokens": 8192, "temperature": 0.7, "top_p": 0.95},
63
+ "Quality": {"max_new_tokens": 16384, "temperature": 0.5, "top_p": 0.90},
64
+ "Unreal Performance": {"max_new_tokens": 32768, "temperature": 0.6, "top_p": 0.75},
65
  }
66
  }
67
 
 
210
 
211
  # Model registry
212
  model_registry = [
 
213
  ("mistralai/Mistral-7B-Instruct-v0.3", "Lake 1 Base", "released"),
214
  ("meta-llama/Llama-3.2-3B-Instruct", "Lake 1 Advanced", "released"),
215
  ("meta-llama/Llama-2-7b-chat-hf", "Lake 2 Chat [Closed Alpha]", "shown"),
 
220
 
221
  # Model information
222
  model_info = {
 
 
 
 
 
 
 
 
 
223
  "Lake 1 Base": {
224
  "description": "Balanced model offering good performance across tasks",
225
  "parameters": "7B",
 
255
  "best_for": "Research & advanced applications",
256
  "architecture": "Mixture of Experts",
257
  "context_window": "16384 tokens"
258
+ },
259
+ "Lake 2 Advanced": {
260
+ "description": "Enhanced performance model with advanced reasoning capabilities.",
261
+ "parameters": "13B",
262
+ "training_data": "Diverse datasets including tech, literature, and general knowledge.",
263
+ "developer": "BI Corp",
264
+ "best_for": "Handling large-scale queries and in-depth topics.",
265
+ "architecture": "Transformer-based with specialized learning capabilities.",
266
+ "context_window": "16384 tokens"
267
+ },
268
+ "Lake 2 Pro [Planned]": {
269
+ "description": "Pro-level model with deep learning architecture for high-end applications.",
270
+ "parameters": "67B",
271
+ "training_data": "Expert-level data across multiple industries.",
272
+ "developer": "BI Corp",
273
+ "best_for": "High-performance computing and enterprise-level tasks.",
274
+ "architecture": "Hybrid architecture leveraging the latest advances in deep learning.",
275
+ "context_window": "32768 tokens"
276
  }
277
  }
278