Hjgugugjhuhjggg commited on
Commit
3294219
·
verified ·
1 Parent(s): e279e64

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -2
app.py CHANGED
@@ -20,6 +20,7 @@ from sklearn.metrics.pairwise import cosine_similarity
20
  from sklearn.feature_extraction.text import TfidfVectorizer
21
  import nltk
22
  from nltk.corpus import stopwords
 
23
 
24
  nltk.download('stopwords')
25
 
@@ -234,7 +235,7 @@ global_data = {
234
  model_configs = [
235
  {"repo_id": "Hjgugugjhuhjggg/testing_semifinal-Q2_K-GGUF", "filename": "testing_semifinal-q2_k.gguf", "name": "testing"},
236
  {"repo_id": "bartowski/Llama-3.2-3B-Instruct-uncensored-GGUF", "filename": "Llama-3.2-3B-Instruct-uncensored-Q2_K.gguf", "name": "Llama-3.2-3B-Instruct"},
237
- {"repo_id": "Ffftdtd5dtft/Meta-Llama-3.1-70B-Q2_K-GGUF", "filename": "meta-llama-3.1-70b-q2_k.gguf", "name": "Meta-Llama-3.1-70B"},
238
  {"repo_id": "Hhhbvvkgh/Heidi-Llama-v4-Q2_K-GGUF", "filename": "heidi-llama-v4-q2_k.gguf", "name": "Heidi-Llama-V4"}
239
  ]
240
 
@@ -244,14 +245,24 @@ def normalize_input(input_text):
244
  filtered_words = [word for word in words if word.lower() not in stop_words]
245
  return " ".join(filtered_words)
246
 
 
 
 
 
 
 
 
 
247
  async def generate_model_response(model, inputs):
248
  try:
249
  response = await model.generate(inputs)
250
  return response
251
  except Exception as e:
252
- return ""
253
 
254
  def get_best_response(responses):
 
 
255
  return max(responses, key=lambda x: x['score'])
256
 
257
  async def process_message(message):
@@ -287,6 +298,7 @@ def run_gradio():
287
  iface.launch(server_port=7862, prevent_thread_lock=True)
288
 
289
  if __name__ == "__main__":
 
290
  Thread(target=run_uvicorn).start()
291
  Thread(target=run_gradio).start()
292
  asyncio.get_event_loop().run_forever()
 
20
  from sklearn.feature_extraction.text import TfidfVectorizer
21
  import nltk
22
  from nltk.corpus import stopwords
23
+ from huggingface_hub import download_url
24
 
25
  nltk.download('stopwords')
26
 
 
235
  model_configs = [
236
  {"repo_id": "Hjgugugjhuhjggg/testing_semifinal-Q2_K-GGUF", "filename": "testing_semifinal-q2_k.gguf", "name": "testing"},
237
  {"repo_id": "bartowski/Llama-3.2-3B-Instruct-uncensored-GGUF", "filename": "Llama-3.2-3B-Instruct-uncensored-Q2_K.gguf", "name": "Llama-3.2-3B-Instruct"},
238
+ {"repo_id": "Ffftdtd5dtft/Meta-Llama-3.1-70B", "filename": "meta-llama-3.1-70B", "name": "Meta-Llama-3.1-70B"},
239
  {"repo_id": "Hhhbvvkgh/Heidi-Llama-v4-Q2_K-GGUF", "filename": "heidi-llama-v4-q2_k.gguf", "name": "Heidi-Llama-V4"}
240
  ]
241
 
 
245
  filtered_words = [word for word in words if word.lower() not in stop_words]
246
  return " ".join(filtered_words)
247
 
248
+ async def load_models():
249
+ for model in model_configs:
250
+ model_path = os.path.join("models", model["filename"])
251
+ if not os.path.exists(model_path):
252
+ url = f"https://huggingface.co/{model['repo_id']}/resolve/main/{model['filename']}"
253
+ download_url(url, model_path)
254
+ global_data['models'][model['name']] = Llama(model_path)
255
+
256
  async def generate_model_response(model, inputs):
257
  try:
258
  response = await model.generate(inputs)
259
  return response
260
  except Exception as e:
261
+ return {"error": str(e)}
262
 
263
  def get_best_response(responses):
264
+ if not responses:
265
+ return {"error": "No valid responses from models."}
266
  return max(responses, key=lambda x: x['score'])
267
 
268
  async def process_message(message):
 
298
  iface.launch(server_port=7862, prevent_thread_lock=True)
299
 
300
  if __name__ == "__main__":
301
+ asyncio.run(load_models())
302
  Thread(target=run_uvicorn).start()
303
  Thread(target=run_gradio).start()
304
  asyncio.get_event_loop().run_forever()