Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -20,6 +20,7 @@ from sklearn.metrics.pairwise import cosine_similarity
|
|
20 |
from sklearn.feature_extraction.text import TfidfVectorizer
|
21 |
import nltk
|
22 |
from nltk.corpus import stopwords
|
|
|
23 |
|
24 |
nltk.download('stopwords')
|
25 |
|
@@ -234,7 +235,7 @@ global_data = {
|
|
234 |
model_configs = [
|
235 |
{"repo_id": "Hjgugugjhuhjggg/testing_semifinal-Q2_K-GGUF", "filename": "testing_semifinal-q2_k.gguf", "name": "testing"},
|
236 |
{"repo_id": "bartowski/Llama-3.2-3B-Instruct-uncensored-GGUF", "filename": "Llama-3.2-3B-Instruct-uncensored-Q2_K.gguf", "name": "Llama-3.2-3B-Instruct"},
|
237 |
-
{"repo_id": "Ffftdtd5dtft/Meta-Llama-3.1-70B
|
238 |
{"repo_id": "Hhhbvvkgh/Heidi-Llama-v4-Q2_K-GGUF", "filename": "heidi-llama-v4-q2_k.gguf", "name": "Heidi-Llama-V4"}
|
239 |
]
|
240 |
|
@@ -244,14 +245,24 @@ def normalize_input(input_text):
|
|
244 |
filtered_words = [word for word in words if word.lower() not in stop_words]
|
245 |
return " ".join(filtered_words)
|
246 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
247 |
async def generate_model_response(model, inputs):
|
248 |
try:
|
249 |
response = await model.generate(inputs)
|
250 |
return response
|
251 |
except Exception as e:
|
252 |
-
return ""
|
253 |
|
254 |
def get_best_response(responses):
|
|
|
|
|
255 |
return max(responses, key=lambda x: x['score'])
|
256 |
|
257 |
async def process_message(message):
|
@@ -287,6 +298,7 @@ def run_gradio():
|
|
287 |
iface.launch(server_port=7862, prevent_thread_lock=True)
|
288 |
|
289 |
if __name__ == "__main__":
|
|
|
290 |
Thread(target=run_uvicorn).start()
|
291 |
Thread(target=run_gradio).start()
|
292 |
asyncio.get_event_loop().run_forever()
|
|
|
20 |
from sklearn.feature_extraction.text import TfidfVectorizer
|
21 |
import nltk
|
22 |
from nltk.corpus import stopwords
|
23 |
+
from huggingface_hub import download_url
|
24 |
|
25 |
nltk.download('stopwords')
|
26 |
|
|
|
235 |
model_configs = [
|
236 |
{"repo_id": "Hjgugugjhuhjggg/testing_semifinal-Q2_K-GGUF", "filename": "testing_semifinal-q2_k.gguf", "name": "testing"},
|
237 |
{"repo_id": "bartowski/Llama-3.2-3B-Instruct-uncensored-GGUF", "filename": "Llama-3.2-3B-Instruct-uncensored-Q2_K.gguf", "name": "Llama-3.2-3B-Instruct"},
|
238 |
+
{"repo_id": "Ffftdtd5dtft/Meta-Llama-3.1-70B", "filename": "meta-llama-3.1-70B", "name": "Meta-Llama-3.1-70B"},
|
239 |
{"repo_id": "Hhhbvvkgh/Heidi-Llama-v4-Q2_K-GGUF", "filename": "heidi-llama-v4-q2_k.gguf", "name": "Heidi-Llama-V4"}
|
240 |
]
|
241 |
|
|
|
245 |
filtered_words = [word for word in words if word.lower() not in stop_words]
|
246 |
return " ".join(filtered_words)
|
247 |
|
248 |
+
async def load_models():
|
249 |
+
for model in model_configs:
|
250 |
+
model_path = os.path.join("models", model["filename"])
|
251 |
+
if not os.path.exists(model_path):
|
252 |
+
url = f"https://huggingface.co/{model['repo_id']}/resolve/main/{model['filename']}"
|
253 |
+
download_url(url, model_path)
|
254 |
+
global_data['models'][model['name']] = Llama(model_path)
|
255 |
+
|
256 |
async def generate_model_response(model, inputs):
|
257 |
try:
|
258 |
response = await model.generate(inputs)
|
259 |
return response
|
260 |
except Exception as e:
|
261 |
+
return {"error": str(e)}
|
262 |
|
263 |
def get_best_response(responses):
|
264 |
+
if not responses:
|
265 |
+
return {"error": "No valid responses from models."}
|
266 |
return max(responses, key=lambda x: x['score'])
|
267 |
|
268 |
async def process_message(message):
|
|
|
298 |
iface.launch(server_port=7862, prevent_thread_lock=True)
|
299 |
|
300 |
if __name__ == "__main__":
|
301 |
+
asyncio.run(load_models())
|
302 |
Thread(target=run_uvicorn).start()
|
303 |
Thread(target=run_gradio).start()
|
304 |
asyncio.get_event_loop().run_forever()
|