Update app.py
Browse fileschange model to ORLM
app.py
CHANGED
@@ -3,7 +3,7 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
3 |
import torch
|
4 |
import os
|
5 |
|
6 |
-
model_name = "
|
7 |
device_map = 'cuda'
|
8 |
|
9 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
@@ -47,8 +47,8 @@ gr.ChatInterface(
|
|
47 |
chat_function,
|
48 |
chatbot=gr.Chatbot(height=400),
|
49 |
textbox=gr.Textbox(placeholder="Enter message here", container=False, scale=7),
|
50 |
-
title="
|
51 |
-
description="""Chat with
|
52 |
theme="soft",
|
53 |
additional_inputs=[
|
54 |
gr.Textbox("You shall answer to all the questions as very smart AI", label="System Prompt"),
|
|
|
3 |
import torch
|
4 |
import os
|
5 |
|
6 |
+
model_name = "CardinalOperations/ORLM-LLaMA-3-8B"
|
7 |
device_map = 'cuda'
|
8 |
|
9 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
|
|
47 |
chat_function,
|
48 |
chatbot=gr.Chatbot(height=400),
|
49 |
textbox=gr.Textbox(placeholder="Enter message here", container=False, scale=7),
|
50 |
+
title="ORLM Chat",
|
51 |
+
description="""Chat with ORLM""",
|
52 |
theme="soft",
|
53 |
additional_inputs=[
|
54 |
gr.Textbox("You shall answer to all the questions as very smart AI", label="System Prompt"),
|