Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -21,7 +21,7 @@ MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "2048"))
|
|
21 |
|
22 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
23 |
|
24 |
-
model_id = "
|
25 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
26 |
model = AutoModelForCausalLM.from_pretrained(
|
27 |
model_id,
|
@@ -130,7 +130,6 @@ chat_interface = gr.ChatInterface(
|
|
130 |
|
131 |
with gr.Blocks(css="style.css", fill_height=True) as demo:
|
132 |
gr.Markdown(DESCRIPTION)
|
133 |
-
gr.DuplicateButton(value="Duplicate Space for private use", elem_id="duplicate-button")
|
134 |
chat_interface.render()
|
135 |
|
136 |
if __name__ == "__main__":
|
|
|
21 |
|
22 |
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
|
23 |
|
24 |
+
model_id = "giustinod/TestLogica-AZService"
|
25 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
26 |
model = AutoModelForCausalLM.from_pretrained(
|
27 |
model_id,
|
|
|
130 |
|
131 |
with gr.Blocks(css="style.css", fill_height=True) as demo:
|
132 |
gr.Markdown(DESCRIPTION)
|
|
|
133 |
chat_interface.render()
|
134 |
|
135 |
if __name__ == "__main__":
|