Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -12,12 +12,14 @@ from transformers import (
|
|
12 |
TextIteratorStreamer,
|
13 |
)
|
14 |
|
15 |
-
|
|
|
|
|
16 |
# ORLM LLaMA-3-8B
|
17 |
|
18 |
Hello! I'm ORLM-LLaMA-3-8B, here to automate your optimization modeling tasks! Check our [repo](https://github.com/Cardinal-Operations/ORLM) and [paper](https://arxiv.org/abs/2405.17743)!
|
19 |
|
20 |
-
Please note that solution generation may be terminated if it exceeds
|
21 |
|
22 |
If the demo successfully generates a code solution, execute it in your Python environment with `coptpy` installed to obtain the final optimal value for your task.
|
23 |
"""
|
@@ -47,7 +49,7 @@ Below is an operations research question. Build a mathematical model and corresp
|
|
47 |
# Response:
|
48 |
"""
|
49 |
|
50 |
-
@spaces.GPU(duration=
|
51 |
def generate(
|
52 |
message: str,
|
53 |
chat_history: list[tuple[str, str]],
|
|
|
12 |
TextIteratorStreamer,
|
13 |
)
|
14 |
|
15 |
+
GENERATION_TIME=90
|
16 |
+
|
17 |
+
DESCRIPTION = f"""\
|
18 |
# ORLM LLaMA-3-8B
|
19 |
|
20 |
Hello! I'm ORLM-LLaMA-3-8B, here to automate your optimization modeling tasks! Check our [repo](https://github.com/Cardinal-Operations/ORLM) and [paper](https://arxiv.org/abs/2405.17743)!
|
21 |
|
22 |
+
Please note that solution generation may be terminated if it exceeds {GENERATION_TIME} seconds. We strongly recommend running the demo locally using our [sample script](https://github.com/Cardinal-Operations/ORLM/blob/master/scripts/inference.py) for a smoother experience.
|
23 |
|
24 |
If the demo successfully generates a code solution, execute it in your Python environment with `coptpy` installed to obtain the final optimal value for your task.
|
25 |
"""
|
|
|
49 |
# Response:
|
50 |
"""
|
51 |
|
52 |
+
@spaces.GPU(duration=GENERATION_TIME)
|
53 |
def generate(
|
54 |
message: str,
|
55 |
chat_history: list[tuple[str, str]],
|