Update app.py
Browse files
app.py
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
from PIL import Image, ImageDraw, ImageFont
|
2 |
import tempfile
|
3 |
import gradio as gr
|
4 |
-
from smolagents import CodeAgent, InferenceClientModel
|
5 |
from smolagents import DuckDuckGoSearchTool, Tool
|
6 |
from huggingface_hub import InferenceClient
|
7 |
from diffusers import DiffusionPipeline
|
@@ -129,7 +129,14 @@ image_generation_tool = Tool.from_space(
|
|
129 |
#image_generation_tool= TextToImageTool()
|
130 |
#image_generation_tool = WrappedTextToImageTool()
|
131 |
search_tool = DuckDuckGoSearchTool()
|
132 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
133 |
|
134 |
agent = CodeAgent(tools=[image_generation_tool, search_tool], model=llm_engine)
|
135 |
|
|
|
1 |
from PIL import Image, ImageDraw, ImageFont
|
2 |
import tempfile
|
3 |
import gradio as gr
|
4 |
+
from smolagents import CodeAgent, InferenceClientModel, TransformersModel
|
5 |
from smolagents import DuckDuckGoSearchTool, Tool
|
6 |
from huggingface_hub import InferenceClient
|
7 |
from diffusers import DiffusionPipeline
|
|
|
129 |
#image_generation_tool= TextToImageTool()
|
130 |
#image_generation_tool = WrappedTextToImageTool()
|
131 |
search_tool = DuckDuckGoSearchTool()
|
132 |
+
print('iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii')
|
133 |
+
#llm_engine = InferenceClientModel("Qwen/Qwen2.5-72B-Instruct")
|
134 |
+
llm_engine = TransformersModel(
|
135 |
+
model_id="Qwen/Qwen2.5-72B-Instruct",
|
136 |
+
device="cuda",
|
137 |
+
max_new_tokens=5000,
|
138 |
+
)
|
139 |
+
print('aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa')
|
140 |
|
141 |
agent = CodeAgent(tools=[image_generation_tool, search_tool], model=llm_engine)
|
142 |
|