Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -37,8 +37,8 @@ def terminate():
|
|
37 |
|
38 |
import ollama
|
39 |
import gradio as gr
|
40 |
-
from ollama import
|
41 |
-
client =
|
42 |
|
43 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
44 |
|
@@ -115,7 +115,7 @@ async def stream_chat(message: str, history: list, model: str, temperature: floa
|
|
115 |
|
116 |
print(f"Conversation is -\n{conversation}")
|
117 |
|
118 |
-
|
119 |
model=model,
|
120 |
stream=True,
|
121 |
messages=conversation,
|
@@ -128,14 +128,8 @@ async def stream_chat(message: str, history: list, model: str, temperature: floa
|
|
128 |
'repeat_penalty': penalty,
|
129 |
'low_vram': True,
|
130 |
},
|
131 |
-
|
132 |
-
|
133 |
-
print(response)
|
134 |
-
|
135 |
-
buffer = ""
|
136 |
-
for chunk in response:
|
137 |
-
buffer += chunk["message"]["content"]
|
138 |
-
yield buffer
|
139 |
|
140 |
|
141 |
async def main(message: str, history: list, model: str, temperature: float, max_new_tokens: int, top_p: float, top_k: int, penalty: float):
|
@@ -145,7 +139,7 @@ async def main(message: str, history: list, model: str, temperature: float, max_
|
|
145 |
else:
|
146 |
if not INIT_SIGN:
|
147 |
yield "Please initialize Ollama"
|
148 |
-
else:
|
149 |
async for response in stream_chat(
|
150 |
message,
|
151 |
history,
|
|
|
37 |
|
38 |
import ollama
|
39 |
import gradio as gr
|
40 |
+
from ollama import AsyncClient
|
41 |
+
client = AsyncClient(host='http://localhost:11434', timeout=120)
|
42 |
|
43 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
44 |
|
|
|
115 |
|
116 |
print(f"Conversation is -\n{conversation}")
|
117 |
|
118 |
+
async for part in await client.chat(
|
119 |
model=model,
|
120 |
stream=True,
|
121 |
messages=conversation,
|
|
|
128 |
'repeat_penalty': penalty,
|
129 |
'low_vram': True,
|
130 |
},
|
131 |
+
):
|
132 |
+
yield part['message']['content']
|
|
|
|
|
|
|
|
|
|
|
|
|
133 |
|
134 |
|
135 |
async def main(message: str, history: list, model: str, temperature: float, max_new_tokens: int, top_p: float, top_k: int, penalty: float):
|
|
|
139 |
else:
|
140 |
if not INIT_SIGN:
|
141 |
yield "Please initialize Ollama"
|
142 |
+
else:
|
143 |
async for response in stream_chat(
|
144 |
message,
|
145 |
history,
|