Gregor Betz commited on
Commit
39d02de
1 Parent(s): bb6b48a
Files changed (1) hide show
  1. app.py +10 -10
app.py CHANGED
@@ -88,9 +88,9 @@ CHATBOT_INSTRUCTIONS = (
88
  with open("config.yaml") as stream:
89
  try:
90
  demo_config = yaml.safe_load(stream)
91
- logging.getLogger(__name__).info(f"Config: {demo_config}")
92
  except yaml.YAMLError as exc:
93
- logging.getLogger(__name__).error(f"Error loading config: {exc}")
94
  gr.Error("Error loading config: {exc}")
95
 
96
  try:
@@ -111,10 +111,10 @@ try:
111
  duration=-1
112
  )
113
  except Exception as exc:
114
- logging.getLogger(__name__).error(f"Error processing config: {exc}")
115
  gr.Error(f"Error processing config: {exc}")
116
 
117
- logging.getLogger(__name__).info(f"Reasoning guide expert model is {guide_kwargs['expert_model']}.")
118
 
119
 
120
 
@@ -128,7 +128,7 @@ def setup_client_llm(**client_kwargs) -> LogitsModel | None:
128
  try:
129
  llm = create_logits_model(**client_kwargs)
130
  except Exception as e:
131
- logging.getLogger(__name__).error(f"When setting up client llm: Error: {e}")
132
  return False
133
  return llm
134
 
@@ -162,13 +162,14 @@ async def bot(
162
  if len(history_langchain_format) <= 1:
163
 
164
  # health check
 
165
  health_check = await guide.health_check()
166
 
167
  if health_check.get("status", None) != "ok":
168
  health_msg = " | ".join([f"{k}: {v}" for k, v in health_check.items()])
169
- logging.getLogger(__name__).error(f"Guide health check failed: {health_msg}")
170
  gr.Error(f"LLM availability / health check failed: {health_msg}")
171
- logging.getLogger(__name__).info(f"Health check: {health_check}")
172
 
173
  message = history[-1][0]
174
 
@@ -176,9 +177,9 @@ async def bot(
176
  artifacts = {}
177
  progress_step = 0
178
  async for otype, ovalue in guide.guide(message):
179
- logging.getLogger(__name__).info(f"Guide output: {otype.value} - {ovalue}")
180
  if otype.value == "progress":
181
- logging.getLogger(__name__).info(f"Progress: {ovalue}")
182
  gr.Info(ovalue, duration=12)
183
  progress((progress_step,4))
184
  progress_step += 1
@@ -186,7 +187,6 @@ async def bot(
186
  artifacts[otype.value] = ovalue
187
  else:
188
  break
189
- asyncio.sleep(0.1)
190
  except asyncio.TimeoutError:
191
  msg = "Guided reasoning process took too long. Please try again."
192
  raise gr.Error(msg)
 
88
  with open("config.yaml") as stream:
89
  try:
90
  demo_config = yaml.safe_load(stream)
91
+ logging.info(f"Config: {demo_config}")
92
  except yaml.YAMLError as exc:
93
+ logging.error(f"Error loading config: {exc}")
94
  gr.Error("Error loading config: {exc}")
95
 
96
  try:
 
111
  duration=-1
112
  )
113
  except Exception as exc:
114
+ logging.error(f"Error processing config: {exc}")
115
  gr.Error(f"Error processing config: {exc}")
116
 
117
+ logging.info(f"Reasoning guide expert model is {guide_kwargs['expert_model']}.")
118
 
119
 
120
 
 
128
  try:
129
  llm = create_logits_model(**client_kwargs)
130
  except Exception as e:
131
+ logging.error(f"When setting up client llm: Error: {e}")
132
  return False
133
  return llm
134
 
 
162
  if len(history_langchain_format) <= 1:
163
 
164
  # health check
165
+ gr.Info("Checking availability and health of inference endpoints ...", duration=6)
166
  health_check = await guide.health_check()
167
 
168
  if health_check.get("status", None) != "ok":
169
  health_msg = " | ".join([f"{k}: {v}" for k, v in health_check.items()])
170
+ logging.error(f"Guide health check failed: {health_msg}")
171
  gr.Error(f"LLM availability / health check failed: {health_msg}")
172
+ logging.info(f"Health check: {health_check}")
173
 
174
  message = history[-1][0]
175
 
 
177
  artifacts = {}
178
  progress_step = 0
179
  async for otype, ovalue in guide.guide(message):
180
+ logging.info(f"Guide output: {otype.value} - {ovalue}")
181
  if otype.value == "progress":
182
+ logging.info(f"Progress: {ovalue}")
183
  gr.Info(ovalue, duration=12)
184
  progress((progress_step,4))
185
  progress_step += 1
 
187
  artifacts[otype.value] = ovalue
188
  else:
189
  break
 
190
  except asyncio.TimeoutError:
191
  msg = "Guided reasoning process took too long. Please try again."
192
  raise gr.Error(msg)