hoshingakag commited on
Commit
5478ed1
·
verified ·
1 Parent(s): 9153d1a

Update src/llamaindex_backend.py

Browse files
Files changed (1) hide show
  1. src/llamaindex_backend.py +4 -5
src/llamaindex_backend.py CHANGED
@@ -63,18 +63,17 @@ class GLlamaIndex():
63
  result = await self.index.retrieve_context(query)
64
  return result["result"]
65
 
66
- extended_query = f"<chat history>]\n{history[-1]}\n</chat history><new query>\n{query}\n</new query>"
67
- print(f"history[-1]: {history[-1]}")
68
- print(f"history[:-1]: {history[:-1]}")
69
  results = await self.index.aretrieve_context_multi(
70
  [query, extended_query]
71
  )
72
- print("retrieval results", results)
73
  eval_results = await self.aevaluate_context_multi(
74
  [query, extended_query],
75
  [r["result"] for r in results]
76
  )
77
- print("eval results", eval_results)
78
  return results[0]["result"] if eval_results[0].score > eval_results[1].score \
79
  else results[1]["result"]
80
 
 
63
  result = await self.index.retrieve_context(query)
64
  return result["result"]
65
 
66
+ extended_query = f"{format_history(history[-2:])}\n{query}"
67
+
 
68
  results = await self.index.aretrieve_context_multi(
69
  [query, extended_query]
70
  )
71
+ self.logger.info(f"retrieval results: {results}")
72
  eval_results = await self.aevaluate_context_multi(
73
  [query, extended_query],
74
  [r["result"] for r in results]
75
  )
76
+ self.logger.info(f"eval results: {eval_results}")
77
  return results[0]["result"] if eval_results[0].score > eval_results[1].score \
78
  else results[1]["result"]
79