awacke1 commited on
Commit
83b1b2c
1 Parent(s): 59b8b42

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -6
app.py CHANGED
@@ -121,7 +121,6 @@ def read_file_content(file,max_length):
121
  else:
122
  return ""
123
 
124
-
125
  def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
126
  model = model_choice
127
  conversation = [{'role': 'system', 'content': 'You are a helpful assistant.'}]
@@ -129,7 +128,6 @@ def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
129
  if len(document_section)>0:
130
  conversation.append({'role': 'assistant', 'content': document_section})
131
 
132
-
133
  # iterate through the stream of events
134
  start_time = time.time()
135
  response = openai.ChatCompletion.create(
@@ -147,13 +145,11 @@ def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
147
  collected_messages.append(chunk_message) # save the message
148
  content=chunk["choices"][0].get("delta",{}).get("content")
149
  st.write(f'*{content}*')
150
- st.markdown(f"Full response received {chunk_time:.2f} seconds after request")
151
  full_reply_content = ''.join([m.get('content', '') for m in collected_messages])
152
- st.markdown(f"Full conversation received: {full_reply_content}")
153
  return full_reply_content
154
 
155
-
156
-
157
  def chat_with_file_contents(prompt, file_content, model_choice='gpt-3.5-turbo'):
158
  conversation = [{'role': 'system', 'content': 'You are a helpful assistant.'}]
159
  conversation.append({'role': 'user', 'content': prompt})
 
121
  else:
122
  return ""
123
 
 
124
  def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
125
  model = model_choice
126
  conversation = [{'role': 'system', 'content': 'You are a helpful assistant.'}]
 
128
  if len(document_section)>0:
129
  conversation.append({'role': 'assistant', 'content': document_section})
130
 
 
131
  # iterate through the stream of events
132
  start_time = time.time()
133
  response = openai.ChatCompletion.create(
 
145
  collected_messages.append(chunk_message) # save the message
146
  content=chunk["choices"][0].get("delta",{}).get("content")
147
  st.write(f'*{content}*')
148
+ st.write(f"Full response received {chunk_time:.2f} seconds after request")
149
  full_reply_content = ''.join([m.get('content', '') for m in collected_messages])
150
+ st.write(f"Full conversation received: {full_reply_content}")
151
  return full_reply_content
152
 
 
 
153
  def chat_with_file_contents(prompt, file_content, model_choice='gpt-3.5-turbo'):
154
  conversation = [{'role': 'system', 'content': 'You are a helpful assistant.'}]
155
  conversation.append({'role': 'user', 'content': prompt})