Update app.py
Browse files
app.py
CHANGED
@@ -425,7 +425,7 @@ def handsome_chat_completions():
|
|
425 |
|
426 |
if delta.get("content") is not None:
|
427 |
if not first_reasoning_chunk:
|
428 |
-
yield f"data: {json.dumps({'choices': [{'delta': {'content': '\n
|
429 |
first_reasoning_chunk = True
|
430 |
yield f"data: {json.dumps({'choices': [{'delta': {'content': delta["content"]}, 'index': 0}]})}\n\n"
|
431 |
|
@@ -525,7 +525,6 @@ def handsome_chat_completions():
|
|
525 |
content_type="text/event-stream"
|
526 |
)
|
527 |
else:
|
528 |
-
# ... (Non-streaming part remains the same as in the previous response)
|
529 |
response.raise_for_status()
|
530 |
end_time = time.time()
|
531 |
response_json = response.json()
|
@@ -597,7 +596,6 @@ def handsome_chat_completions():
|
|
597 |
request_timestamps.append(time.time())
|
598 |
token_counts.append(prompt_tokens + completion_tokens)
|
599 |
|
600 |
-
# Reformat the response to standard OpenAI format for non-streaming responses
|
601 |
formatted_response = {
|
602 |
"id": response_json.get("id", ""),
|
603 |
"object": "chat.completion",
|
|
|
425 |
|
426 |
if delta.get("content") is not None:
|
427 |
if not first_reasoning_chunk:
|
428 |
+
yield f"data: {json.dumps({'choices': [{'delta': {'content': '\n```\n'}, 'index': 0}]})}\n\n"
|
429 |
first_reasoning_chunk = True
|
430 |
yield f"data: {json.dumps({'choices': [{'delta': {'content': delta["content"]}, 'index': 0}]})}\n\n"
|
431 |
|
|
|
525 |
content_type="text/event-stream"
|
526 |
)
|
527 |
else:
|
|
|
528 |
response.raise_for_status()
|
529 |
end_time = time.time()
|
530 |
response_json = response.json()
|
|
|
596 |
request_timestamps.append(time.time())
|
597 |
token_counts.append(prompt_tokens + completion_tokens)
|
598 |
|
|
|
599 |
formatted_response = {
|
600 |
"id": response_json.get("id", ""),
|
601 |
"object": "chat.completion",
|