Update app.py
Browse files
app.py
CHANGED
@@ -1305,11 +1305,11 @@ def handsome_chat_completions():
|
|
1305 |
try:
|
1306 |
start_time = time.time()
|
1307 |
response = requests.post(
|
1308 |
-
|
1309 |
headers=headers,
|
1310 |
json=data,
|
1311 |
-
|
1312 |
-
|
1313 |
)
|
1314 |
|
1315 |
if response.status_code == 429:
|
@@ -1317,33 +1317,116 @@ def handsome_chat_completions():
|
|
1317 |
|
1318 |
if data.get("stream", False):
|
1319 |
def generate():
|
1320 |
-
|
1321 |
-
|
1322 |
-
|
1323 |
-
|
1324 |
-
|
1325 |
-
|
1326 |
-
|
1327 |
-
|
1328 |
-
|
1329 |
-
|
1330 |
-
|
1331 |
-
|
1332 |
-
|
1333 |
-
|
1334 |
-
|
1335 |
-
|
1336 |
-
|
1337 |
-
|
1338 |
-
|
1339 |
-
|
1340 |
-
|
1341 |
-
|
1342 |
-
]
|
1343 |
-
|
1344 |
-
|
1345 |
-
|
1346 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1347 |
else:
|
1348 |
response.raise_for_status()
|
1349 |
end_time = time.time()
|
@@ -1351,74 +1434,66 @@ def handsome_chat_completions():
|
|
1351 |
total_time = end_time - start_time
|
1352 |
|
1353 |
try:
|
1354 |
-
|
1355 |
-
|
1356 |
-
|
1357 |
-
|
1358 |
-
|
1359 |
-
|
1360 |
-
|
1361 |
-
"created": int(time.time()),
|
1362 |
-
"model": model_name,
|
1363 |
-
"choices": [
|
1364 |
-
{
|
1365 |
-
"index": 0,
|
1366 |
-
"message": {
|
1367 |
-
"role": "assistant",
|
1368 |
-
"content": content
|
1369 |
-
},
|
1370 |
-
"finish_reason": "stop"
|
1371 |
-
}
|
1372 |
-
],
|
1373 |
-
}
|
1374 |
-
else:
|
1375 |
-
response_data = {
|
1376 |
-
"id": f"chatcmpl-{uuid.uuid4()}",
|
1377 |
-
"object": "chat.completion",
|
1378 |
-
"created": int(time.time()),
|
1379 |
-
"model": model_name,
|
1380 |
-
"choices": [
|
1381 |
-
{
|
1382 |
-
"index": 0,
|
1383 |
-
"message": {
|
1384 |
-
"role": "assistant",
|
1385 |
-
"content": "No response content"
|
1386 |
-
},
|
1387 |
-
"finish_reason": "stop"
|
1388 |
-
}
|
1389 |
-
]
|
1390 |
-
}
|
1391 |
except (KeyError, ValueError, IndexError) as e:
|
1392 |
logging.error(
|
1393 |
-
f"
|
1394 |
f"完整内容: {response_json}"
|
1395 |
)
|
1396 |
-
|
1397 |
-
|
1398 |
-
|
1399 |
-
|
1400 |
-
|
1401 |
-
|
1402 |
-
|
1403 |
-
|
1404 |
-
|
1405 |
-
|
1406 |
-
|
1407 |
-
|
1408 |
-
|
1409 |
-
|
1410 |
-
|
1411 |
-
|
1412 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1413 |
logging.info(
|
1414 |
-
|
1415 |
-
|
1416 |
-
|
|
|
|
|
|
|
|
|
|
|
1417 |
)
|
1418 |
with data_lock:
|
1419 |
request_timestamps.append(time.time())
|
1420 |
-
|
1421 |
-
|
|
|
|
|
|
|
|
|
|
|
1422 |
except requests.exceptions.RequestException as e:
|
1423 |
logging.error(f"请求转发异常: {e}")
|
1424 |
return jsonify({"error": str(e)}), 500
|
|
|
1305 |
try:
|
1306 |
start_time = time.time()
|
1307 |
response = requests.post(
|
1308 |
+
TEST_MODEL_ENDPOINT,
|
1309 |
headers=headers,
|
1310 |
json=data,
|
1311 |
+
stream=data.get("stream", False),
|
1312 |
+
timeout=60
|
1313 |
)
|
1314 |
|
1315 |
if response.status_code == 429:
|
|
|
1317 |
|
1318 |
if data.get("stream", False):
|
1319 |
def generate():
|
1320 |
+
first_chunk_time = None
|
1321 |
+
full_response_content = ""
|
1322 |
+
for chunk in response.iter_content(chunk_size=1024):
|
1323 |
+
if chunk:
|
1324 |
+
if first_chunk_time is None:
|
1325 |
+
first_chunk_time = time.time()
|
1326 |
+
full_response_content += chunk.decode("utf-8")
|
1327 |
+
yield chunk
|
1328 |
+
|
1329 |
+
end_time = time.time()
|
1330 |
+
first_token_time = (
|
1331 |
+
first_chunk_time - start_time
|
1332 |
+
if first_chunk_time else 0
|
1333 |
+
)
|
1334 |
+
total_time = end_time - start_time
|
1335 |
+
|
1336 |
+
prompt_tokens = 0
|
1337 |
+
completion_tokens = 0
|
1338 |
+
response_content = ""
|
1339 |
+
for line in full_response_content.splitlines():
|
1340 |
+
if line.startswith("data:"):
|
1341 |
+
line = line[5:].strip()
|
1342 |
+
if line == "[DONE]":
|
1343 |
+
continue
|
1344 |
+
try:
|
1345 |
+
response_json = json.loads(line)
|
1346 |
+
|
1347 |
+
if (
|
1348 |
+
"usage" in response_json and
|
1349 |
+
"completion_tokens" in response_json["usage"]
|
1350 |
+
):
|
1351 |
+
completion_tokens = response_json[
|
1352 |
+
"usage"
|
1353 |
+
]["completion_tokens"]
|
1354 |
+
|
1355 |
+
if (
|
1356 |
+
"choices" in response_json and
|
1357 |
+
len(response_json["choices"]) > 0 and
|
1358 |
+
"delta" in response_json["choices"][0] and
|
1359 |
+
"content" in response_json[
|
1360 |
+
"choices"
|
1361 |
+
][0]["delta"]
|
1362 |
+
):
|
1363 |
+
response_content += response_json[
|
1364 |
+
"choices"
|
1365 |
+
][0]["delta"]["content"]
|
1366 |
+
|
1367 |
+
if (
|
1368 |
+
"usage" in response_json and
|
1369 |
+
"prompt_tokens" in response_json["usage"]
|
1370 |
+
):
|
1371 |
+
prompt_tokens = response_json[
|
1372 |
+
"usage"
|
1373 |
+
]["prompt_tokens"]
|
1374 |
+
|
1375 |
+
except (
|
1376 |
+
KeyError,
|
1377 |
+
ValueError,
|
1378 |
+
IndexError
|
1379 |
+
) as e:
|
1380 |
+
logging.error(
|
1381 |
+
f"解析流式响应单行 JSON 失败: {e}, "
|
1382 |
+
f"行内容: {line}"
|
1383 |
+
)
|
1384 |
+
|
1385 |
+
user_content = ""
|
1386 |
+
messages = data.get("messages", [])
|
1387 |
+
for message in messages:
|
1388 |
+
if message["role"] == "user":
|
1389 |
+
if isinstance(message["content"], str):
|
1390 |
+
user_content += message["content"] + " "
|
1391 |
+
elif isinstance(message["content"], list):
|
1392 |
+
for item in message["content"]:
|
1393 |
+
if (
|
1394 |
+
isinstance(item, dict) and
|
1395 |
+
item.get("type") == "text"
|
1396 |
+
):
|
1397 |
+
user_content += (
|
1398 |
+
item.get("text", "") +
|
1399 |
+
" "
|
1400 |
+
)
|
1401 |
+
|
1402 |
+
user_content = user_content.strip()
|
1403 |
+
|
1404 |
+
user_content_replaced = user_content.replace(
|
1405 |
+
'\n', '\\n'
|
1406 |
+
).replace('\r', '\\n')
|
1407 |
+
response_content_replaced = response_content.replace(
|
1408 |
+
'\n', '\\n'
|
1409 |
+
).replace('\r', '\\n')
|
1410 |
+
|
1411 |
+
logging.info(
|
1412 |
+
f"使用的key: {api_key}, "
|
1413 |
+
f"提示token: {prompt_tokens}, "
|
1414 |
+
f"输出token: {completion_tokens}, "
|
1415 |
+
f"首字用时: {first_token_time:.4f}秒, "
|
1416 |
+
f"总共用时: {total_time:.4f}秒, "
|
1417 |
+
f"使用的模型: {model_name}, "
|
1418 |
+
f"用户的内容: {user_content_replaced}, "
|
1419 |
+
f"输出的内容: {response_content_replaced}"
|
1420 |
+
)
|
1421 |
+
|
1422 |
+
with data_lock:
|
1423 |
+
request_timestamps.append(time.time())
|
1424 |
+
token_counts.append(prompt_tokens+completion_tokens)
|
1425 |
+
|
1426 |
+
return Response(
|
1427 |
+
stream_with_context(generate()),
|
1428 |
+
content_type=response.headers['Content-Type']
|
1429 |
+
)
|
1430 |
else:
|
1431 |
response.raise_for_status()
|
1432 |
end_time = time.time()
|
|
|
1434 |
total_time = end_time - start_time
|
1435 |
|
1436 |
try:
|
1437 |
+
prompt_tokens = response_json["usage"]["prompt_tokens"]
|
1438 |
+
completion_tokens = response_json[
|
1439 |
+
"usage"
|
1440 |
+
]["completion_tokens"]
|
1441 |
+
response_content = response_json[
|
1442 |
+
"choices"
|
1443 |
+
][0]["message"]["content"]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1444 |
except (KeyError, ValueError, IndexError) as e:
|
1445 |
logging.error(
|
1446 |
+
f"解析非流式响应 JSON 失败: {e}, "
|
1447 |
f"完整内容: {response_json}"
|
1448 |
)
|
1449 |
+
prompt_tokens = 0
|
1450 |
+
completion_tokens = 0
|
1451 |
+
response_content = ""
|
1452 |
+
|
1453 |
+
user_content = ""
|
1454 |
+
messages = data.get("messages", [])
|
1455 |
+
for message in messages:
|
1456 |
+
if message["role"] == "user":
|
1457 |
+
if isinstance(message["content"], str):
|
1458 |
+
user_content += message["content"] + " "
|
1459 |
+
elif isinstance(message["content"], list):
|
1460 |
+
for item in message["content"]:
|
1461 |
+
if (
|
1462 |
+
isinstance(item, dict) and
|
1463 |
+
item.get("type") == "text"
|
1464 |
+
):
|
1465 |
+
user_content += (
|
1466 |
+
item.get("text", "") + " "
|
1467 |
+
)
|
1468 |
+
|
1469 |
+
user_content = user_content.strip()
|
1470 |
+
|
1471 |
+
user_content_replaced = user_content.replace(
|
1472 |
+
'\n', '\\n'
|
1473 |
+
).replace('\r', '\\n')
|
1474 |
+
response_content_replaced = response_content.replace(
|
1475 |
+
'\n', '\\n'
|
1476 |
+
).replace('\r', '\\n')
|
1477 |
+
|
1478 |
logging.info(
|
1479 |
+
f"使用的key: {api_key}, "
|
1480 |
+
f"提示token: {prompt_tokens}, "
|
1481 |
+
f"输出token: {completion_tokens}, "
|
1482 |
+
f"首字用时: 0, "
|
1483 |
+
f"总共用时: {total_time:.4f}秒, "
|
1484 |
+
f"使用的模型: {model_name}, "
|
1485 |
+
f"用户的内容: {user_content_replaced}, "
|
1486 |
+
f"输出的内容: {response_content_replaced}"
|
1487 |
)
|
1488 |
with data_lock:
|
1489 |
request_timestamps.append(time.time())
|
1490 |
+
if "prompt_tokens" in response_json["usage"] and "completion_tokens" in response_json["usage"]:
|
1491 |
+
token_counts.append(response_json["usage"]["prompt_tokens"] + response_json["usage"]["completion_tokens"])
|
1492 |
+
else:
|
1493 |
+
token_counts.append(0)
|
1494 |
+
|
1495 |
+
return jsonify(response_json)
|
1496 |
+
|
1497 |
except requests.exceptions.RequestException as e:
|
1498 |
logging.error(f"请求转发异常: {e}")
|
1499 |
return jsonify({"error": str(e)}), 500
|