Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -27,14 +27,14 @@ from datetime import datetime
|
|
27 |
from openai import OpenAI
|
28 |
|
29 |
torch.backends.cuda.matmul.allow_tf32 = True
|
30 |
-
torch.backends.cuda.matmul.allow_bf16_reduced_precision_reduction =
|
31 |
torch.backends.cuda.matmul.allow_fp16_reduced_precision_reduction = False
|
32 |
torch.backends.cudnn.allow_tf32 = True
|
33 |
torch.backends.cudnn.deterministic = False
|
34 |
|
35 |
torch.backends.cudnn.benchmark = False
|
36 |
|
37 |
-
torch.set_float32_matmul_precision("
|
38 |
|
39 |
# Load Hugging Face token if needed
|
40 |
hf_token = os.getenv("HF_TOKEN")
|
|
|
27 |
from openai import OpenAI
|
28 |
|
29 |
torch.backends.cuda.matmul.allow_tf32 = True
|
30 |
+
torch.backends.cuda.matmul.allow_bf16_reduced_precision_reduction = True
|
31 |
torch.backends.cuda.matmul.allow_fp16_reduced_precision_reduction = False
|
32 |
torch.backends.cudnn.allow_tf32 = True
|
33 |
torch.backends.cudnn.deterministic = False
|
34 |
|
35 |
torch.backends.cudnn.benchmark = False
|
36 |
|
37 |
+
torch.set_float32_matmul_precision("high")
|
38 |
|
39 |
# Load Hugging Face token if needed
|
40 |
hf_token = os.getenv("HF_TOKEN")
|