gorkemgoknar commited on
Commit
a51d57b
1 Parent(s): e41429f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -158,7 +158,7 @@ from llama_cpp import Llama
158
  # set GPU_LAYERS to 15 if you have a 8GB GPU so both models can fit in
159
  # else 35 full layers + XTTS works fine on T4 16GB
160
  # 5gb per llm, 4gb XTTS -> full layers should fit T4 16GB , 2LLM + XTTS
161
- GPU_LAYERS=int(os.environ.get("GPU_LAYERS", 35))
162
 
163
  LLAMA_VERBOSE=False
164
  print("Running LLM Mistral")
 
158
  # set GPU_LAYERS to 15 if you have a 8GB GPU so both models can fit in
159
  # else 35 full layers + XTTS works fine on T4 16GB
160
  # 5gb per llm, 4gb XTTS -> full layers should fit T4 16GB , 2LLM + XTTS
161
+ GPU_LAYERS=int(os.environ.get("GPU_LAYERS", 30))
162
 
163
  LLAMA_VERBOSE=False
164
  print("Running LLM Mistral")