Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ from peft import PeftModel
|
|
11 |
|
12 |
# Constants
|
13 |
MAX_MAX_NEW_TOKENS = 2048
|
14 |
-
DEFAULT_MAX_NEW_TOKENS =
|
15 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
16 |
|
17 |
# Description and License Texts
|
@@ -30,7 +30,7 @@ if not torch.cuda.is_available():
|
|
30 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
31 |
|
32 |
# Model and Tokenizer Configuration
|
33 |
-
model_id = "meta-llama/Llama-2-7b-
|
34 |
bnb_config = BitsAndBytesConfig(
|
35 |
load_in_4bit=True,
|
36 |
bnb_4bit_use_double_quant=False,
|
|
|
11 |
|
12 |
# Constants
|
13 |
MAX_MAX_NEW_TOKENS = 2048
|
14 |
+
DEFAULT_MAX_NEW_TOKENS = 950
|
15 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
16 |
|
17 |
# Description and License Texts
|
|
|
30 |
DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
|
31 |
|
32 |
# Model and Tokenizer Configuration
|
33 |
+
model_id = "meta-llama/Llama-2-7b-hf"
|
34 |
bnb_config = BitsAndBytesConfig(
|
35 |
load_in_4bit=True,
|
36 |
bnb_4bit_use_double_quant=False,
|