Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,23 +1,35 @@
|
|
1 |
import torch
|
2 |
-
from
|
3 |
-
from
|
|
|
4 |
from threading import Thread
|
5 |
import gradio as gr
|
6 |
|
7 |
MODEL_PATH = "Ozaii/zephyr-bae"
|
|
|
8 |
|
9 |
print("Attempting to load Zephyr... Cross your fingers! 🤞")
|
10 |
|
11 |
try:
|
12 |
peft_config = PeftConfig.from_pretrained(MODEL_PATH)
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
)
|
18 |
-
|
19 |
-
|
20 |
-
tokenizer.pad_token = tokenizer.eos_token
|
21 |
print("Zephyr loaded successfully! Time to charm!")
|
22 |
except Exception as e:
|
23 |
print(f"Oops! Zephyr seems to be playing hide and seek. Error: {str(e)}")
|
|
|
1 |
import torch
|
2 |
+
from unsloth import FastLanguageModel
|
3 |
+
from peft import PeftConfig
|
4 |
+
from transformers import AutoTokenizer, TextIteratorStreamer
|
5 |
from threading import Thread
|
6 |
import gradio as gr
|
7 |
|
8 |
MODEL_PATH = "Ozaii/zephyr-bae"
|
9 |
+
max_seq_length = 2048
|
10 |
|
11 |
print("Attempting to load Zephyr... Cross your fingers! 🤞")
|
12 |
|
13 |
try:
|
14 |
peft_config = PeftConfig.from_pretrained(MODEL_PATH)
|
15 |
+
base_model_name = peft_config.base_model_name_or_path
|
16 |
+
print(f"Loading base model: {base_model_name}")
|
17 |
+
|
18 |
+
model, tokenizer = FastLanguageModel.from_pretrained(
|
19 |
+
model_name=base_model_name,
|
20 |
+
max_seq_length=max_seq_length,
|
21 |
+
dtype=None, # Auto-detect
|
22 |
+
load_in_4bit=True,
|
23 |
+
)
|
24 |
+
|
25 |
+
model = FastLanguageModel.get_peft_model(
|
26 |
+
model,
|
27 |
+
peft_config=peft_config,
|
28 |
+
adapter_name="default",
|
29 |
+
use_gradient_checkpointing=True,
|
30 |
)
|
31 |
+
|
32 |
+
FastLanguageModel.for_inference(model)
|
|
|
33 |
print("Zephyr loaded successfully! Time to charm!")
|
34 |
except Exception as e:
|
35 |
print(f"Oops! Zephyr seems to be playing hide and seek. Error: {str(e)}")
|