Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
-
import spaces
|
2 |
import torch
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM, TextIteratorStreamer
|
4 |
from peft import PeftConfig, PeftModel
|
5 |
from threading import Thread
|
6 |
import gradio as gr
|
|
|
7 |
|
8 |
MODEL_PATH = "Ozaii/zephyr-bae"
|
9 |
-
BASE_MODEL = "unsloth/
|
10 |
max_seq_length = 2048
|
11 |
|
12 |
print("Zephyr is getting ready to charm! π")
|
@@ -25,7 +25,8 @@ def load_model():
|
|
25 |
BASE_MODEL,
|
26 |
torch_dtype=torch.float16,
|
27 |
device_map="auto",
|
28 |
-
load_in_4bit=True
|
|
|
29 |
)
|
30 |
|
31 |
model = PeftModel.from_pretrained(base_model, MODEL_PATH)
|
|
|
|
|
1 |
import torch
|
2 |
from transformers import AutoTokenizer, AutoModelForCausalLM, TextIteratorStreamer
|
3 |
from peft import PeftConfig, PeftModel
|
4 |
from threading import Thread
|
5 |
import gradio as gr
|
6 |
+
import spaces
|
7 |
|
8 |
MODEL_PATH = "Ozaii/zephyr-bae"
|
9 |
+
BASE_MODEL = "unsloth/llama-3-8b-bnb-4bit"
|
10 |
max_seq_length = 2048
|
11 |
|
12 |
print("Zephyr is getting ready to charm! π")
|
|
|
25 |
BASE_MODEL,
|
26 |
torch_dtype=torch.float16,
|
27 |
device_map="auto",
|
28 |
+
load_in_4bit=True,
|
29 |
+
trust_remote_code=True
|
30 |
)
|
31 |
|
32 |
model = PeftModel.from_pretrained(base_model, MODEL_PATH)
|