Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -12,13 +12,13 @@ os.environ["PYTORCH_CUDA_ALLOC_CONF"] = "expandable_segments:True"
|
|
12 |
accelerator = Accelerator()
|
13 |
|
14 |
gpt2_pipeline = accelerator.prepare(
|
15 |
-
pipeline("text-generation", model="Qwen/Qwen-1_8B-Chat", device=accelerator.device)
|
16 |
)
|
17 |
Najeb_pipeline = accelerator.prepare(
|
18 |
-
pipeline("text-generation", model="najeebjust/Najeeb", device=accelerator.device)
|
19 |
)
|
20 |
llama2_pipeline = accelerator.prepare(
|
21 |
-
pipeline("text-generation", model="Harikrishnan46624/finetuned_llama2-1.1b-chat", device=accelerator.device)
|
22 |
)
|
23 |
'''
|
24 |
gpt2_pipeline = pipeline("text-generation", model="Qwen/Qwen-1_8B-Chat", device=0 if torch.cuda.is_available() else -1, trust_remote_code=True)
|
|
|
12 |
accelerator = Accelerator()
|
13 |
|
14 |
gpt2_pipeline = accelerator.prepare(
|
15 |
+
pipeline("text-generation", model="Qwen/Qwen-1_8B-Chat", device=accelerator.device , trust_remote_code=True)
|
16 |
)
|
17 |
Najeb_pipeline = accelerator.prepare(
|
18 |
+
pipeline("text-generation", model="najeebjust/Najeeb", device=accelerator.device , trust_remote_code=True)
|
19 |
)
|
20 |
llama2_pipeline = accelerator.prepare(
|
21 |
+
pipeline("text-generation", model="Harikrishnan46624/finetuned_llama2-1.1b-chat", device=accelerator.device , trust_remote_code=True)
|
22 |
)
|
23 |
'''
|
24 |
gpt2_pipeline = pipeline("text-generation", model="Qwen/Qwen-1_8B-Chat", device=0 if torch.cuda.is_available() else -1, trust_remote_code=True)
|