Update app.py
Browse files
app.py
CHANGED
@@ -1,11 +1,10 @@
|
|
1 |
import torch
|
2 |
-
from transformers import
|
3 |
|
4 |
model_name = 'prithivida/parrot_paraphraser_on_T5'
|
5 |
torch_device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
6 |
-
tokenizer =
|
7 |
-
model = AutoModelForSeq2SeqLM.from_pretrained(
|
8 |
-
|
9 |
def get_response(input_text,num_return_sequences):
|
10 |
batch = tokenizer.prepare_seq2seq_batch([input_text],truncation=True,padding='longest',max_length=60, return_tensors="pt").to(torch_device)
|
11 |
translated = model.generate(**batch,max_length=60,num_beams=4, num_return_sequences=num_return_sequences, temperature=0.5)
|
|
|
1 |
import torch
|
2 |
+
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
3 |
|
4 |
model_name = 'prithivida/parrot_paraphraser_on_T5'
|
5 |
torch_device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
6 |
+
tokenizer = AutoTokenizer.from_pretrained("prithivida/parrot_paraphraser_on_T5")
|
7 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("prithivida/parrot_paraphraser_on_T5")
|
|
|
8 |
def get_response(input_text,num_return_sequences):
|
9 |
batch = tokenizer.prepare_seq2seq_batch([input_text],truncation=True,padding='longest',max_length=60, return_tensors="pt").to(torch_device)
|
10 |
translated = model.generate(**batch,max_length=60,num_beams=4, num_return_sequences=num_return_sequences, temperature=0.5)
|