Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
from transformers import pipeline
|
2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
import gradio as gr
|
|
|
4 |
|
5 |
model = "janny127/autotrain-7qmts-cs1er"
|
6 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
@@ -31,9 +32,12 @@ def generate_answer(query, sample_num=3):
|
|
31 |
for seq in sequences:
|
32 |
answer = seq['generated_text'].replace(formatted_prompt, "")
|
33 |
answers.append(answer)
|
34 |
-
|
35 |
-
|
36 |
-
|
|
|
|
|
|
|
37 |
|
38 |
interface = gr.ChatInterface(
|
39 |
fn=generate_answer,
|
|
|
1 |
from transformers import pipeline
|
2 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
import gradio as gr
|
4 |
+
from nltk.tokenize import sent_tokenize
|
5 |
|
6 |
model = "janny127/autotrain-7qmts-cs1er"
|
7 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
|
|
32 |
for seq in sequences:
|
33 |
answer = seq['generated_text'].replace(formatted_prompt, "")
|
34 |
answers.append(answer)
|
35 |
+
answer = sent_tokenize(answers[0])
|
36 |
+
final_answer = ''
|
37 |
+
for an in answer:
|
38 |
+
final_answer += an.strip()
|
39 |
+
return final_answer
|
40 |
+
|
41 |
|
42 |
interface = gr.ChatInterface(
|
43 |
fn=generate_answer,
|