sandeepmajumdar commited on
Commit
6d0f31a
·
1 Parent(s): 1694937

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -1,13 +1,13 @@
1
  import gradio as gr
2
  import torch
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
- #torch.set_default_tensor_type(torch.cuda.FloatTensor)
5
 
6
- def generate(text):
7
  tokenizer = AutoTokenizer.from_pretrained("bigscience/bloom-1b1")
8
  model = AutoModelForCausalLM.from_pretrained("bigscience/bloom-1b1", use_cache=True)
9
- input_ids = tokenizer(text, return_tensors="pt").to(0)
10
- sample = model.generate(**input_ids, max_length=300, num_beams = 2, num_beam_groups = 2, top_k=1, temperature=0.9, repetition_penalty = 2.0, diversity_penalty = 0.9)
11
  return tokenizer.decode(sample[0], truncate_before_pattern=[r"\n\n^#", "^'''", "\n\n\n"])
12
 
13
  with gr.Blocks() as demo:
 
1
  import gradio as gr
2
  import torch
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
+ torch.set_default_tensor_type(torch.cuda.FloatTensor)
5
 
6
+ def generate(prompt):
7
  tokenizer = AutoTokenizer.from_pretrained("bigscience/bloom-1b1")
8
  model = AutoModelForCausalLM.from_pretrained("bigscience/bloom-1b1", use_cache=True)
9
+ input_ids = tokenizer(prompt, return_tensors="pt").to(0)
10
+ sample = model.generate(**input_ids, max_length=100, num_beams = 2, num_beam_groups = 2, top_k=1, temperature=0.9, repetition_penalty = 2.0, diversity_penalty = 0.9)
11
  return tokenizer.decode(sample[0], truncate_before_pattern=[r"\n\n^#", "^'''", "\n\n\n"])
12
 
13
  with gr.Blocks() as demo: