hey1800 commited on
Commit
a211c4e
·
verified ·
1 Parent(s): 588386c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -3
app.py CHANGED
@@ -1,12 +1,13 @@
1
  import gradio as gr
2
  from transformers import AutoTokenizer, GPTJForCausalLM
 
3
 
4
  model_name = "rycont/kakaobrain__kogpt-6b-8bit"
5
  tokenizer = AutoTokenizer.from_pretrained(model_name)
6
- model = GPTJForCausalLM.from_pretrained(model_name)
7
 
8
  def generate_response(prompt):
9
- inputs = tokenizer(prompt, return_tensors="pt")
10
  outputs = model.generate(inputs['input_ids'], max_new_tokens=50)
11
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
12
  return response
@@ -15,7 +16,7 @@ iface = gr.Interface(
15
  fn=generate_response,
16
  inputs="text",
17
  outputs="text",
18
- title="KoGPT-6B Chatbot",
19
  description="Enter a prompt and the model will generate a response."
20
  )
21
 
 
1
  import gradio as gr
2
  from transformers import AutoTokenizer, GPTJForCausalLM
3
+ import torch
4
 
5
  model_name = "rycont/kakaobrain__kogpt-6b-8bit"
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
+ model = GPTJForCausalLM.from_pretrained(model_name, device_map="auto", load_in_8bit=True)
8
 
9
  def generate_response(prompt):
10
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
11
  outputs = model.generate(inputs['input_ids'], max_new_tokens=50)
12
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
13
  return response
 
16
  fn=generate_response,
17
  inputs="text",
18
  outputs="text",
19
+ title="KoGPT-6B 8bit Chatbot",
20
  description="Enter a prompt and the model will generate a response."
21
  )
22