DR-Rakshitha commited on
Commit
8384864
·
1 Parent(s): 57991f9

update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -30
app.py CHANGED
@@ -2,42 +2,22 @@ import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
4
  from gpt4all import GPT4All
 
5
 
6
- model = GPT4All("wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin")
 
7
 
8
- # model = AutoModelForCausalLM.from_pretrained(
9
- # "tiiuae/falcon-7b-instruct",
10
- # torch_dtype=torch.bfloat16,
11
- # trust_remote_code=True,
12
- # device_map="auto",
13
- # low_cpu_mem_usage=True,
14
- # )
15
- # tokenizer = AutoTokenizer.from_pretrained("tiiuae/falcon-7b-instruct")
16
 
 
 
17
 
18
  def generate_text(input_text):
19
- # input_ids = tokenizer.encode(input_text, return_tensors="pt")
20
- # attention_mask = torch.ones(input_ids.shape)
21
-
22
- output = model.generate(
23
- input_text
24
- # input_ids,
25
- # attention_mask=attention_mask,
26
- # max_length=200,
27
- # do_sample=True,
28
- # top_k=10,
29
- # num_return_sequences=1,
30
- # eos_token_id=tokenizer.eos_token_id,
31
- )
32
-
33
- # output_text = tokenizer.decode(output[0], skip_special_tokens=True)
34
- # print(output_text)
35
-
36
- # Remove Prompt Echo from Generated Text
37
- # cleaned_output_text = output_text.replace(input_text, "")
38
  return output
39
 
40
-
41
  text_generation_interface = gr.Interface(
42
  fn=generate_text,
43
  inputs=[
@@ -45,4 +25,4 @@ text_generation_interface = gr.Interface(
45
  ],
46
  outputs=gr.inputs.Textbox(label="Generated Text"),
47
  title="Falcon-7B Instruct",
48
- ).launch()
 
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
  import torch
4
  from gpt4all import GPT4All
5
+ import os
6
 
7
+ # Specify the local path to the downloaded model file
8
+ model_path = "path/to/wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin"
9
 
10
+ # Check if the model file exists locally
11
+ if not os.path.exists(model_path):
12
+ raise FileNotFoundError(f"Model file not found at {model_path}. Please download it manually.")
 
 
 
 
 
13
 
14
+ # Initialize the GPT4All model
15
+ model = GPT4All(model_path)
16
 
17
  def generate_text(input_text):
18
+ output = model.generate(input_text)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  return output
20
 
 
21
  text_generation_interface = gr.Interface(
22
  fn=generate_text,
23
  inputs=[
 
25
  ],
26
  outputs=gr.inputs.Textbox(label="Generated Text"),
27
  title="Falcon-7B Instruct",
28
+ ).launch()