Spaces:
Sleeping
Sleeping
| import gradio as gr | |
| from transformers import AutoModelForCausalLM, AutoTokenizer,BitsAndBytesConfig | |
| from transformers.generation import GenerationConfig | |
| quantization_config = BitsAndBytesConfig( | |
| load_in_4bit=True, | |
| bnb_4bit_quant_type='int8', | |
| bnb_4bit_compute_dtype=torch.bfloat16) | |
| # Note: The default behavior now has injection attack prevention off. | |
| tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen-7B-Chat", trust_remote_code=True) | |
| model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen-7B-Chat", device_map="auto", trust_remote_code=True,quantization_config=quantization_config).eval() | |
| # Specify hyperparameters for generation | |
| model.generation_config = GenerationConfig.from_pretrained("Qwen/Qwen-7B-Chat", trust_remote_code=True) # 可指定不同的生成长度、top_p等相关超参 | |
| def generate(text): | |
| response, history = model.chat(tokenizer, text, history=None) | |
| return response | |
| examples = [ | |
| ["The Moon's orbit around Earth has"], | |
| ["The smooth Borealis basin in the Northern Hemisphere covers 40%"], | |
| ] | |
| demo = gr.Interface( | |
| fn=generate, | |
| inputs=gr.inputs.Textbox(lines=5, label="Input Text"), | |
| outputs=gr.outputs.Textbox(label="Generated Text"), | |
| examples=examples | |
| ) | |
| demo.launch() | |