Spaces:
Sleeping
Sleeping
# | |
# Simple example. | |
# | |
import spaces | |
from diffusers import DiffusionPipeline | |
import os | |
import torch | |
from transformers import pipeline | |
import gradio as gr | |
token = os.getenv("HUGGINGFACE_API_TOKEN") | |
model = "meta-llama/Meta-Llama-3-8B-Instruct" | |
model = "instructlab/granite-7b-lab" | |
model = "ibm/granite-7b-base" | |
model = "ibm-granite/granite-3b-code-instruct" | |
print(f'Loading model {model}') | |
pipe = pipeline("text-generation", model, torch_dtype=torch.bfloat16, device_map="auto", token=token) | |
# pipe.to('cuda') | |
def generate(prompt): | |
response = pipe(prompt, max_new_tokens=512) | |
# r = response[0]['generated_text'][-1]['content'] | |
print(f'Response received!') | |
r = response[0]['generated_text'] | |
return r | |
input_textbox = gr.Textbox( | |
label="Prompt", | |
info="Ask me something.", | |
lines=3, | |
value="# Write a python function to read a csv file using pandas and print rows 20 through 25." | |
) | |
gr.Interface( | |
fn=generate, | |
inputs=input_textbox, | |
outputs=gr.Text(), | |
title=model | |
).launch() | |