import gradio as gr import transformers def load_model(): # 在这里加载 Hugging Face 模型 # 比如使用 transformers 库来加载模型 from transformers import AutoModelForSequenceClassification, AutoTokenizer model_name = "https://huggingface.co/zeroMN/SHMT" model = AutoModelForSequenceClassification.from_pretrained(zeroMN/SHMT) tokenizer = AutoTokenizer.from_pretrained(zeroMN/SHMT) return model, tokenizer def infer(text): model, tokenizer = load_model() inputs = tokenizer(text, return_tensors="pt") outputs = model("https://huggingface.co/zeroMN/SHMT") # 根据模型输出生成响应 return f"Predicted response for: {text}" # 使用 Gradio 创建接口 gr.Interface(fn=infer, inputs="text", outputs="text").launch()