|
--- |
|
base_model: openbmb/MiniCPM-Llama3-V-2_5 |
|
datasets: |
|
- OpenGVLab/ShareGPT-4o |
|
--- |
|
|
|
``` |
|
import torch |
|
from PIL import Image |
|
from transformers import AutoModel, AutoTokenizer |
|
|
|
model = AutoModel.from_pretrained('macadeliccc/ShareGPT-4o-MiniCPM-Llama-3-V-2_5', trust_remote_code=True, torch_dtype=torch.float16) |
|
model = model.to(device='cuda') |
|
|
|
tokenizer = AutoTokenizer.from_pretrained('macadeliccc/ShareGPT-4o-MiniCPM-Llama-3-V-2_5', trust_remote_code=True) |
|
model.eval() |
|
|
|
image = Image.open('xx.png').convert('RGB') |
|
question = 'What is in the image?' |
|
msgs = [{'role': 'user', 'content': question}] |
|
|
|
res = model.chat( |
|
image=image, |
|
msgs=msgs, |
|
tokenizer=tokenizer, |
|
sampling=True, |
|
temperature=0.7, |
|
stream=True |
|
) |
|
|
|
generated_text = "" |
|
for new_text in res: |
|
generated_text += new_text |
|
print(new_text, flush=True, end='') |
|
|
|
``` |