Safe-o1-V Model Card π€β¨
Model Overview π
Safe-o1-V
is an innovative multi-modal language model that introduces a self-monitoring thinking process to detect and filter unsafe content, achieving more robust safety performance π.
Features and Highlights π
- Safety First π: Through a self-monitoring mechanism, it detects potential unsafe content in the thinking process in real-time, ensuring outputs consistently align with ethical and safety standards.
- Enhanced Robustness π‘: Compared to traditional models,
Safe-o1-V
performs more stably in complex scenarios, reducing unexpected "derailments." - User-Friendly π: Designed to provide users with a trustworthy conversational partner, suitable for various application scenarios, striking a balance between helpfulness and harmfulness.
Usage π
You can load Safe-o1-V
using the Hugging Face transformers
library:
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("PKU-Alignment/Safe-o1-V")
model = AutoModelForCausalLM.from_pretrained("PKU-Alignment/Safe-o1-V")
input_text = "Hello, World!"
inputs = tokenizer(input_text, return_tensors="pt")
outputs = model.generate(**inputs)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
- Downloads last month
- 8
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
HF Inference deployability: The model has no library tag.