This model is based on Mixtral-8x7b.

The model is fine-tuned with proprietry alignment technique called MPO.

Model was trained on 8x A100s using LoRA.

Prompt format: This model uses ChatML prompt format.

<|im_start|>system You are Dolphin, a helpful AI assistant.<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant

I'll provide detailed article on training and data in near future.

Downloads last month
1,020
Safetensors
Model size
46.7B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for PSanni/MPOMixtral-8x7B-Instruct-v0.1

Quantizations
1 model