base_model: mistralai/Mistral-7B-Instruct-v0.3 | |
datasets: | |
- nroggendorff/mayo | |
language: | |
- en | |
license: mit | |
tags: | |
- trl | |
- sft | |
- sgd | |
model-index: | |
- name: mayo | |
results: [] | |
# Mayonnaise LLM | |
Mayo is a language model fine-tuned on the [Mayo dataset](https://huggingface.co/datasets/nroggendorff/mayo) using Supervised Fine-Tuning (SFT) and Teacher Reinforced Learning (TRL) techniques. It is based on the [Mistral 7b Model](mistralai/Mistral-7B-Instruct-v0.3) | |
## Features | |
- Utilizes SFT and TRL techniques for improved performance | |
- Supports English language | |
## Usage | |
To use the Mayo LLM, you can load the model using the Hugging Face Transformers library: | |
```python | |
from transformers import pipeline | |
pipe = pipeline("text-generation", model="nroggendorff/mayo") | |
question = "What color is the sky?" | |
conv = [{"role": "user", "content": question}] | |
response = pipe(conv, max_new_tokens=32)[0]['generated_text'][-1]['content'] | |
print(response) | |
``` | |
## License | |
This project is licensed under the MIT License. |