|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- timdettmers/openassistant-guanaco |
|
- databricks/databricks-dolly-15k |
|
--- |
|
# Mastermax Llama 7B |
|
|
|
This is a a Llama2 7B base model that was fined tuned on additional datasets, in attempts improve performance. |
|
|
|
## How to use with HugginFace pipeline |
|
|
|
``` |
|
from transformers import AutoModelForCausalLM, AutoTokenizer, pineline |
|
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
"lifeofcoding/mastermax-llama-7b", |
|
load_in_4bit=True) |
|
tokenizer = AutoTokenizer.from_pretrained("lifeofcoding/mastermax-llama-7b", trust_remote_code=True) |
|
|
|
# Generate text using the pipeline |
|
pipe = pipeline(task="text-generation", |
|
model=model, |
|
tokenizer=tokenizer, |
|
max_length=200) |
|
|
|
result = pipe(f"<s>[INST] {prompt} [/INST]") |
|
generated_text = result[0]['generated_text'] |
|
``` |
|
|
|
|
|
|
|
- **Developed by:** [LifeOfCoding](https://www.livingthecode.life/) |
|
- **Finetuned from model:** [GuardRail Llama2 7B Sharded](https://huggingface.co/guardrail/llama-2-7b-guanaco-instruct-sharded) |
|
|