nroggendorff commited on
Commit
3db71a2
1 Parent(s): 9155296

End of training

Browse files
README.md CHANGED
@@ -1,71 +1,50 @@
1
  ---
2
- license: mit
3
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
4
  tags:
5
  - trl
6
  - sft
7
- - sgd
8
  model-index:
9
  - name: mayo
10
  results: []
11
- datasets:
12
- - nroggendorff/mayo
13
- language:
14
- - en
15
  ---
16
 
17
- # Mayonnaise LLM
 
18
 
19
- Mayo is a language model fine-tuned on the [Mayo dataset](https://huggingface.co/datasets/nroggendorff/mayo) using Supervised Fine-Tuning (SFT) and Teacher Reinforced Learning (TRL) techniques. It is based on the [TinyLlama/TinyLlama-1.1B-Chat-v1.0 model](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0).
20
 
21
- ## Features
22
 
23
- - Utilizes SFT and TRL techniques for improved performance
24
- - Supports English language
25
 
26
- ## Usage
27
 
28
- To use the Mayo LLM, you can load the model using the Hugging Face Transformers library:
29
 
30
- ```python
31
- from transformers import pipeline
32
 
33
- pipe = pipeline("text-generation", model="nroggendorff/mayo")
34
 
35
- question = "What color is the sky?"
36
- conv = [{"role": "user", "content": question}]
37
 
38
- response = pipe(conv, max_new_tokens=32)[0]['generated_text'][-1]['content']
39
- print(response)
40
- ```
41
 
42
- To use the model with quantization:
43
 
44
- ```python
45
- from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
46
- import torch
 
 
 
 
 
47
 
48
- bnb_config = BitsAndBytesConfig(
49
- load_in_4bit=True,
50
- bnb_4bit_use_double_quant=True,
51
- bnb_4bit_quant_type="nf4",
52
- bnb_4bit_compute_dtype=torch.bfloat16
53
- )
54
 
55
- model_id = "nroggendorff/mayo"
56
-
57
- tokenizer = AutoTokenizer.from_pretrained(model_id)
58
- model = AutoModelForCausalLM.from_pretrained(model_id, quantization_config=bnb_config)
59
-
60
- prompt = "<|user|>\nWhat color is the sky?</s>\n"
61
- inputs = tokenizer(prompt, return_tensors="pt")
62
-
63
- outputs = model.generate(**inputs, max_new_tokens=32)
64
-
65
- generated_text = tokenizer.batch_decode(outputs)[0]
66
- print(generated_text)
67
- ```
68
-
69
- ## License
70
-
71
- This project is licensed under the MIT License.
 
1
  ---
2
+ license: apache-2.0
3
  base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
4
  tags:
5
  - trl
6
  - sft
7
+ - generated_from_trainer
8
  model-index:
9
  - name: mayo
10
  results: []
 
 
 
 
11
  ---
12
 
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # mayo
17
 
18
+ This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on an unknown dataset.
19
 
20
+ ## Model description
 
21
 
22
+ More information needed
23
 
24
+ ## Intended uses & limitations
25
 
26
+ More information needed
 
27
 
28
+ ## Training and evaluation data
29
 
30
+ More information needed
 
31
 
32
+ ## Training procedure
 
 
33
 
34
+ ### Training hyperparameters
35
 
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 0.0001
38
+ - train_batch_size: 16
39
+ - eval_batch_size: 16
40
+ - seed: 42
41
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
+ - lr_scheduler_type: linear
43
+ - training_steps: 650
44
 
45
+ ### Framework versions
 
 
 
 
 
46
 
47
+ - Transformers 4.39.3
48
+ - Pytorch 2.1.2
49
+ - Datasets 2.18.0
50
+ - Tokenizers 0.15.2
 
 
 
 
 
 
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4678b8c283713bb83d9fe047db16f087707bc5b2c53206fe90af9c0c6f4878cd
3
  size 4400216536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98550081dc345f70f3b835d7bd4a709d55de1bafaed609303c1ab6a5d926ab5b
3
  size 4400216536
runs/Jun03_20-42-00_a7f4754c6b47/events.out.tfevents.1717447323.a7f4754c6b47.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49e7e563abc07d36027bf66b7005cfd7257a7d5f44287791b3de8c911381e400
3
+ size 9192
runs/Jun03_20-46-02_a7f4754c6b47/events.out.tfevents.1717447572.a7f4754c6b47.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f238cafa2faf83b40b86fe91f856b602c4cbb315a529273575128d90af2a907
3
+ size 4640
runs/Jun03_20-47-10_a7f4754c6b47/events.out.tfevents.1717447634.a7f4754c6b47.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f7ee633ce478b8cdf313b69048a2b2a487ae61acb99dad4c9cf12f8a5aee023
3
+ size 4184
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63563c986bbdde5fddfb470e84f681a592ca55592fd812fec4d2ac172deb0cf1
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8994d669d7f58dd765f8f16f3850c8523c83adc26e951c258408f22c5ef0d68
3
  size 4920