msaavedra1234 commited on
Commit
a87fb4d
1 Parent(s): ee3708a

Upload 9 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ ggml-model-f16.gguf filter=lfs diff=lfs merge=lfs -text
Modelfile ADDED
@@ -0,0 +1 @@
 
 
1
+ FROM ./ggml-model-f16.gguf
README.md ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: tinyllama-out
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
+ <details><summary>See axolotl config</summary>
16
+
17
+ axolotl version: `0.3.0`
18
+ ```yaml
19
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
20
+ model_type: LlamaForCausalLM
21
+ tokenizer_type: LlamaTokenizer
22
+ is_llama_derived_model: true
23
+
24
+
25
+ eval_sample_packing: False #Poco dato
26
+
27
+ load_in_8bit: false
28
+ load_in_4bit: false
29
+ strict: false
30
+
31
+ datasets:
32
+ - path: data.json # or json
33
+ ds_type: json # see other options below
34
+ type: completion
35
+
36
+ dataset_prepared_path:
37
+ val_set_size: 0.05
38
+ # output_dir: ./lora-out
39
+
40
+ sequence_len: 4096
41
+ sample_packing: true
42
+ pad_to_sequence_len: true
43
+
44
+ # adapter: lora
45
+ # lora_model_dir:
46
+ # lora_r: 32
47
+ # lora_alpha: 16
48
+ # lora_dropout: 0.05
49
+ # lora_target_linear: true
50
+ # lora_fan_in_fan_out:
51
+
52
+ wandb_project:
53
+ wandb_entity:
54
+ wandb_watch:
55
+ wandb_name:
56
+ wandb_log_model:
57
+ output_dir: ./tinyllama-out
58
+ gradient_accumulation_steps: 4
59
+ micro_batch_size: 2
60
+ num_epochs: 8 #2
61
+ optimizer: adamw_bnb_8bit
62
+ lr_scheduler: cosine
63
+ learning_rate: 0.0002
64
+
65
+ train_on_inputs: false
66
+ group_by_length: false
67
+ bf16: true
68
+ fp16: false #TODO: change to true
69
+ tf32: false
70
+
71
+ gradient_checkpointing: true
72
+ early_stopping_patience:
73
+ resume_from_checkpoint:
74
+ local_rank:
75
+ logging_steps: 1
76
+ xformers_attention:
77
+ flash_attention: true
78
+
79
+ save_strategy: "no"
80
+
81
+ warmup_steps: 10
82
+ evals_per_epoch: 4
83
+ # saves_per_epoch: 1
84
+ debug:
85
+ deepspeed:
86
+ weight_decay: 0.0
87
+ fsdp:
88
+ fsdp_config:
89
+ special_tokens:
90
+
91
+ ```
92
+
93
+ </details><br>
94
+
95
+ # tinyllama-out
96
+
97
+ This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on the None dataset.
98
+ It achieves the following results on the evaluation set:
99
+ - Loss: 1.8806
100
+
101
+ ## Model description
102
+
103
+ More information needed
104
+
105
+ ## Intended uses & limitations
106
+
107
+ More information needed
108
+
109
+ ## Training and evaluation data
110
+
111
+ More information needed
112
+
113
+ ## Training procedure
114
+
115
+ ### Training hyperparameters
116
+
117
+ The following hyperparameters were used during training:
118
+ - learning_rate: 0.0002
119
+ - train_batch_size: 2
120
+ - eval_batch_size: 2
121
+ - seed: 42
122
+ - gradient_accumulation_steps: 4
123
+ - total_train_batch_size: 8
124
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
125
+ - lr_scheduler_type: cosine
126
+ - lr_scheduler_warmup_steps: 10
127
+ - num_epochs: 8
128
+
129
+ ### Training results
130
+
131
+ | Training Loss | Epoch | Step | Validation Loss |
132
+ |:-------------:|:-----:|:----:|:---------------:|
133
+ | 1.9894 | 0.13 | 1 | 1.5790 |
134
+ | 1.915 | 0.26 | 2 | 1.4849 |
135
+ | 1.642 | 0.52 | 4 | 1.4032 |
136
+ | 1.5396 | 0.77 | 6 | 1.4059 |
137
+ | 1.3746 | 1.03 | 8 | 1.4101 |
138
+ | 0.9355 | 1.23 | 10 | 1.5147 |
139
+ | 0.9266 | 1.48 | 12 | 1.5291 |
140
+ | 0.8006 | 1.74 | 14 | 1.4724 |
141
+ | 0.7664 | 2.0 | 16 | 1.4965 |
142
+ | 0.4813 | 2.16 | 18 | 1.5715 |
143
+ | 0.4193 | 2.42 | 20 | 1.5436 |
144
+ | 0.364 | 2.68 | 22 | 1.6040 |
145
+ | 0.3592 | 2.94 | 24 | 1.5823 |
146
+ | 0.1884 | 3.13 | 26 | 1.6850 |
147
+ | 0.159 | 3.39 | 28 | 1.8316 |
148
+ | 0.1641 | 3.65 | 30 | 1.7286 |
149
+ | 0.1512 | 3.9 | 32 | 1.7029 |
150
+ | 0.1563 | 4.06 | 34 | 1.7033 |
151
+ | 0.0696 | 4.32 | 36 | 1.7482 |
152
+ | 0.0643 | 4.58 | 38 | 1.8069 |
153
+ | 0.0662 | 4.84 | 40 | 1.8410 |
154
+ | 0.0709 | 5.1 | 42 | 1.8529 |
155
+ | 0.0344 | 5.26 | 44 | 1.8626 |
156
+ | 0.0468 | 5.52 | 46 | 1.8716 |
157
+ | 0.0328 | 5.77 | 48 | 1.8761 |
158
+ | 0.0353 | 6.03 | 50 | 1.8789 |
159
+ | 0.0375 | 6.23 | 52 | 1.8803 |
160
+ | 0.0345 | 6.48 | 54 | 1.8802 |
161
+ | 0.0346 | 6.74 | 56 | 1.8806 |
162
+
163
+
164
+ ### Framework versions
165
+
166
+ - Transformers 4.37.0.dev0
167
+ - Pytorch 2.0.1
168
+ - Datasets 2.16.1
169
+ - Tokenizers 0.15.0
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 2048,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 5632,
14
+ "max_position_embeddings": 4096,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 32,
17
+ "num_hidden_layers": 22,
18
+ "num_key_value_heads": 4,
19
+ "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.37.0.dev0",
26
+ "use_cache": false,
27
+ "vocab_size": 32000
28
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "max_length": 2048,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.37.0.dev0"
7
+ }
ggml-model-f16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89a729680294ea5f37ce4d9ceea6faf4b8d3a3c065083707d085b97d27cb2dc5
3
+ size 2201017248
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c118c16a1a80d74217d3334fd295f2ca3a096b21ef81535bc804508a06403be
3
+ size 2200164273
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "bos_token": "<s>",
31
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": false,
35
+ "model_max_length": 2048,
36
+ "pad_token": "</s>",
37
+ "padding_side": "right",
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "trust_remote_code": false,
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false,
44
+ "use_fast": true
45
+ }