Tippawan commited on
Commit
ce1bc2d
·
verified ·
1 Parent(s): b4c6da9

End of training

Browse files
Files changed (2) hide show
  1. README.md +13 -14
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -1,12 +1,12 @@
1
  ---
2
- base_model: SeaLLMs/SeaLLM3-7B-Chat
3
  library_name: peft
4
  license: other
 
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
- - name: proof-reading-SeaLLM3-7B-Chat-3090-v10
10
  results: []
11
  ---
12
 
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
  <details><summary>See axolotl config</summary>
18
 
19
- axolotl version: `0.4.1`
20
  ```yaml
21
  base_model: SeaLLMs/SeaLLM3-7B-Chat
22
  trust_remote_code: true
@@ -26,9 +26,8 @@ load_in_4bit: true
26
  strict: false
27
 
28
  datasets:
29
- - path: Tippawan/pr-10-wiki-seallm
30
- type: sharegpt
31
- split: 'train[:100000]'
32
  conversation: chatml
33
  field_messages: messages
34
  chat_template: chatml
@@ -42,7 +41,7 @@ eval_sample_packing: false
42
  pad_to_sequence_len: false
43
 
44
  push_to_hub: true
45
- hub_model_id: Tippawan/proof-reading-SeaLLM3-7B-Chat-3090-v10 # Replace with your Hugging Face repo ID
46
  use_auth_token: true # Ensure you have set your Hugging Face API token in the environment
47
  hub_private_repo: true # Set to true if you want the repository to be private
48
  hub_strategy: all_checkpoints
@@ -57,7 +56,7 @@ lora_dropout: 0.05
57
  lora_target_linear: true
58
  lora_fan_in_fan_out:
59
 
60
- wandb_project: proof-reading-SeaLLM3-7B-Chat-3090-v10
61
  wandb_entity:
62
  wandb_watch:
63
  wandb_name:
@@ -97,7 +96,7 @@ special_tokens:
97
 
98
  </details><br>
99
 
100
- # proof-reading-SeaLLM3-7B-Chat-3090-v10
101
 
102
  This model is a fine-tuned version of [SeaLLMs/SeaLLM3-7B-Chat](https://huggingface.co/SeaLLMs/SeaLLM3-7B-Chat) on the None dataset.
103
 
@@ -124,7 +123,7 @@ The following hyperparameters were used during training:
124
  - seed: 42
125
  - gradient_accumulation_steps: 4
126
  - total_train_batch_size: 8
127
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
128
  - lr_scheduler_type: cosine
129
  - lr_scheduler_warmup_steps: 10
130
  - num_epochs: 1
@@ -135,8 +134,8 @@ The following hyperparameters were used during training:
135
 
136
  ### Framework versions
137
 
138
- - PEFT 0.12.0
139
- - Transformers 4.45.0.dev0
140
  - Pytorch 2.3.1+cu121
141
- - Datasets 2.21.0
142
- - Tokenizers 0.19.1
 
1
  ---
 
2
  library_name: peft
3
  license: other
4
+ base_model: SeaLLMs/SeaLLM3-7B-Chat
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
+ - name: proof-reading-SeaLLM3-7B-Chat-3090-v11
10
  results: []
11
  ---
12
 
 
16
  [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
  <details><summary>See axolotl config</summary>
18
 
19
+ axolotl version: `0.5.0`
20
  ```yaml
21
  base_model: SeaLLMs/SeaLLM3-7B-Chat
22
  trust_remote_code: true
 
26
  strict: false
27
 
28
  datasets:
29
+ - path: Tippawan/p11-seallm
30
+ type: chat_template
 
31
  conversation: chatml
32
  field_messages: messages
33
  chat_template: chatml
 
41
  pad_to_sequence_len: false
42
 
43
  push_to_hub: true
44
+ hub_model_id: Tippawan/proof-reading-SeaLLM3-7B-Chat-3090-v11 # Replace with your Hugging Face repo ID
45
  use_auth_token: true # Ensure you have set your Hugging Face API token in the environment
46
  hub_private_repo: true # Set to true if you want the repository to be private
47
  hub_strategy: all_checkpoints
 
56
  lora_target_linear: true
57
  lora_fan_in_fan_out:
58
 
59
+ wandb_project: proof-reading-SeaLLM3-7B-Chat-3090-v11
60
  wandb_entity:
61
  wandb_watch:
62
  wandb_name:
 
96
 
97
  </details><br>
98
 
99
+ # proof-reading-SeaLLM3-7B-Chat-3090-v11
100
 
101
  This model is a fine-tuned version of [SeaLLMs/SeaLLM3-7B-Chat](https://huggingface.co/SeaLLMs/SeaLLM3-7B-Chat) on the None dataset.
102
 
 
123
  - seed: 42
124
  - gradient_accumulation_steps: 4
125
  - total_train_batch_size: 8
126
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
127
  - lr_scheduler_type: cosine
128
  - lr_scheduler_warmup_steps: 10
129
  - num_epochs: 1
 
134
 
135
  ### Framework versions
136
 
137
+ - PEFT 0.13.2
138
+ - Transformers 4.46.1
139
  - Pytorch 2.3.1+cu121
140
+ - Datasets 3.0.1
141
+ - Tokenizers 0.20.3
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:510bb9fcb5e688917e13ab4eb4ad4b47014c4f16f157be766a5439ece5fe30b1
3
  size 161621802
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5be0710fe18dc5c0bce44d297de83dbbff8402c49f8a3cf6c7284c445680f90f
3
  size 161621802