beamaia commited on
Commit
32faaa7
·
verified ·
1 Parent(s): 4ccd466

Model save

Browse files
README.md CHANGED
@@ -18,13 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.0052
22
- - Rewards/chosen: 6.8557
23
- - Rewards/rejected: -40.2341
24
- - Rewards/margins: 47.0897
25
- - Kl: 0.0
26
- - Logps/chosen: -138.8257
27
- - Logps/rejected: -603.3300
28
 
29
  ## Model description
30
 
@@ -45,35 +45,21 @@ More information needed
45
  The following hyperparameters were used during training:
46
  - learning_rate: 0.0002
47
  - train_batch_size: 4
48
- - eval_batch_size: 4
49
  - seed: 42
50
  - gradient_accumulation_steps: 8
51
  - total_train_batch_size: 32
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
  - lr_scheduler_warmup_ratio: 0.03
55
- - training_steps: 786
56
  - mixed_precision_training: Native AMP
57
 
58
  ### Training results
59
 
60
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/margins | Kl | Logps/chosen | Logps/rejected |
61
- |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:---------------:|:---:|:------------:|:--------------:|
62
- | 0.1911 | 0.38 | 50 | 0.0198 | 5.0250 | -23.0117 | 28.0367 | 0.0 | -157.1321 | -431.1063 |
63
- | 0.0525 | 0.76 | 100 | 0.0155 | 5.6300 | -30.1143 | 35.7443 | 0.0 | -151.0826 | -502.1325 |
64
- | 0.0684 | 1.14 | 150 | 0.0104 | 6.1959 | -32.9347 | 39.1306 | 0.0 | -145.4235 | -530.3358 |
65
- | 0.0253 | 1.52 | 200 | 0.0126 | 5.9530 | -28.8204 | 34.7734 | 0.0 | -147.8525 | -489.1933 |
66
- | 0.0177 | 1.9 | 250 | 0.0163 | 6.4453 | -53.9510 | 60.3963 | 0.0 | -142.9294 | -740.4993 |
67
- | 0.0126 | 2.28 | 300 | 0.0206 | 6.2002 | -65.0484 | 71.2486 | 0.0 | -145.3807 | -851.4736 |
68
- | 0.007 | 2.66 | 350 | 0.0068 | 6.6872 | -41.5169 | 48.2041 | 0.0 | -140.5102 | -616.1578 |
69
- | 0.0085 | 3.04 | 400 | 0.0076 | 6.7677 | -52.6223 | 59.3899 | 0.0 | -139.7057 | -727.2117 |
70
- | 0.0057 | 3.43 | 450 | 0.0060 | 6.5419 | -38.1888 | 44.7308 | 0.0 | -141.9630 | -582.8775 |
71
- | 0.0029 | 3.81 | 500 | 0.0061 | 6.7190 | -37.0418 | 43.7608 | 0.0 | -140.1928 | -571.4076 |
72
- | 0.003 | 4.19 | 550 | 0.0062 | 6.8462 | -36.6584 | 43.5045 | 0.0 | -138.9207 | -567.5727 |
73
- | 0.0023 | 4.57 | 600 | 0.0058 | 6.8631 | -38.2151 | 45.0782 | 0.0 | -138.7514 | -583.1401 |
74
- | 0.0028 | 4.95 | 650 | 0.0050 | 6.8261 | -40.4010 | 47.2271 | 0.0 | -139.1210 | -604.9990 |
75
- | 0.0029 | 5.33 | 700 | 0.0052 | 6.8557 | -40.2341 | 47.0897 | 0.0 | -138.8257 | -603.3300 |
76
- | 0.0022 | 5.71 | 750 | 0.0052 | 6.8628 | -40.4610 | 47.3239 | 0.0 | -138.7541 | -605.5995 |
77
 
78
 
79
  ### Framework versions
@@ -82,4 +68,4 @@ The following hyperparameters were used during training:
82
  - Transformers 4.39.1
83
  - Pytorch 2.1.0+cu118
84
  - Datasets 2.18.0
85
- - Tokenizers 0.15.2
 
18
 
19
  This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.0306
22
+ - Rewards/chosen: 5.5134
23
+ - Logps/chosen: -155.1431
24
+ - Rewards/rejected: -19.7208
25
+ - Logps/rejected: -384.9552
26
+ - Kl: 0.3506
27
+ - Rewards/margins: 25.2388
28
 
29
  ## Model description
30
 
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 0.0002
47
  - train_batch_size: 4
48
+ - eval_batch_size: 2
49
  - seed: 42
50
  - gradient_accumulation_steps: 8
51
  - total_train_batch_size: 32
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
  - lr_scheduler_warmup_ratio: 0.03
55
+ - training_steps: 96
56
  - mixed_precision_training: Native AMP
57
 
58
  ### Training results
59
 
60
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Logps/chosen | Rewards/rejected | Logps/rejected | Kl | Rewards/margins |
61
+ |:-------------:|:-----:|:----:|:---------------:|:--------------:|:------------:|:----------------:|:--------------:|:------:|:---------------:|
62
+ | 0.0687 | 0.38 | 50 | 0.0442 | 4.9556 | -160.7203 | -13.7592 | -325.3391 | 0.6980 | 18.5049 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
63
 
64
 
65
  ### Framework versions
 
68
  - Transformers 4.39.1
69
  - Pytorch 2.1.0+cu118
70
  - Datasets 2.18.0
71
+ - Tokenizers 0.15.1
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "q_proj",
25
  "o_proj",
26
- "v_proj"
 
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "o_proj",
24
+ "q_proj",
25
+ "v_proj",
26
+ "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9ff12f3f4d5befc156462ce388fd17f745898fa371492d3ca8fd33ef9dd7c58
3
- size 54560368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:099fac32cd1fbefde302f0c47045a7a1fe2d172ff0a9c19cabee49171a0d67d0
3
+ size 1103203256
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32001,
3
+ "<|im_start|>": 32000
4
+ }
special_tokens_map.json CHANGED
@@ -1,24 +1,23 @@
1
  {
2
  "additional_special_tokens": [
3
- "<unk>",
4
- "<s>",
5
- "</s>"
 
 
 
 
 
 
 
 
 
 
 
6
  ],
7
- "bos_token": {
8
- "content": "<s>",
9
- "lstrip": false,
10
- "normalized": false,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "eos_token": {
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": false,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "pad_token": "<unk>",
22
  "unk_token": {
23
  "content": "<unk>",
24
  "lstrip": false,
 
1
  {
2
  "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|im_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
  ],
18
+ "bos_token": "<|im_start|>",
19
+ "eos_token": "<|im_end|>",
20
+ "pad_token": "<|im_end|>",
 
 
 
 
 
 
 
 
 
 
 
 
21
  "unk_token": {
22
  "content": "<unk>",
23
  "lstrip": false,
tokenizer.json CHANGED
@@ -29,6 +29,24 @@
29
  "rstrip": false,
30
  "normalized": false,
31
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  }
33
  ],
34
  "normalizer": {
 
29
  "rstrip": false,
30
  "normalized": false,
31
  "special": true
32
+ },
33
+ {
34
+ "id": 32000,
35
+ "content": "<|im_start|>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ },
42
+ {
43
+ "id": 32001,
44
+ "content": "<|im_end|>",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
  }
51
  ],
52
  "normalizer": {
tokenizer_config.json CHANGED
@@ -25,21 +25,36 @@
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  }
29
  },
30
  "additional_special_tokens": [
31
- "<unk>",
32
- "<s>",
33
- "</s>"
34
  ],
35
- "bos_token": "<s>",
36
- "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
37
  "clean_up_tokenization_spaces": false,
38
- "eos_token": "</s>",
39
  "legacy": true,
40
  "max_lenght": 8192,
41
  "model_max_length": 1000000000000000019884624838656,
42
- "pad_token": "<unk>",
43
  "padding": true,
44
  "sp_model_kwargs": {},
45
  "spaces_between_special_tokens": false,
 
25
  "rstrip": false,
26
  "single_word": false,
27
  "special": true
28
+ },
29
+ "32000": {
30
+ "content": "<|im_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "32001": {
38
+ "content": "<|im_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
  }
45
  },
46
  "additional_special_tokens": [
47
+ "<|im_start|>",
48
+ "<|im_end|>"
 
49
  ],
50
+ "bos_token": "<|im_start|>",
51
+ "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
52
  "clean_up_tokenization_spaces": false,
53
+ "eos_token": "<|im_end|>",
54
  "legacy": true,
55
  "max_lenght": 8192,
56
  "model_max_length": 1000000000000000019884624838656,
57
+ "pad_token": "<|im_end|>",
58
  "padding": true,
59
  "sp_model_kwargs": {},
60
  "spaces_between_special_tokens": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72ac543b1af4f9675ddc3d87c79e0f49fbd145783e3bc0097b4be05d98c85865
3
- size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d7d2996f7c9647166fa0a6de85aa22f4274af7424b79d82eb5d5b498f157d56
3
+ size 5688