ChenWu98 commited on
Commit
8a27aef
1 Parent(s): 062b67a

Model save

Browse files
README.md CHANGED
@@ -1,15 +1,9 @@
1
  ---
2
  library_name: peft
3
  tags:
4
- - alignment-handbook
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- - trl
9
- - sft
10
- - generated_from_trainer
11
- datasets:
12
- - ChenWu98/skills_metaphor_chat
13
  base_model: merged_ckpts/skills_red_herring_chat-lora
14
  model-index:
15
  - name: skills_red_herring_chat-then-skills_metaphor_chat-lora
@@ -21,9 +15,9 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # skills_red_herring_chat-then-skills_metaphor_chat-lora
23
 
24
- This model is a fine-tuned version of [merged_ckpts/skills_red_herring_chat-lora](https://huggingface.co/merged_ckpts/skills_red_herring_chat-lora) on the ChenWu98/skills_metaphor_chat dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 0.2163
27
 
28
  ## Model description
29
 
@@ -52,14 +46,13 @@ The following hyperparameters were used during training:
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: cosine
54
  - lr_scheduler_warmup_ratio: 0.1
55
- - num_epochs: 2.0
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss |
60
  |:-------------:|:-----:|:----:|:---------------:|
61
- | 0.4383 | 0.96 | 9 | 0.2286 |
62
- | 0.2047 | 1.92 | 18 | 0.2163 |
63
 
64
 
65
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  tags:
 
4
  - trl
5
  - sft
6
  - generated_from_trainer
 
 
 
 
 
7
  base_model: merged_ckpts/skills_red_herring_chat-lora
8
  model-index:
9
  - name: skills_red_herring_chat-then-skills_metaphor_chat-lora
 
15
 
16
  # skills_red_herring_chat-then-skills_metaphor_chat-lora
17
 
18
+ This model was trained from scratch on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.2417
21
 
22
  ## Model description
23
 
 
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
  - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 1.0
50
 
51
  ### Training results
52
 
53
  | Training Loss | Epoch | Step | Validation Loss |
54
  |:-------------:|:-----:|:----:|:---------------:|
55
+ | 0.408 | 0.96 | 9 | 0.2417 |
 
56
 
57
 
58
  ### Framework versions
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "gate_proj",
23
- "up_proj",
24
- "q_proj",
25
- "k_proj",
26
- "down_proj",
27
  "o_proj",
28
- "v_proj"
 
 
 
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "v_proj",
23
  "gate_proj",
 
 
 
 
24
  "o_proj",
25
+ "k_proj",
26
+ "q_proj",
27
+ "up_proj",
28
+ "down_proj"
29
  ],
30
  "task_type": "CAUSAL_LM"
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f4458a655d1b37b0847767ec7e99c7c07d8847970c528fb8d8fb9a490552aac
3
  size 83946192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ef8afaffc00a0a1ed24e6bf2eb5020016451e603284826da6f8d2a45dc00a7
3
  size 83946192
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 1.92,
3
- "eval_loss": 0.21630115807056427,
4
- "eval_runtime": 4.1792,
5
  "eval_samples": 100,
6
- "eval_samples_per_second": 23.928,
7
- "eval_steps_per_second": 3.111,
8
- "train_loss": 0.2861863639619615,
9
- "train_runtime": 207.9549,
10
  "train_samples": 300,
11
- "train_samples_per_second": 2.885,
12
- "train_steps_per_second": 0.087
13
  }
 
1
  {
2
+ "epoch": 0.96,
3
+ "eval_loss": 0.24174152314662933,
4
+ "eval_runtime": 5.1895,
5
  "eval_samples": 100,
6
+ "eval_samples_per_second": 19.27,
7
+ "eval_steps_per_second": 2.505,
8
+ "train_loss": 0.3598402738571167,
9
+ "train_runtime": 147.976,
10
  "train_samples": 300,
11
+ "train_samples_per_second": 2.027,
12
+ "train_steps_per_second": 0.061
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.92,
3
- "eval_loss": 0.21630115807056427,
4
- "eval_runtime": 4.1792,
5
  "eval_samples": 100,
6
- "eval_samples_per_second": 23.928,
7
- "eval_steps_per_second": 3.111
8
  }
 
1
  {
2
+ "epoch": 0.96,
3
+ "eval_loss": 0.24174152314662933,
4
+ "eval_runtime": 5.1895,
5
  "eval_samples": 100,
6
+ "eval_samples_per_second": 19.27,
7
+ "eval_steps_per_second": 2.505
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 1.92,
3
- "train_loss": 0.2861863639619615,
4
- "train_runtime": 207.9549,
5
  "train_samples": 300,
6
- "train_samples_per_second": 2.885,
7
- "train_steps_per_second": 0.087
8
  }
 
1
  {
2
+ "epoch": 0.96,
3
+ "train_loss": 0.3598402738571167,
4
+ "train_runtime": 147.976,
5
  "train_samples": 300,
6
+ "train_samples_per_second": 2.027,
7
+ "train_steps_per_second": 0.061
8
  }
trainer_state.json CHANGED
@@ -1,69 +1,49 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.92,
5
  "eval_steps": 500,
6
- "global_step": 18,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.11,
13
- "learning_rate": 0.0001,
14
  "loss": 0.6469,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.53,
19
- "learning_rate": 0.00018314696123025454,
20
- "loss": 0.4383,
21
  "step": 5
22
  },
23
  {
24
  "epoch": 0.96,
25
- "eval_loss": 0.2286338359117508,
26
- "eval_runtime": 5.2482,
27
- "eval_samples_per_second": 19.054,
28
- "eval_steps_per_second": 2.477,
29
  "step": 9
30
  },
31
  {
32
- "epoch": 1.07,
33
- "learning_rate": 0.0001,
34
- "loss": 0.23,
35
- "step": 10
36
- },
37
- {
38
- "epoch": 1.6,
39
- "learning_rate": 1.6853038769745467e-05,
40
- "loss": 0.2047,
41
- "step": 15
42
- },
43
- {
44
- "epoch": 1.92,
45
- "eval_loss": 0.21630115807056427,
46
- "eval_runtime": 4.1963,
47
- "eval_samples_per_second": 23.83,
48
- "eval_steps_per_second": 3.098,
49
- "step": 18
50
- },
51
- {
52
- "epoch": 1.92,
53
- "step": 18,
54
- "total_flos": 19174999326720.0,
55
- "train_loss": 0.2861863639619615,
56
- "train_runtime": 207.9549,
57
- "train_samples_per_second": 2.885,
58
- "train_steps_per_second": 0.087
59
  }
60
  ],
61
  "logging_steps": 5,
62
- "max_steps": 18,
63
  "num_input_tokens_seen": 0,
64
- "num_train_epochs": 2,
65
  "save_steps": 500,
66
- "total_flos": 19174999326720.0,
67
  "train_batch_size": 4,
68
  "trial_name": null,
69
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.96,
5
  "eval_steps": 500,
6
+ "global_step": 9,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.11,
13
+ "learning_rate": 0.0002,
14
  "loss": 0.6469,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.53,
19
+ "learning_rate": 0.0001,
20
+ "loss": 0.408,
21
  "step": 5
22
  },
23
  {
24
  "epoch": 0.96,
25
+ "eval_loss": 0.24174152314662933,
26
+ "eval_runtime": 6.0031,
27
+ "eval_samples_per_second": 16.658,
28
+ "eval_steps_per_second": 2.166,
29
  "step": 9
30
  },
31
  {
32
+ "epoch": 0.96,
33
+ "step": 9,
34
+ "total_flos": 9507448946688.0,
35
+ "train_loss": 0.3598402738571167,
36
+ "train_runtime": 147.976,
37
+ "train_samples_per_second": 2.027,
38
+ "train_steps_per_second": 0.061
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
  }
40
  ],
41
  "logging_steps": 5,
42
+ "max_steps": 9,
43
  "num_input_tokens_seen": 0,
44
+ "num_train_epochs": 1,
45
  "save_steps": 500,
46
+ "total_flos": 9507448946688.0,
47
  "train_batch_size": 4,
48
  "trial_name": null,
49
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a792dd90748717635ab48db1bfbcda9995f658c6cd97df432481828a1ad566cf
3
  size 6008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c507c7139f6ecd836dacd1b5db7cfff68767efc1e8ff6754fd31a5df7f0db554
3
  size 6008