Minbyul commited on
Commit
d1bf441
1 Parent(s): d8e5459

Model save

Browse files
README.md CHANGED
@@ -1,15 +1,10 @@
1
  ---
2
  base_model: Minbyul/selfbiorag-7b-wo-medication_qa-sft
3
  tags:
4
- - alignment-handbook
5
- - trl
6
- - dpo
7
- - generated_from_trainer
8
  - trl
9
  - dpo
 
10
  - generated_from_trainer
11
- datasets:
12
- - HuggingFaceH4/ultrafeedback_binarized
13
  model-index:
14
  - name: selfbiorag-7b-dpo-full-sft-wo-medication_qa
15
  results: []
@@ -20,17 +15,17 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  # selfbiorag-7b-dpo-full-sft-wo-medication_qa
22
 
23
- This model is a fine-tuned version of [Minbyul/selfbiorag-7b-wo-medication_qa-sft](https://huggingface.co/Minbyul/selfbiorag-7b-wo-medication_qa-sft) on the HuggingFaceH4/ultrafeedback_binarized dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 0.2758
26
- - Rewards/chosen: -1.2291
27
- - Rewards/rejected: -7.1014
28
- - Rewards/accuracies: 0.8949
29
- - Rewards/margins: 5.8722
30
- - Logps/rejected: -1442.6945
31
- - Logps/chosen: -679.7597
32
- - Logits/rejected: -0.3284
33
- - Logits/chosen: -0.3526
34
 
35
  ## Model description
36
 
@@ -65,11 +60,11 @@ The following hyperparameters were used during training:
65
 
66
  ### Training results
67
 
68
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
69
- |:-------------:|:-----:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
70
- | 0.2249 | 0.32 | 100 | 0.3821 | -0.9339 | -4.5771 | 0.8551 | 3.6432 | -1190.2701 | -650.2339 | -0.0290 | -0.1107 |
71
- | 0.1549 | 0.65 | 200 | 0.2709 | -0.9607 | -5.7585 | 0.8977 | 4.7978 | -1308.4048 | -652.9113 | -0.3222 | -0.3180 |
72
- | 0.0946 | 0.97 | 300 | 0.2756 | -1.2277 | -7.0991 | 0.8920 | 5.8714 | -1442.4718 | -679.6155 | -0.3283 | -0.3523 |
73
 
74
 
75
  ### Framework versions
 
1
  ---
2
  base_model: Minbyul/selfbiorag-7b-wo-medication_qa-sft
3
  tags:
 
 
 
 
4
  - trl
5
  - dpo
6
+ - alignment-handbook
7
  - generated_from_trainer
 
 
8
  model-index:
9
  - name: selfbiorag-7b-dpo-full-sft-wo-medication_qa
10
  results: []
 
15
 
16
  # selfbiorag-7b-dpo-full-sft-wo-medication_qa
17
 
18
+ This model is a fine-tuned version of [Minbyul/selfbiorag-7b-wo-medication_qa-sft](https://huggingface.co/Minbyul/selfbiorag-7b-wo-medication_qa-sft) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Logits/chosen: -0.3523
21
+ - Logits/rejected: -0.3283
22
+ - Logps/chosen: -679.6155
23
+ - Logps/rejected: -1442.4718
24
+ - Loss: 0.2756
25
+ - Rewards/accuracies: 0.8920
26
+ - Rewards/chosen: -1.2277
27
+ - Rewards/margins: 5.8714
28
+ - Rewards/rejected: -7.0991
29
 
30
  ## Model description
31
 
 
60
 
61
  ### Training results
62
 
63
+ | Training Loss | Epoch | Step | Logits/chosen | Logits/rejected | Logps/chosen | Logps/rejected | Validation Loss | Rewards/accuracies | Rewards/chosen | Rewards/margins | Rewards/rejected |
64
+ |:-------------:|:-----:|:----:|:-------------:|:---------------:|:------------:|:--------------:|:---------------:|:------------------:|:--------------:|:---------------:|:----------------:|
65
+ | 0.2249 | 0.32 | 100 | -0.1107 | -0.0290 | -650.2339 | -1190.2701 | 0.3821 | 0.8551 | -0.9339 | 3.6432 | -4.5771 |
66
+ | 0.1549 | 0.65 | 200 | -0.3180 | -0.3222 | -652.9113 | -1308.4048 | 0.2709 | 0.8977 | -0.9607 | 4.7978 | -5.7585 |
67
+ | 0.0946 | 0.97 | 300 | -0.3523 | -0.3283 | -679.6155 | -1442.4718 | 0.2756 | 0.8920 | -1.2277 | 5.8714 | -7.0991 |
68
 
69
 
70
  ### Framework versions
all_results.json CHANGED
@@ -13,9 +13,9 @@
13
  "eval_samples": 1392,
14
  "eval_samples_per_second": 8.061,
15
  "eval_steps_per_second": 0.255,
16
- "train_loss": 0.23439283587014406,
17
- "train_runtime": 5872.9997,
18
  "train_samples": 19761,
19
- "train_samples_per_second": 3.365,
20
- "train_steps_per_second": 0.053
21
  }
 
13
  "eval_samples": 1392,
14
  "eval_samples_per_second": 8.061,
15
  "eval_steps_per_second": 0.255,
16
+ "train_loss": 0.0030385508506429234,
17
+ "train_runtime": 144.6456,
18
  "train_samples": 19761,
19
+ "train_samples_per_second": 136.617,
20
+ "train_steps_per_second": 2.136
21
  }
config.json CHANGED
@@ -23,6 +23,6 @@
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.39.0.dev0",
26
- "use_cache": true,
27
  "vocab_size": 32016
28
  }
 
23
  "tie_word_embeddings": false,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.39.0.dev0",
26
+ "use_cache": false,
27
  "vocab_size": 32016
28
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e89a2cf022407c47d43554bd2be48b24f21d95171afc83f83b7a5fbc6c1c2091
3
  size 4939116424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34a6842492517bd5147607d339f469a2810aa4ef18c9520443ca11d3c43380f3
3
  size 4939116424
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9aaf261990d4a881df076ec2e3b70271760eb1d4650860e21e7dac23d6253532
3
  size 4947390880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d5a9b4e1a52b8244e7ca6922b06cc6e2e89dac0dc0236fb1b472bb2fe3c2d2a
3
  size 4947390880
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56834f893605fbd9b04fae4860c65ca40ad37951f73d2e824a2edba88a18cc12
3
  size 3590619888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:603cb88c567f7f6f78be3d4bef7e5dfd47d99a547b30001d846984fcc555a0d8
3
  size 3590619888
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "train_loss": 0.23439283587014406,
4
- "train_runtime": 5872.9997,
5
  "train_samples": 19761,
6
- "train_samples_per_second": 3.365,
7
- "train_steps_per_second": 0.053
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "train_loss": 0.0030385508506429234,
4
+ "train_runtime": 144.6456,
5
  "train_samples": 19761,
6
+ "train_samples_per_second": 136.617,
7
+ "train_steps_per_second": 2.136
8
  }
trainer_state.json CHANGED
@@ -525,10 +525,10 @@
525
  "epoch": 1.0,
526
  "step": 309,
527
  "total_flos": 0.0,
528
- "train_loss": 0.23439283587014406,
529
- "train_runtime": 5872.9997,
530
- "train_samples_per_second": 3.365,
531
- "train_steps_per_second": 0.053
532
  }
533
  ],
534
  "logging_steps": 10,
 
525
  "epoch": 1.0,
526
  "step": 309,
527
  "total_flos": 0.0,
528
+ "train_loss": 0.0030385508506429234,
529
+ "train_runtime": 144.6456,
530
+ "train_samples_per_second": 136.617,
531
+ "train_steps_per_second": 2.136
532
  }
533
  ],
534
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b89935d144b21f9eae1842dca6021f49e0972822d5cb86deaca42054492c5e32
3
  size 6264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ea0f89586293f82aba15efa6e181e34eada67564061d2a833ae7eb15e0b550a
3
  size 6264