uine commited on
Commit
e6129fa
·
verified ·
1 Parent(s): 794d15a

End of training

Browse files
README.md CHANGED
@@ -35,12 +35,12 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 0.0003
39
- - train_batch_size: 1
40
  - eval_batch_size: 8
41
  - seed: 42
42
- - gradient_accumulation_steps: 2
43
- - total_train_batch_size: 2
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 5e-05
39
+ - train_batch_size: 2
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 8
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
adapter_config.json CHANGED
@@ -9,13 +9,13 @@
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
- "lora_alpha": 16,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
- "r": 16,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 128,
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 256,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07c7da2aaa7395b9546939aba64c0f3c0fa8c526db71790465458a7f56d5cc02
3
- size 84972816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd04306d3fc4122de7a33d8c6d774d43473d80d1ee2e9514f619d55bccd231b5
3
+ size 1358993696
runs/May12_16-56-55_d3d5113b4054/events.out.tfevents.1715533093.d3d5113b4054.3173.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52337c0db0a970425ca174f0386a486cc22a2e9f92bf0387cd099ccf2cec0f7d
3
+ size 170522
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c75c59987b0229b9a7b60dfa28eb45c374234e885772a7098e050400bd303d0
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86146dee46cb197e09f8c736f9dd4112a1e96f57d53d0312e0e732c82d0fc162
3
  size 4728