EdBerg commited on
Commit
0a5d032
·
verified ·
1 Parent(s): 938bfaf

EdBerg/Baha_1A_cupper_in_gold

Browse files
README.md CHANGED
@@ -27,14 +27,14 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/jrtabletsms/huggingface/runs/cunjvc0i)
31
 
32
  This model was trained with SFT.
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.0
37
- - Transformers: 4.49.0
38
  - Pytorch: 2.5.1+cu124
39
  - Datasets: 3.3.2
40
  - Tokenizers: 0.21.0
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/harpermia882/huggingface/runs/yyxdqfqf)
31
 
32
  This model was trained with SFT.
33
 
34
  ### Framework versions
35
 
36
  - TRL: 0.12.0
37
+ - Transformers: 4.48.3
38
  - Pytorch: 2.5.1+cu124
39
  - Datasets: 3.3.2
40
  - Tokenizers: 0.21.0
adapter_config.json CHANGED
@@ -24,13 +24,13 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "k_proj",
28
- "v_proj",
29
- "o_proj",
30
- "q_proj",
31
  "down_proj",
 
 
 
32
  "up_proj",
33
- "gate_proj"
 
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
 
 
 
 
27
  "down_proj",
28
+ "gate_proj",
29
+ "q_proj",
30
+ "o_proj",
31
  "up_proj",
32
+ "k_proj",
33
+ "v_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b10a6566782afbd6728c6fb88be3fe801f60c042b64c7608ae4b8d58d08d4cd7
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:143355e2d637602f6780bd386f3ba0b53283ec5768961bd172b38a9c0ed24cc2
3
  size 335604696
runs/Mar11_15-24-19_c93113ea815d/events.out.tfevents.1741706660.c93113ea815d.366.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c2a49f8c80d185cd62246ea577b7dd860c9b4dc429dff13c59411d665ac5087
3
+ size 111512
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13574c77ca1d572525cfa7caac46cee99309100524dad568a7ef85ae383df39f
3
- size 17210018
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json CHANGED
@@ -2060,5 +2060,5 @@
2060
  ],
2061
  "model_max_length": 131072,
2062
  "pad_token": "<|eot_id|>",
2063
- "tokenizer_class": "PreTrainedTokenizer"
2064
  }
 
2060
  ],
2061
  "model_max_length": 131072,
2062
  "pad_token": "<|eot_id|>",
2063
+ "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f36bbd9d89e59542b8445818b109ae3b1304a0f3a976dc3a973f048f4af72c44
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1123a6cbbe374404a8c2cd451f808da466e1f022d6c117bb63740a9d4c694124
3
  size 5624