meetrathi97 commited on
Commit
93ff25d
·
verified ·
1 Parent(s): 72553a2

End of training

Browse files
README.md CHANGED
@@ -27,17 +27,18 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
 
30
 
31
 
32
  This model was trained with SFT.
33
 
34
  ### Framework versions
35
 
36
- - TRL: 0.12.2
37
- - Transformers: 4.46.3
38
  - Pytorch: 2.5.1+cu121
39
  - Datasets: 3.2.0
40
- - Tokenizers: 0.20.3
41
 
42
  ## Citations
43
 
 
27
 
28
  ## Training procedure
29
 
30
+
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.13.0
38
+ - Transformers: 4.47.1
39
  - Pytorch: 2.5.1+cu121
40
  - Datasets: 3.2.0
41
+ - Tokenizers: 0.21.0
42
 
43
  ## Citations
44
 
adapter_config.json CHANGED
@@ -3,6 +3,8 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "meta-llama/Llama-3.2-3B",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
@@ -11,6 +13,7 @@
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 64,
 
14
  "lora_dropout": 0.1,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
@@ -20,10 +23,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
  "o_proj",
25
  "v_proj",
26
- "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "meta-llama/Llama-3.2-3B",
5
  "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 64,
16
+ "lora_bias": false,
17
  "lora_dropout": 0.1,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
  "o_proj",
28
  "v_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f347f3db07578c68c18a72614ac3bae1c3b66754bbb6388673bd7a8c3170a72
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b02e62f6165d9ab598f6f361e9bbf455cc83738d2c8b2a82c2106b80b8f86fac
3
  size 36730224
runs/Dec25_20-37-10_192eac3a7992/events.out.tfevents.1735159270.192eac3a7992.568.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:267805afeccad04f34d63f2dc4ff2829b4b4706e56c1b5320458e63e6c8f794d
3
+ size 16653
runs/Dec25_21-00-46_192eac3a7992/events.out.tfevents.1735160681.192eac3a7992.568.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33c764d1e71ddc587aab0180644395672dda5043c9cf532649b9ef625b82de7e
3
+ size 111442
tokenizer_config.json CHANGED
@@ -2052,6 +2052,7 @@
2052
  "bos_token": "<|begin_of_text|>",
2053
  "clean_up_tokenization_spaces": true,
2054
  "eos_token": "<|end_of_text|>",
 
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
 
2052
  "bos_token": "<|begin_of_text|>",
2053
  "clean_up_tokenization_spaces": true,
2054
  "eos_token": "<|end_of_text|>",
2055
+ "extra_special_tokens": {},
2056
  "model_input_names": [
2057
  "input_ids",
2058
  "attention_mask"
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98ec612874ca34750e3a27345a41d46becd0fec48ec1a83f24676ca11f056672
3
- size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef0b4fbb33c70ac0ad87f565705ca01f2c24ba341dc3fc349e50d17cd404b7d4
3
+ size 5624