Adil1567 commited on
Commit
21a6b6d
·
verified ·
1 Parent(s): 6b4e90c

Model save

Browse files
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 1.0178
22
 
23
  ## Model description
24
 
@@ -53,7 +53,7 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|
56
- | 0.9809 | 1.0 | 20 | 1.0178 |
57
 
58
 
59
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [meta-llama/Llama-3.1-70B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-70B-Instruct) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6089
22
 
23
  ## Model description
24
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|
56
+ | 0.6126 | 1.0 | 200 | 0.6089 |
57
 
58
 
59
  ### Framework versions
adapter_config.json CHANGED
@@ -23,13 +23,13 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "v_proj",
27
  "o_proj",
28
  "q_proj",
29
- "k_proj",
30
  "down_proj",
31
  "up_proj",
32
- "gate_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "gate_proj",
27
  "v_proj",
28
  "o_proj",
29
  "q_proj",
 
30
  "down_proj",
31
  "up_proj",
32
+ "k_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3af779e366c638a194f86a7c739fa2410c3b76dd8c4205710484d9f4b88d3599
3
  size 414337624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac66d622181fcbc5e33f12553515eeaf7cd109b0e3ffcf59bc482977c1c0aac3
3
  size 414337624
runs/Jan04_19-46-04_gpu-server/events.out.tfevents.1736020144.gpu-server.2012116.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8181d3b2c899e88ed75207cc2c2ef82aa177cbf51d3a9ba69636ed69bd537e5
3
- size 13982
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed95fbc0b684887f501e48372a7cec9e604ef9505926f7bd600c97fed2c32071
3
+ size 14607