YYYYYYibo commited on
Commit
b03cb12
1 Parent(s): 96a9f4c

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "mistralai/Mistral-7B-v0.1",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -19,11 +19,11 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "k_proj",
23
  "gate_proj",
24
  "v_proj",
25
  "down_proj",
26
- "o_proj",
27
  "up_proj",
28
  "q_proj"
29
  ],
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "alignment-handbook/zephyr-7b-sft-full",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "o_proj",
23
  "k_proj",
24
  "gate_proj",
25
  "v_proj",
26
  "down_proj",
 
27
  "up_proj",
28
  "q_proj"
29
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4d9bf9987226be438eb58d72b399060d4e34e3c069ffd40fe4304f388bddebf
3
  size 671150064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f87f55ddc463dfe97b79100617ebc334333146776dd3d46f370acf70aa84d10
3
  size 671150064
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a135e6d79034267e47a5f6bac112759419a4e5e36f186ec21020a7bd9b34bf14
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39194ec5d1fee52fe6ef42b72b41c64d96c0b552b0a56d6f3005a15ee52d3d7c
3
  size 4920