rathi2023 commited on
Commit
e7f13a4
1 Parent(s): ebc6407

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -1,9 +1,8 @@
1
  {
2
- "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "TheBloke/zephyr-7B-beta-GPTQ",
5
  "bias": "none",
6
- "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
@@ -13,11 +12,10 @@
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 16,
16
- "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
- "q_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
 
1
  {
 
2
  "auto_mapping": null,
3
+ "base_model_name_or_path": "TheBloke/zephyr-7B-alpha-GPTQ",
4
  "bias": "none",
5
+ "fan_in_fan_out": true,
6
  "inference_mode": true,
7
  "init_lora_weights": true,
8
  "layers_pattern": null,
 
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
  "r": 16,
 
15
  "revision": null,
16
  "target_modules": [
17
+ "q_proj",
18
+ "v_proj"
19
  ],
20
  "task_type": "CAUSAL_LM"
21
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f370663cd60042e1ff81f555d2dd1504b440d9d16c13c8d922345fda912f82
3
- size 27309386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1a31e3ea1677c35dbfde79d138f02f510d542ff8dd9164b1df7f7f35f28cb8f
3
+ size 27308941
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<s>": 1,
4
+ "<unk>": 0
5
+ }
special_tokens_map.json CHANGED
@@ -4,26 +4,8 @@
4
  "<s>",
5
  "</s>"
6
  ],
7
- "bos_token": {
8
- "content": "<s>",
9
- "lstrip": false,
10
- "normalized": false,
11
- "rstrip": false,
12
- "single_word": false
13
- },
14
- "eos_token": {
15
- "content": "</s>",
16
- "lstrip": false,
17
- "normalized": false,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
  "pad_token": "</s>",
22
- "unk_token": {
23
- "content": "<unk>",
24
- "lstrip": false,
25
- "normalized": false,
26
- "rstrip": false,
27
- "single_word": false
28
- }
29
  }
 
4
  "<s>",
5
  "</s>"
6
  ],
7
+ "bos_token": "<s>",
8
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
9
  "pad_token": "</s>",
10
+ "unk_token": "<unk>"
 
 
 
 
 
 
11
  }
tokenizer_config.json CHANGED
@@ -31,7 +31,6 @@
31
  "</s>"
32
  ],
33
  "bos_token": "<s>",
34
- "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "</s>",
37
  "legacy": true,
 
31
  "</s>"
32
  ],
33
  "bos_token": "<s>",
 
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
36
  "legacy": true,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7d542a42a7c4b9d0af70c01c106dc95980649f138df61c13cf4730e376240a6
3
- size 4536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca910b11804dde54e7509f592e4df53ec6dd8a88ffc4f1aa1d7e845c250e45b9
3
+ size 4091