sengi commited on
Commit
f06ceba
1 Parent(s): 6ecd899

Training in progress, step 100

Browse files
adapter_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-v0.1",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "loftq_config": {},
12
+ "lora_alpha": 128,
13
+ "lora_dropout": 0.05,
14
+ "megatron_config": null,
15
+ "megatron_core": "megatron.core",
16
+ "modules_to_save": null,
17
+ "peft_type": "LORA",
18
+ "r": 128,
19
+ "rank_pattern": {},
20
+ "revision": null,
21
+ "target_modules": [
22
+ "up_proj",
23
+ "q_proj",
24
+ "v_proj",
25
+ "o_proj",
26
+ "down_proj",
27
+ "gate_proj",
28
+ "k_proj"
29
+ ],
30
+ "task_type": "CAUSAL_LM"
31
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e44ce263e6fd885f50d82ca515b9325375b43ee36ededb75acf161ce88bc2e41
3
+ size 48
runs/Mar06_22-59-48_g3034/events.out.tfevents.1709794860.g3034.55555.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78782d745d08e1ed93017bc952c2823b8b513d811d67abbfa8662d3b3626fdcf
3
+ size 4754
runs/Mar06_23-11-56_g3034/events.out.tfevents.1709795613.g3034.71721.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4b2c9bb6f87f9e4f1ea0292cf4ac3b97fbc087c955886124cddfce4800d28ca
3
+ size 4754
runs/Mar06_23-15-12_g3034/events.out.tfevents.1709795783.g3034.75513.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9510f4b3aeabf10b1d8fc1610d8d1c355df1d9bedfd9de0526aaa859a04ac6
3
+ size 4754
runs/Mar07_00-07-18_g3034/events.out.tfevents.1709799093.g3034.37435.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63806b8ffbb82114d87a0eec855c63ae324a9e354122da453cf74bfe85ab489e
3
+ size 5196
runs/Mar07_00-24-06_g3022/events.out.tfevents.1709800125.g3022.52852.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07d0e764674c2c9601ef5220245f56c9aa48a46fb9b85006b81644c169439906
3
+ size 5196
runs/Mar07_10-06-31_g3023/events.out.tfevents.1709834966.g3023.41230.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e753705b19b0ca07ed2c54e4b9d7933c459aad5cdfc0c083f3a33112a33f869f
3
+ size 5195
runs/Mar07_10-14-19_g3023/events.out.tfevents.1709835334.g3023.48210.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54ed9a3a7985b87f6644877f3cbc8220c16aaaec72fc219b8f1dff743ec1413c
3
+ size 4753
runs/Mar07_12-51-52_g3022/events.out.tfevents.1709845251.g3022.23987.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a134ceb995043026187804cd62cb90873a75a6417a182d75d0c08aa94ae03af8
3
+ size 5195
runs/Mar07_13-04-04_g3022/events.out.tfevents.1709845966.g3022.32730.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8d5ac1095efcee7fda098647cf1aeea5538ce306e5ca37440429a6675500329
3
+ size 5195
runs/Mar07_13-18-10_g3022/events.out.tfevents.1709846819.g3022.41751.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a604f57232aa9bb00072d797bf7ab2b1bbf1b4870a9c36c5935265d9d70d14
3
+ size 5195
runs/Mar07_13-51-11_g3022/events.out.tfevents.1709848348.g3022.61000.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:425fd5460abef42b5f0e641659e66bda136543ce271590a34119ea24d8a9d481
3
+ size 5195
runs/Mar07_14-15-44_g3022/events.out.tfevents.1709850189.g3022.76081.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99ec1d6f93926069598252a9737c717e9460af0a53ec397539b066e016dd5e4a
3
+ size 5195
runs/Mar08_11-58-43_g3022/events.out.tfevents.1709928013.g3022.49666.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e451ca8b153ba1cf6cea0dac6b00a8d3b3b425f4d285c6417de5ff67f7c518b
3
+ size 4753
runs/Mar08_12-11-46_g3022/events.out.tfevents.1709928931.g3022.56950.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b53f78aea4893e050c9ca02772e5ccc9ad6dc85c2936b25783d4105e55d4b7
3
+ size 4753
runs/Mar08_15-20-15_g3010/events.out.tfevents.1709940095.g3010.21285.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bca8d1a475bcf2a20f81befd74a88ba5ed5b1c24546b0cc9616dbeace8415bfb
3
+ size 6105
runs/Mar08_15-40-26_g3010/events.out.tfevents.1709941483.g3010.44948.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:525caaa11a5ec354ddcb3f73bafc3a25792a7544d8b3434c870d546bafb0a253
3
+ size 12916
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": true,
36
+ "model_max_length": 2048,
37
+ "pad_token": "</s>",
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a3ba19f4b5785ca3b7c71eb9977a50cd94fe004cf7ee23b36269feed1b38a00
3
+ size 6520