tomrb commited on
Commit
220bb24
·
verified ·
1 Parent(s): 838636a

Training in progress, step 100

Browse files
adapter_config.json CHANGED
@@ -21,9 +21,9 @@
21
  "revision": null,
22
  "target_modules": [
23
  "o_proj",
24
- "v_proj",
25
  "q_proj",
26
- "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
21
  "revision": null,
22
  "target_modules": [
23
  "o_proj",
24
+ "k_proj",
25
  "q_proj",
26
+ "v_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cdaac026dd535edf5358d17b052d0cfbded5333d52d9288413772e445716abe7
3
  size 18893616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:394c06f6148f27735d5e62ee0ee1979a2ba6aa87859a71c473ef156e93dfe00e
3
  size 18893616
tokenizer_config.json CHANGED
@@ -46,7 +46,7 @@
46
  "cls_token": "[CLS]",
47
  "eos_token": "[SEP]",
48
  "mask_token": "[MASK]",
49
- "model_max_length": 1000000000000000019884624838656,
50
  "pad_token": "[PAD]",
51
  "sep_token": "[SEP]",
52
  "tokenizer_class": "PreTrainedTokenizerFast",
 
46
  "cls_token": "[CLS]",
47
  "eos_token": "[SEP]",
48
  "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
  "pad_token": "[PAD]",
51
  "sep_token": "[SEP]",
52
  "tokenizer_class": "PreTrainedTokenizerFast",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bc157820ea8e71249c6e81853e9ad3af4c38a702b5d1cebb110b6df5e4e1fd8
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10f71107cf5f5a08c4340435a84536e2f5faed2f2db9b42e7d10e1f6b7669ce5
3
  size 5176