semran1 commited on
Commit
95baacb
·
verified ·
1 Parent(s): c543b52

Upload folder using huggingface_hub

Browse files
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9023aaf6076a1273038a94324c5a5d937fedc7b190a3d0047f5da522414da95
3
  size 4965799096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b594353ab63527420015e922d5b183eedca772fd537493519c0f9670f681b10
3
  size 4965799096
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:783981dd41ffa293477156fd3b002f625e649c5211f7b010c9a681a7a14ce957
3
  size 2247734992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1584bdafefe77d92afd776bc37ad189100c155571de707ebaa495c811d6f707
3
  size 2247734992
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:186d020555b67bb109bbc2fd4b592c8b8ebef786395f57fa503e3ee648dd2668
3
  size 14427231362
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5bbed1d27f43f70856deb2c884d1a19f58a75d0bdd5c62f5455d1e24f362058
3
  size 14427231362
special_tokens_map.json CHANGED
@@ -12,5 +12,6 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
- }
 
16
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
+ },
16
+ "pad_token": "<|end_of_text|>"
17
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79e3e522635f3171300913bb421464a87de6222182a0570b9b2ccba2a964b2b4
3
- size 9085657
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c85066e7642934ed09b44155e6566b0b5dab2637fb9433439ba5c9c7f8b50d3
3
+ size 17210018
tokenizer_config.json CHANGED
@@ -2050,12 +2050,15 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
 
2053
  "clean_up_tokenization_spaces": true,
2054
  "eos_token": "<|end_of_text|>",
 
2055
  "model_input_names": [
2056
  "input_ids",
2057
  "attention_mask"
2058
  ],
2059
  "model_max_length": 131072,
 
2060
  "tokenizer_class": "PreTrainedTokenizerFast"
2061
  }
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|end_of_text|>",
2056
+ "extra_special_tokens": {},
2057
  "model_input_names": [
2058
  "input_ids",
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 131072,
2062
+ "pad_token": "<|end_of_text|>",
2063
  "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }