timlrx-resaro commited on
Commit
711c235
1 Parent(s): 8e33a9f

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,12 +1,16 @@
1
  {
2
- "_name_or_path": "unsloth/Meta-Llama-3.1-8B-bnb-4bit",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
- "eos_token_id": 128001,
 
 
 
 
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
@@ -17,7 +21,7 @@
17
  "num_attention_heads": 32,
18
  "num_hidden_layers": 32,
19
  "num_key_value_heads": 8,
20
- "pad_token_id": 128004,
21
  "pretraining_tp": 1,
22
  "rms_norm_eps": 1e-05,
23
  "rope_scaling": {
 
1
  {
2
+ "_name_or_path": "mlabonne/Meta-Llama-3.1-8B-Instruct-abliterated",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": [
10
+ 128001,
11
+ 128008,
12
+ 128009
13
+ ],
14
  "hidden_act": "silu",
15
  "hidden_size": 4096,
16
  "initializer_range": 0.02,
 
21
  "num_attention_heads": 32,
22
  "num_hidden_layers": 32,
23
  "num_key_value_heads": 8,
24
+ "pad_token_id": 128255,
25
  "pretraining_tp": 1,
26
  "rms_norm_eps": 1e-05,
27
  "rope_scaling": {
generation_config.json CHANGED
@@ -3,9 +3,9 @@
3
  "do_sample": true,
4
  "eos_token_id": [
5
  128001,
6
- 128008,
7
  128009
8
  ],
 
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
  "transformers_version": "4.43.3"
 
3
  "do_sample": true,
4
  "eos_token_id": [
5
  128001,
 
6
  128009
7
  ],
8
+ "max_length": 4096,
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
  "transformers_version": "4.43.3"
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9f4038149d776b1978ff3394733b9ae8de630796bf10f098464c51614036ce55
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbfd68c9dec5bfd73b9bf92f6c8db944879d6278ba736c250ae00046bd588923
3
  size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4791437c1fb2bdce039f1485d3ac897bbfb0e67d867f6ff4461d03bc9f4280ad
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:673b53ef5f58596cbf101f5b9f904f9cb86b2524b01df45d6e69a97a0c9d1810
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61e787b6501be145be192f70c8c86586d2b334f1ff155b99cd8fc91c645d2885
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98128ee4b5acd8d8a93b3ee04b549d929313f2f37bf3dfff495de0367013342b
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed9d8dc6b934adab4cbe65fc28ecb38e099e1ea736e6e3a5b2b6ed03f4d39854
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:909a71c9e5505950c3e036e1a2f10258ccee5c43c83d72cefb217f9d2e0cfa82
3
  size 1168138808
special_tokens_map.json CHANGED
@@ -1,5 +1,17 @@
1
  {
2
- "bos_token": "<|begin_of_text|>",
3
- "eos_token": "<|im_end|>",
4
- "pad_token": "<|finetune_right_pad_id|>"
 
 
 
 
 
 
 
 
 
 
 
 
5
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|eot_id|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<|reserved_special_token_247|>"
17
  }
tokenizer.json CHANGED
@@ -14,7 +14,7 @@
14
  },
15
  {
16
  "id": 128001,
17
- "content": "<|im_end|>",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
 
14
  },
15
  {
16
  "id": 128001,
17
+ "content": "<|end_of_text|>",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -9,7 +9,7 @@
9
  "special": true
10
  },
11
  "128001": {
12
- "content": "<|im_end|>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
@@ -2050,11 +2050,15 @@
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
- "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'<|im_start|>user\n' + message['content'] + '<|im_end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['content'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2054
  "clean_up_tokenization_spaces": true,
2055
- "eos_token": "<|im_end|>",
2056
- "model_max_length": 1000000000000000019884624838656,
2057
- "pad_token": "<|finetune_right_pad_id|>",
2058
- "tokenizer_class": "PreTrainedTokenizerFast",
2059
- "unk_token": null
 
 
 
 
2060
  }
 
9
  "special": true
10
  },
11
  "128001": {
12
+ "content": "<|end_of_text|>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
 
2050
  }
2051
  },
2052
  "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}",
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|eot_id|>",
2056
+ "model_input_names": [
2057
+ "input_ids",
2058
+ "attention_mask"
2059
+ ],
2060
+ "model_max_length": 131072,
2061
+ "pad_token": "<|reserved_special_token_247|>",
2062
+ "padding_side": "left",
2063
+ "tokenizer_class": "PreTrainedTokenizerFast"
2064
  }