_id
stringlengths 24
24
| id
stringlengths 14
79
| author
stringlengths 3
21
| gated
stringclasses 3
values | inference
stringclasses 9
values | lastModified
stringlengths 24
24
| likes
int64 0
4.13k
| private
bool 1
class | sha
stringlengths 40
40
| config
stringlengths 23
5.92k
⌀ | downloads
int64 1.8k
3.13M
| tags
listlengths 5
62
| pipeline_tag
stringclasses 6
values | library_name
stringclasses 4
values | createdAt
stringlengths 24
24
| modelId
stringlengths 14
79
| siblings
listlengths 4
285
| model_type
stringclasses 13
values | base_model
stringlengths 13
43
| base_model_downloads
float64 98
28.9k
| base_model_pipeline
stringclasses 1
value | base_model_children_count
float64 98
28.9k
| adapter
float64 0
377
| merge
float64 0
94
| finetune
float64 0
263
| quantized
float64 0
182
| __index_level_0__
int64 86k
992k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
65d6acbc29af34543aeeb91d
|
mlc-ai/gemma-2b-it-q4f16_1-MLC
|
mlc-ai
|
False
|
pipeline-not-detected
|
2024-08-01T15:18:31.000Z
| 4 | false |
d7ecc5f26ff1d62dc4e69bd7bfd41fe26d9f51e2
|
{"tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 6,964 |
[
"mlc-llm",
"web-llm",
"base_model:google/gemma-2b-it",
"base_model:quantized:google/gemma-2b-it",
"region:us"
] | null |
mlc-llm
|
2024-02-22T02:09:00.000Z
|
mlc-ai/gemma-2b-it-q4f16_1-MLC
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "logs.txt"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | null |
google/gemma-2b-it
| 576 |
text-generation
| 576 | 0 | 0 | 0 | 0 | 502,646 |
659f8757d2e705b3fb660a96
|
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
|
NousResearch
|
False
|
warm
|
2024-04-30T18:09:51.000Z
| 412 | false |
286ae6737d048ad1d965c2e830864df02db50f2f
|
{"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 6,955 |
[
"transformers",
"safetensors",
"mixtral",
"text-generation",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:finetune:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-01-11T06:14:47.000Z
|
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00019.safetensors"
},
{
"rfilename": "model-00002-of-00019.safetensors"
},
{
"rfilename": "model-00003-of-00019.safetensors"
},
{
"rfilename": "model-00004-of-00019.safetensors"
},
{
"rfilename": "model-00005-of-00019.safetensors"
},
{
"rfilename": "model-00006-of-00019.safetensors"
},
{
"rfilename": "model-00007-of-00019.safetensors"
},
{
"rfilename": "model-00008-of-00019.safetensors"
},
{
"rfilename": "model-00009-of-00019.safetensors"
},
{
"rfilename": "model-00010-of-00019.safetensors"
},
{
"rfilename": "model-00011-of-00019.safetensors"
},
{
"rfilename": "model-00012-of-00019.safetensors"
},
{
"rfilename": "model-00013-of-00019.safetensors"
},
{
"rfilename": "model-00014-of-00019.safetensors"
},
{
"rfilename": "model-00015-of-00019.safetensors"
},
{
"rfilename": "model-00016-of-00019.safetensors"
},
{
"rfilename": "model-00017-of-00019.safetensors"
},
{
"rfilename": "model-00018-of-00019.safetensors"
},
{
"rfilename": "model-00019-of-00019.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "transformers_inference_example.py"
}
] |
mixtral
|
mistralai/Mixtral-8x7B-v0.1
| 182 |
text-generation
| 182 | 0 | 5 | 3 | 9 | 448,581 |
653bb81b7f5fc9ccb1617aa3
|
TheBloke/zephyr-7B-beta-GGUF
|
TheBloke
|
False
|
explicit-opt-out
|
2023-10-27T14:56:25.000Z
| 222 | false |
e4714d14e9652aa9658fa937732cceadc63ac42e
|
{"model_type": "mistral"}
| 6,913 |
[
"transformers",
"gguf",
"mistral",
"generated_from_trainer",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:quantized:HuggingFaceH4/zephyr-7b-beta",
"license:mit",
"text-generation-inference",
"region:us"
] | null |
transformers
|
2023-10-27T13:16:11.000Z
|
TheBloke/zephyr-7B-beta-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "zephyr-7b-beta.Q2_K.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_L.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_0.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_0.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q6_K.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q8_0.gguf"
}
] |
mistral
|
HuggingFaceH4/zephyr-7b-beta
| 439 |
text-generation
| 439 | 0 | 0 | 0 | 0 | 358,502 |
6580aa20419afba19a692cc8
|
TheBloke/phi-2-GGUF
|
TheBloke
|
False
|
explicit-opt-out
|
2023-12-18T20:25:44.000Z
| 190 | false |
5a454d977c6438bb9fb2df233c8ca70f21c87420
|
{"model_type": "phi-msft"}
| 6,843 |
[
"transformers",
"gguf",
"phi-msft",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/phi-2",
"base_model:quantized:microsoft/phi-2",
"license:other",
"region:us"
] |
text-generation
|
transformers
|
2023-12-18T20:22:56.000Z
|
TheBloke/phi-2-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "phi-2.Q2_K.gguf"
},
{
"rfilename": "phi-2.Q3_K_L.gguf"
},
{
"rfilename": "phi-2.Q3_K_M.gguf"
},
{
"rfilename": "phi-2.Q3_K_S.gguf"
},
{
"rfilename": "phi-2.Q4_0.gguf"
},
{
"rfilename": "phi-2.Q4_K_M.gguf"
},
{
"rfilename": "phi-2.Q4_K_S.gguf"
},
{
"rfilename": "phi-2.Q5_0.gguf"
},
{
"rfilename": "phi-2.Q5_K_M.gguf"
},
{
"rfilename": "phi-2.Q5_K_S.gguf"
},
{
"rfilename": "phi-2.Q6_K.gguf"
},
{
"rfilename": "phi-2.Q8_0.gguf"
}
] |
phi-msft
|
microsoft/phi-2
| 922 |
text-generation
| 922 | 0 | 0 | 0 | 0 | 423,143 |
653bb81b8502e1fd87503201
|
TheBloke/zephyr-7B-beta-GPTQ
|
TheBloke
|
False
|
explicit-opt-out
|
2023-10-27T14:56:24.000Z
| 56 | false |
8128029fba795c423004d08695fdda8491289748
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
| 6,572 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:quantized:HuggingFaceH4/zephyr-7b-beta",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] |
text-generation
|
transformers
|
2023-10-27T13:16:11.000Z
|
TheBloke/zephyr-7B-beta-GPTQ
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
}
] |
mistral
|
HuggingFaceH4/zephyr-7b-beta
| 439 |
text-generation
| 439 | 39 | 0 | 18 | 0 | 358,503 |
6644446bc3d4867f3a4d246c
|
neuralmagic/Llama-2-7b-gsm8k
|
neuralmagic
|
False
|
not-popular-enough
|
2024-06-20T09:17:42.000Z
| 0 | false |
1d435c94098a9407b0cf12f7452631d243823340
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
| 6,528 |
[
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"dataset:openai/gsm8k",
"arxiv:2405.03594",
"arxiv:2110.14168",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-15T05:13:15.000Z
|
neuralmagic/Llama-2-7b-gsm8k
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00003.bin"
},
{
"rfilename": "pytorch_model-00002-of-00003.bin"
},
{
"rfilename": "pytorch_model-00003-of-00003.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-2-7b-hf
| 1,595 |
text-generation
| 1,595 | 0 | 0 | 0 | 0 | 639,403 |
6642c636f8149cb6237b6f57
|
Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3
|
Kukedlc
|
False
|
not-popular-enough
|
2024-09-22T19:08:57.000Z
| 0 | false |
8945b7f810a793dffb3b66f959ab66c8f42dbd01
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}}
| 6,466 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"dataset:mlabonne/orpo-dpo-mix-40k",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-14T02:02:30.000Z
|
Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3
|
[
{
"rfilename": ".config/.last_opt_in_prompt.yaml"
},
{
"rfilename": ".config/.last_survey_prompt.yaml"
},
{
"rfilename": ".config/.last_update_check.json"
},
{
"rfilename": ".config/active_config"
},
{
"rfilename": ".config/config_sentinel"
},
{
"rfilename": ".config/configurations/config_default"
},
{
"rfilename": ".config/default_configs.db"
},
{
"rfilename": ".config/gce"
},
{
"rfilename": ".config/logs/2024.05.23/13.23.45.334202.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.10.458450.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.22.323397.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.32.030423.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.43.927079.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.44.609427.log"
},
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "sample_data/README.md"
},
{
"rfilename": "sample_data/anscombe.json"
},
{
"rfilename": "sample_data/california_housing_test.csv"
},
{
"rfilename": "sample_data/california_housing_train.csv"
},
{
"rfilename": "sample_data/mnist_test.csv"
},
{
"rfilename": "sample_data/mnist_train_small.csv"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-3.1-8B-Instruct
| 695 |
text-generation
| 695 | 0 | 0 | 0 | 0 | 637,407 |
66c87db20d38f15892b5aade
|
anakin87/Phi-3.5-mini-ITA
|
anakin87
|
False
|
not-popular-enough
|
2024-09-02T21:34:04.000Z
| 9 | false |
eb9c2869b8c92ad495d589a4d3c910aefb74f87e
|
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3.5-mini-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3.5-mini-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 6,464 |
[
"transformers",
"safetensors",
"phi3",
"text-generation",
"trl",
"spectrum",
"conversational",
"custom_code",
"it",
"en",
"dataset:mlabonne/FineTome-100k",
"dataset:efederici/capybara-claude-15k-ita",
"arxiv:2406.06623",
"base_model:microsoft/Phi-3.5-mini-instruct",
"base_model:finetune:microsoft/Phi-3.5-mini-instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-08-23T12:16:50.000Z
|
anakin87/Phi-3.5-mini-ITA
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "assets/phi_35_mini_ita.png"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "notebooks/training.ipynb"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
phi3
|
microsoft/Phi-3.5-mini-instruct
| 147 |
text-generation
| 147 | 0 | 0 | 0 | 1 | 852,190 |
66e1fc522c252209c2756a43
|
silent666/01-ai-Yi-1.5-9B-1726086226
|
silent666
|
False
|
pipeline-not-detected
|
2024-09-11T20:24:41.000Z
| 0 | false |
b0b156299e33a8cd90dedd020e15975f3829774e
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 6,239 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null |
peft
|
2024-09-11T20:23:46.000Z
|
silent666/01-ai-Yi-1.5-9B-1726086226
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 931,657 |
66cde4fc3f233bf2c7c7e2b3
|
garak-llm/artgpt2tox
|
garak-llm
|
False
|
library-not-detected
|
2024-08-27T15:12:43.000Z
| 0 | false |
31fcab3082a612c844aad1662ee74ac27400dcf1
|
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
| 6,022 |
[
"safetensors",
"gpt2",
"en",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"region:us"
] | null | null |
2024-08-27T14:38:52.000Z
|
garak-llm/artgpt2tox
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] |
gpt2
|
openai-community/gpt2
| 2,686 |
text-generation
| 2,686 | 0 | 0 | 0 | 0 | 866,425 |
6628e4587ef8a073dda15122
|
jondurbin/bagel-8b-v1.0
|
jondurbin
|
False
|
not-popular-enough
|
2024-05-12T20:48:33.000Z
| 48 | false |
b7f4d46d8a935623a47c9d6b3845443f20413a5f
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}}
| 5,999 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"bagel",
"conversational",
"dataset:ai2_arc",
"dataset:allenai/ultrafeedback_binarized_cleaned",
"dataset:argilla/distilabel-intel-orca-dpo-pairs",
"dataset:jondurbin/airoboros-3.2",
"dataset:codeparrot/apps",
"dataset:facebook/belebele",
"dataset:bluemoon-fandom-1-1-rp-cleaned",
"dataset:boolq",
"dataset:camel-ai/biology",
"dataset:camel-ai/chemistry",
"dataset:camel-ai/math",
"dataset:camel-ai/physics",
"dataset:jondurbin/contextual-dpo-v0.1",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:jondurbin/py-dpo-v0.1",
"dataset:jondurbin/truthy-dpo-v0.1",
"dataset:LDJnr/Capybara",
"dataset:jondurbin/cinematika-v0.1",
"dataset:WizardLM/WizardLM_evol_instruct_70k",
"dataset:glaiveai/glaive-function-calling-v2",
"dataset:grimulkan/LimaRP-augmented",
"dataset:lmsys/lmsys-chat-1m",
"dataset:ParisNeo/lollms_aware_dataset",
"dataset:TIGER-Lab/MathInstruct",
"dataset:Muennighoff/natural-instructions",
"dataset:openbookqa",
"dataset:kingbri/PIPPA-shareGPT",
"dataset:piqa",
"dataset:Vezora/Tested-22k-Python-Alpaca",
"dataset:ropes",
"dataset:cakiki/rosetta-code",
"dataset:Open-Orca/SlimOrca",
"dataset:b-mc2/sql-create-context",
"dataset:squad_v2",
"dataset:mattpscott/airoboros-summarization",
"dataset:migtissera/Synthia-v1.3",
"dataset:unalignment/toxic-dpo-v0.2",
"dataset:WhiteRabbitNeo/WRN-Chapter-1",
"dataset:WhiteRabbitNeo/WRN-Chapter-2",
"dataset:winogrande",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-24T10:52:08.000Z
|
jondurbin/bagel-8b-v1.0
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "bagel.png"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 13 | 0 | 3 | 602,738 |
669932464ab3500b30940ff5
|
bartowski/Mistral-Nemo-Instruct-2407-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-08-28T20:16:33.000Z
| 31 | false |
e9cdc9d71317c0911875031d1c22f6d9231b6715
| null | 5,943 |
[
"gguf",
"text-generation",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ru",
"zh",
"ja",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"region:us"
] |
text-generation
| null |
2024-07-18T15:18:30.000Z
|
bartowski/Mistral-Nemo-Instruct-2407-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ2_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ3_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ3_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ4_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q2_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q2_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_XL.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_4.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_8.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_8_8.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q6_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q6_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q8_0.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-f16.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-f32.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
mistralai/Mistral-Nemo-Instruct-2407
| 101 |
text-generation
| 101 | 0 | 0 | 0 | 0 | 761,505 |
662f699466debc4f8e1df110
|
swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA
|
swap-uniba
|
False
|
not-popular-enough
|
2024-07-09T09:03:55.000Z
| 22 | false |
2b6e46e4c9d341dc8bf8350a167492c880116b66
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
| 5,924 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pythorch",
"llama-3",
"llamantino",
"conversational",
"en",
"it",
"dataset:gsarti/clean_mc4_it",
"dataset:Chat-Error/wizard_alpaca_dolly_orca",
"dataset:mlabonne/orpo-dpo-mix-40k",
"arxiv:2405.07101",
"arxiv:2312.09993",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-29T09:34:12.000Z
|
swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA
|
[
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 5 | 2 | 0 | 611,101 |
663db3c285a3bbab3554b789
|
cognitivecomputations/dolphin-2.9.1-llama-3-8b
|
cognitivecomputations
|
False
|
not-popular-enough
|
2024-05-20T14:36:52.000Z
| 36 | false |
924427715104c3667868d9297e4069b4c62bfd88
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}}
| 5,905 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-10T05:42:26.000Z
|
cognitivecomputations/dolphin-2.9.1-llama-3-8b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 3 | 5 | 4 | 630,642 |
66d7dac66ff4d32337e4fa66
|
Bllossom/llama-3.1-Korean-Bllossom-Vision-8B
|
Bllossom
|
False
|
pipeline-not-detected
|
2024-09-11T01:42:35.000Z
| 86 | false |
fdc00628f5624b5671db2bf09024545dac5bd3b3
|
{"architectures": ["LlavaNextForConditionalGeneration"], "model_type": "llava_next", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}\n{% set loop_messages = messages %}\n{% set default_system_prompt = \"You are a helpful AI assistant. Please answer the user's questions kindly. \ub2f9\uc2e0\uc740 \uc720\ub2a5\ud55c AI \uc5b4\uc2dc\uc2a4\ud134\ud2b8 \uc785\ub2c8\ub2e4. \uc0ac\uc6a9\uc790\uc758 \uc9c8\ubb38\uc5d0 \ub300\ud574 \uce5c\uc808\ud558\uac8c \ub2f5\ubcc0\ud574\uc8fc\uc138\uc694.\" %}\n{% if not loop_messages | selectattr('role', 'equalto', 'system') | list %}\n{% set system_message = {'role': 'system', 'content': default_system_prompt} %}\n{% set loop_messages = [system_message] + loop_messages %}\n{% endif %}\n{% for message in loop_messages %}\n{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n' + message['content'] | trim + '<|eot_id|>' %}\n{% if loop.index0 == 0 %}\n{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}}
| 5,879 |
[
"transformers",
"safetensors",
"llava_next",
"pretraining",
"en",
"ko",
"arxiv:2403.10882",
"arxiv:2403.11399",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"license:llama3.1",
"endpoints_compatible",
"region:us"
] | null |
transformers
|
2024-09-04T03:57:58.000Z
|
Bllossom/llama-3.1-Korean-Bllossom-Vision-8B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "preprocessor_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llava_next
|
meta-llama/Llama-3.1-8B
| 254 |
text-generation
| 254 | 2 | 0 | 0 | 0 | 895,159 |
66c49cefea476bea050b3140
|
XSCP/Qwen-Qwen1.5-7B-1724161263
|
XSCP
|
False
|
pipeline-not-detected
|
2024-08-20T13:41:06.000Z
| 0 | false |
624fa1791d14ac34c77adeed1dd81025efa5129c
|
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}}
| 5,857 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null |
peft
|
2024-08-20T13:41:03.000Z
|
XSCP/Qwen-Qwen1.5-7B-1724161263
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "checkpoint-26/README.md"
},
{
"rfilename": "checkpoint-26/adapter_config.json"
},
{
"rfilename": "checkpoint-26/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-26/added_tokens.json"
},
{
"rfilename": "checkpoint-26/merges.txt"
},
{
"rfilename": "checkpoint-26/optimizer.pt"
},
{
"rfilename": "checkpoint-26/rng_state.pth"
},
{
"rfilename": "checkpoint-26/scheduler.pt"
},
{
"rfilename": "checkpoint-26/special_tokens_map.json"
},
{
"rfilename": "checkpoint-26/tokenizer.json"
},
{
"rfilename": "checkpoint-26/tokenizer_config.json"
},
{
"rfilename": "checkpoint-26/trainer_state.json"
},
{
"rfilename": "checkpoint-26/training_args.bin"
},
{
"rfilename": "checkpoint-26/vocab.json"
},
{
"rfilename": "checkpoint-52/README.md"
},
{
"rfilename": "checkpoint-52/adapter_config.json"
},
{
"rfilename": "checkpoint-52/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-52/added_tokens.json"
},
{
"rfilename": "checkpoint-52/merges.txt"
},
{
"rfilename": "checkpoint-52/optimizer.pt"
},
{
"rfilename": "checkpoint-52/rng_state.pth"
},
{
"rfilename": "checkpoint-52/scheduler.pt"
},
{
"rfilename": "checkpoint-52/special_tokens_map.json"
},
{
"rfilename": "checkpoint-52/tokenizer.json"
},
{
"rfilename": "checkpoint-52/tokenizer_config.json"
},
{
"rfilename": "checkpoint-52/trainer_state.json"
},
{
"rfilename": "checkpoint-52/training_args.bin"
},
{
"rfilename": "checkpoint-52/vocab.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null |
Qwen/Qwen1.5-7B
| 6,515 |
text-generation
| 6,515 | 0 | 0 | 0 | 0 | 841,275 |
6683fcf2a02d9851e16590ed
|
bartowski/Phi-3.1-mini-4k-instruct-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-08-03T22:54:40.000Z
| 41 | false |
66a614ace4d069a12c2f6043f4ea92621c898d4a
| null | 5,565 |
[
"gguf",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"region:us"
] |
text-generation
| null |
2024-07-02T13:13:22.000Z
|
bartowski/Phi-3.1-mini-4k-instruct-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ2_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q2_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q2_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_XL.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-f32.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
microsoft/Phi-3-mini-4k-instruct
| 470 |
text-generation
| 470 | 0 | 0 | 0 | 0 | 733,199 |
657778ad4d989b0a6876e6da
|
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ
|
TheBloke
|
False
|
explicit-opt-out
|
2023-12-14T14:30:42.000Z
| 57 | false |
9afb6f0a7d7fe9ecebdda1baa4ff4e13e73e97d7
|
{"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}"}}
| 5,395 |
[
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] |
text-generation
|
transformers
|
2023-12-11T21:01:33.000Z
|
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mixtral
|
mistralai/Mixtral-8x7B-Instruct-v0.1
| 166 |
text-generation
| 166 | 0 | 0 | 0 | 0 | 414,713 |
664198ddb30e3ea1a6984be2
|
NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF
|
NousResearch
|
False
|
library-not-detected
|
2024-05-14T17:31:19.000Z
| 84 | false |
ded6bdf5c60c7211eb8cb5cec6c91fa358b509b7
| null | 5,388 |
[
"gguf",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"merges",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B",
"region:us"
] | null | null |
2024-05-13T04:36:45.000Z
|
NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-F16.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q4_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q5_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q6_K.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null |
NousResearch/Hermes-2-Pro-Llama-3-8B
| 131 |
text-generation
| 131 | 0 | 0 | 0 | 0 | 635,787 |
6628328a73af5913c6b06bd0
|
solidrust/Meta-Llama-3-8B-AWQ
|
solidrust
|
False
|
explicit-opt-out
|
2024-09-03T08:11:02.000Z
| 0 | false |
bac14f82f7ffcd808e3346615cc291e813bfc640
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
| 5,248 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"4-bit",
"AWQ",
"autotrain_compatible",
"endpoints_compatible",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"text-generation-inference",
"awq",
"region:us"
] |
text-generation
|
transformers
|
2024-04-23T22:13:30.000Z
|
solidrust/Meta-Llama-3-8B-AWQ
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 0 | 601,760 |
6632e960060ab1f666c13fce
|
third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit
|
third-intellect
|
False
|
not-popular-enough
|
2024-09-15T16:39:57.000Z
| 0 | false |
b78320a549884e10558866480daf25ed84348774
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] in ['user', 'system']) %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif message['role'] == 'assistant' %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 5,242 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"math",
"conversational",
"en",
"base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"base_model:finetune:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-02T01:16:16.000Z
|
third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
unsloth/Phi-3-mini-4k-instruct-bnb-4bit
| 626 |
text-generation
| 626 | 0 | 0 | 0 | 0 | 615,912 |
6680ed1247f284fc37511f13
|
ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf
|
ruslandev
|
False
|
library-not-detected
|
2024-06-30T06:10:11.000Z
| 15 | false |
c23b81fa47299676f1190d382fe4fc7adbca0a0d
| null | 5,156 |
[
"gguf",
"generated_from_trainer",
"dataset:ruslandev/tagengo-rus-gpt-4o",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | null | null |
2024-06-30T05:28:50.000Z
|
ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "ggml-model-Q2_K.gguf"
},
{
"rfilename": "ggml-model-Q4_K_M.gguf"
},
{
"rfilename": "ggml-model-Q8_0.gguf"
},
{
"rfilename": "ggml-model-f16.gguf"
}
] | null |
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 0 | 0 | 0 | 729,350 |
65e76fcdc7a0617cc7ed56ea
|
neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin
|
neuralmagic
|
False
|
not-popular-enough
|
2024-03-06T01:00:38.000Z
| 1 | false |
29715d88e87589af7a9c6992ad5a94a2b8f1792a
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 5,152 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"nm-vllm",
"marlin",
"int4",
"conversational",
"arxiv:2210.17323",
"base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"base_model:quantized:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] |
text-generation
|
transformers
|
2024-03-05T19:17:33.000Z
|
neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantization/apply_gptq_save_marlin.py"
},
{
"rfilename": "quantization/requirements.txt"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
TinyLlama/TinyLlama-1.1B-Chat-v1.0
| 687 |
text-generation
| 687 | 0 | 0 | 0 | 0 | 521,648 |
65d1e0f29e6805a7b1faa8b6
|
NousResearch/Nous-Hermes-2-Mistral-7B-DPO
|
NousResearch
|
False
|
not-popular-enough
|
2024-04-30T18:10:15.000Z
| 164 | false |
ebec0a691037d38955727d6949798429a63929dd
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
| 5,110 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"Mistral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-02-18T10:50:26.000Z
|
NousResearch/Nous-Hermes-2-Mistral-7B-DPO
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 18 | 10 | 8 | 497,305 |
66de08a7becd5c1c0c11b3d5
|
silent666/01-ai-Yi-1.5-9B-1725827239
|
silent666
|
False
|
pipeline-not-detected
|
2024-09-08T20:28:02.000Z
| 0 | false |
93a6b7b8673c593fe208aae80125f4526f64a26e
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 5,065 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null |
peft
|
2024-09-08T20:27:19.000Z
|
silent666/01-ai-Yi-1.5-9B-1725827239
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 916,165 |
66867c2143738edbd3361dd4
|
bartowski/Phi-3.1-mini-128k-instruct-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-08-03T22:54:37.000Z
| 27 | false |
32f6acf8f29d7293ef5a43718796aff2a719e44e
| null | 5,063 |
[
"gguf",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/Phi-3-mini-128k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-128k-instruct",
"license:mit",
"region:us"
] |
text-generation
| null |
2024-07-04T10:40:33.000Z
|
bartowski/Phi-3.1-mini-128k-instruct-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XXS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q2_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q2_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_XL.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q6_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q8_0.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-f32.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
microsoft/Phi-3-mini-128k-instruct
| 196 |
text-generation
| 196 | 0 | 0 | 0 | 0 | 736,398 |
66a07e74dc2d9be9833d1702
|
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8
|
neuralmagic
|
False
|
not-popular-enough
|
2024-09-27T21:16:49.000Z
| 9 | false |
1343a1c1803acdfc4bcfc96b0bc3bb6f905bd9d4
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}}
| 4,998 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"int8",
"vllm",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2210.17323",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-07-24T04:09:24.000Z
|
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "recipe.yaml"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-3.1-8B-Instruct
| 695 |
text-generation
| 695 | 0 | 0 | 0 | 0 | 773,970 |
66620154cb9f38e8797b88e8
|
legraphista/Qwen2-1.5B-Instruct-IMat-GGUF
|
legraphista
|
False
|
explicit-opt-out
|
2024-06-06T19:06:17.000Z
| 1 | false |
b0566f98939abff7830b1e91117b338c01f60bdc
| null | 4,814 |
[
"gguf",
"chat",
"quantized",
"GGUF",
"imatrix",
"quantization",
"imat",
"static",
"16bit",
"8bit",
"6bit",
"5bit",
"4bit",
"3bit",
"2bit",
"1bit",
"text-generation",
"en",
"base_model:Qwen/Qwen2-1.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-1.5B-Instruct",
"license:apache-2.0",
"region:us"
] |
text-generation
|
gguf
|
2024-06-06T18:35:00.000Z
|
legraphista/Qwen2-1.5B-Instruct-IMat-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-1.5B-Instruct.BF16.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.FP16.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ1_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ1_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_XXS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_XXS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ4_NL.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ4_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q2_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q2_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q4_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q4_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q5_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q5_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q6_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q8_0.gguf"
},
{
"rfilename": "README.md"
},
{
"rfilename": "imatrix.dat"
},
{
"rfilename": "imatrix.dataset"
},
{
"rfilename": "imatrix.log"
}
] | null |
Qwen/Qwen2-1.5B-Instruct
| 495 |
text-generation
| 495 | 0 | 0 | 0 | 0 | 686,334 |
65531bd06a12e47a509523fe
|
Intel/neural-chat-7b-v3-1
|
Intel
|
False
|
not-popular-enough
|
2024-09-09T09:24:54.000Z
| 542 | false |
8189385e1dc17bcff572511123b0ba7dc553f743
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": true, "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if loop.first and system_message is defined %}\n {{- '### System:\n' + system_message + '\n' }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {{- '### User:\n' + message['content'] + ' \n' }}\n {%- elif message['role'] == 'assistant' %}\n {{- '### Assistant:\n' + message['content'] + eos_token + '\n'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}{% if add_generation_prompt %}{{ '### Assistant:\n' }}{% endif %}\n"}}
| 4,644 |
[
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"LLMs",
"Intel",
"conversational",
"en",
"dataset:Open-Orca/SlimOrca",
"arxiv:2306.02707",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2023-11-14T07:03:44.000Z
|
Intel/neural-chat-7b-v3-1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 8 | 3 | 21 | 7 | 378,001 |
667e45e2b3072cb50c85e51f
|
lmstudio-community/gemma-2-9b-it-GGUF
|
lmstudio-community
|
False
|
not-popular-enough
|
2024-07-16T17:36:07.000Z
| 18 | false |
8fcfcc720a20ee55fa1c9ed602dc12aabc04e525
| null | 4,628 |
[
"transformers",
"gguf",
"conversational",
"text-generation",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"license:gemma",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-28T05:10:58.000Z
|
lmstudio-community/gemma-2-9b-it-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "gemma-2-9b-it-IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q4_K_M.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q5_K_M.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q6_K.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q8_0.gguf"
}
] | null |
google/gemma-2-9b-it
| 131 |
text-generation
| 131 | 0 | 0 | 0 | 0 | 726,142 |
66b3eb1723d744f31d8ca777
|
mradermacher/Mistral-Nemo-Instruct-2407-GGUF
|
mradermacher
|
False
|
pipeline-not-detected
|
2024-08-09T12:18:41.000Z
| 9 | false |
19d4046af26dc3f2aa901fecb8f11dcb106efeae
| null | 4,528 |
[
"transformers",
"gguf",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ru",
"zh",
"ja",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null |
transformers
|
2024-08-07T21:45:59.000Z
|
mradermacher/Mistral-Nemo-Instruct-2407-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ4_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q2_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q6_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null |
mistralai/Mistral-Nemo-Instruct-2407
| 101 |
text-generation
| 101 | 0 | 0 | 0 | 0 | 806,266 |
66258007c8920ec3510b4847
|
ruslanmv/Medical-Llama3-8B
|
ruslanmv
|
False
|
not-popular-enough
|
2024-05-15T08:53:33.000Z
| 74 | false |
82cb0a9d1d30a9532253d21c2793f7163535357b
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|reserved_special_token_250|>"}}
| 4,497 |
[
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"ruslanmv",
"trl",
"llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"heathcare",
"medical",
"clinical",
"med",
"lifescience",
"Pharmaceutical",
"Pharma",
"en",
"dataset:ruslanmv/ai-medical-chatbot",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-21T21:07:19.000Z
|
ruslanmv/Medical-Llama3-8B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "future.jpg"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 5 | 2 | 7 | 597,866 |
6661549b4e5af73c6a61831a
|
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
|
Qwen
|
False
|
not-popular-enough
|
2024-08-21T10:34:22.000Z
| 22 | false |
9d6101c42b48464aeb055c343b4f67ab0aa7a2cb
|
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
| 4,494 |
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] |
text-generation
|
transformers
|
2024-06-06T06:18:03.000Z
|
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] |
qwen2
|
Qwen/Qwen2-7B-Instruct
| 161 |
text-generation
| 161 | 0 | 0 | 0 | 0 | 685,152 |
669207a83bd55e86d861533e
|
Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3
|
Magpie-Align
|
False
|
not-popular-enough
|
2024-07-19T20:33:26.000Z
| 2 | false |
d2578eb754d1c20efe604749296580f680950917
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
| 4,472 |
[
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"axolotl",
"generated_from_trainer",
"conversational",
"en",
"zh",
"dataset:Magpie-Align/Magpie-Reasoning-150K",
"dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1",
"dataset:Magpie-Align/Magpie-Qwen2-Pro-200K-Chinese",
"arxiv:2406.08464",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-07-13T04:50:48.000Z
|
Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 1 | 0 | 749,775 |
663473a227f10f5fb3d56c7b
|
FairMind/Llama-3-8B-4bit-UltraChat-Ita
|
FairMind
|
False
|
not-popular-enough
|
2024-05-06T07:26:05.000Z
| 0 | false |
473a2ee0296e1471ae635fd8977dd3b44404f0e8
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
| 4,459 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"it",
"dataset:mii-community/ultrafeedback-translated-ita",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-03T05:18:26.000Z
|
FairMind/Llama-3-8B-4bit-UltraChat-Ita
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
unsloth/llama-3-8b-bnb-4bit
| 3,018 |
text-generation
| 3,018 | 0 | 0 | 0 | 0 | 618,127 |
655614a6bf4c000cc62ad949
|
argilla/notus-7b-v1
|
argilla
|
False
|
not-popular-enough
|
2023-12-05T07:49:46.000Z
| 121 | false |
30172203a2d41cb487bf7e2b92a821080783b2c9
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
| 4,445 |
[
"transformers",
"tensorboard",
"safetensors",
"mistral",
"text-generation",
"dpo",
"rlaif",
"preference",
"ultrafeedback",
"conversational",
"en",
"dataset:argilla/ultrafeedback-binarized-preferences",
"base_model:alignment-handbook/zephyr-7b-sft-full",
"base_model:finetune:alignment-handbook/zephyr-7b-sft-full",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2023-11-16T13:09:58.000Z
|
argilla/notus-7b-v1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "runs/Nov18_19-37-37_130-61-173-32/events.out.tfevents.1700336346.130-61-173-32.83547.0"
},
{
"rfilename": "runs/Nov18_19-43-49_130-61-173-32/events.out.tfevents.1700336689.130-61-173-32.84702.0"
},
{
"rfilename": "runs/Nov19_08-40-16_130-61-173-32/events.out.tfevents.1700383276.130-61-173-32.110246.0"
},
{
"rfilename": "runs/Nov19_08-42-46_130-61-173-32/events.out.tfevents.1700383428.130-61-173-32.110979.0"
},
{
"rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700383712.130-61-173-32.111866.0"
},
{
"rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700427124.130-61-173-32.111866.1"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] |
mistral
|
alignment-handbook/zephyr-7b-sft-full
| 401 |
text-generation
| 401 | 1 | 12 | 2 | 5 | 381,123 |
66d932fde781122aaef3e373
|
Skywork/Skywork-Reward-Llama-3.1-8B
|
Skywork
|
False
|
not-popular-enough
|
2024-09-10T08:42:26.000Z
| 22 | false |
ddea46c1b5888d2c9d0c3c9323d1d1f8b383864c
|
{"architectures": ["LlamaForSequenceClassification"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}}
| 4,417 |
[
"transformers",
"safetensors",
"llama",
"text-classification",
"dataset:Skywork/Skywork-Reward-Preference-80K-v0.1",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-classification
|
transformers
|
2024-09-05T04:26:37.000Z
|
Skywork/Skywork-Reward-Llama-3.1-8B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "misc/Skywork Community License.pdf"
},
{
"rfilename": "misc/Skywork 模型社区许可协议.pdf"
},
{
"rfilename": "misc/fig.jpg"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-3.1-8B-Instruct
| 695 |
text-generation
| 695 | 0 | 0 | 0 | 0 | 900,143 |
66e25f10a352e926aeb6093b
|
jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597
|
jfranklin-foundry
|
False
|
pipeline-not-detected
|
2024-09-12T03:26:26.000Z
| 0 | false |
3768384a81df5041d0d5c567291a0ad933c85cd2
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 4,377 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null |
peft
|
2024-09-12T03:25:04.000Z
|
jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1042/README.md"
},
{
"rfilename": "checkpoint-1042/adapter_config.json"
},
{
"rfilename": "checkpoint-1042/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1042/optimizer.pt"
},
{
"rfilename": "checkpoint-1042/rng_state.pth"
},
{
"rfilename": "checkpoint-1042/scheduler.pt"
},
{
"rfilename": "checkpoint-1042/special_tokens_map.json"
},
{
"rfilename": "checkpoint-1042/tokenizer.json"
},
{
"rfilename": "checkpoint-1042/tokenizer.model"
},
{
"rfilename": "checkpoint-1042/tokenizer_config.json"
},
{
"rfilename": "checkpoint-1042/trainer_state.json"
},
{
"rfilename": "checkpoint-1042/training_args.bin"
},
{
"rfilename": "checkpoint-1080/README.md"
},
{
"rfilename": "checkpoint-1080/adapter_config.json"
},
{
"rfilename": "checkpoint-1080/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1080/optimizer.pt"
},
{
"rfilename": "checkpoint-1080/rng_state.pth"
},
{
"rfilename": "checkpoint-1080/scheduler.pt"
},
{
"rfilename": "checkpoint-1080/special_tokens_map.json"
},
{
"rfilename": "checkpoint-1080/tokenizer.json"
},
{
"rfilename": "checkpoint-1080/tokenizer.model"
},
{
"rfilename": "checkpoint-1080/tokenizer_config.json"
},
{
"rfilename": "checkpoint-1080/trainer_state.json"
},
{
"rfilename": "checkpoint-1080/training_args.bin"
},
{
"rfilename": "checkpoint-109/README.md"
},
{
"rfilename": "checkpoint-109/adapter_config.json"
},
{
"rfilename": "checkpoint-109/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-109/optimizer.pt"
},
{
"rfilename": "checkpoint-109/rng_state.pth"
},
{
"rfilename": "checkpoint-109/scheduler.pt"
},
{
"rfilename": "checkpoint-109/special_tokens_map.json"
},
{
"rfilename": "checkpoint-109/tokenizer.json"
},
{
"rfilename": "checkpoint-109/tokenizer.model"
},
{
"rfilename": "checkpoint-109/tokenizer_config.json"
},
{
"rfilename": "checkpoint-109/trainer_state.json"
},
{
"rfilename": "checkpoint-109/training_args.bin"
},
{
"rfilename": "checkpoint-164/README.md"
},
{
"rfilename": "checkpoint-164/adapter_config.json"
},
{
"rfilename": "checkpoint-164/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-164/optimizer.pt"
},
{
"rfilename": "checkpoint-164/rng_state.pth"
},
{
"rfilename": "checkpoint-164/scheduler.pt"
},
{
"rfilename": "checkpoint-164/special_tokens_map.json"
},
{
"rfilename": "checkpoint-164/tokenizer.json"
},
{
"rfilename": "checkpoint-164/tokenizer.model"
},
{
"rfilename": "checkpoint-164/tokenizer_config.json"
},
{
"rfilename": "checkpoint-164/trainer_state.json"
},
{
"rfilename": "checkpoint-164/training_args.bin"
},
{
"rfilename": "checkpoint-219/README.md"
},
{
"rfilename": "checkpoint-219/adapter_config.json"
},
{
"rfilename": "checkpoint-219/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-219/optimizer.pt"
},
{
"rfilename": "checkpoint-219/rng_state.pth"
},
{
"rfilename": "checkpoint-219/scheduler.pt"
},
{
"rfilename": "checkpoint-219/special_tokens_map.json"
},
{
"rfilename": "checkpoint-219/tokenizer.json"
},
{
"rfilename": "checkpoint-219/tokenizer.model"
},
{
"rfilename": "checkpoint-219/tokenizer_config.json"
},
{
"rfilename": "checkpoint-219/trainer_state.json"
},
{
"rfilename": "checkpoint-219/training_args.bin"
},
{
"rfilename": "checkpoint-270/README.md"
},
{
"rfilename": "checkpoint-270/adapter_config.json"
},
{
"rfilename": "checkpoint-270/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-270/optimizer.pt"
},
{
"rfilename": "checkpoint-270/rng_state.pth"
},
{
"rfilename": "checkpoint-270/scheduler.pt"
},
{
"rfilename": "checkpoint-270/special_tokens_map.json"
},
{
"rfilename": "checkpoint-270/tokenizer.json"
},
{
"rfilename": "checkpoint-270/tokenizer.model"
},
{
"rfilename": "checkpoint-270/tokenizer_config.json"
},
{
"rfilename": "checkpoint-270/trainer_state.json"
},
{
"rfilename": "checkpoint-270/training_args.bin"
},
{
"rfilename": "checkpoint-274/README.md"
},
{
"rfilename": "checkpoint-274/adapter_config.json"
},
{
"rfilename": "checkpoint-274/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-274/optimizer.pt"
},
{
"rfilename": "checkpoint-274/rng_state.pth"
},
{
"rfilename": "checkpoint-274/scheduler.pt"
},
{
"rfilename": "checkpoint-274/special_tokens_map.json"
},
{
"rfilename": "checkpoint-274/tokenizer.json"
},
{
"rfilename": "checkpoint-274/tokenizer.model"
},
{
"rfilename": "checkpoint-274/tokenizer_config.json"
},
{
"rfilename": "checkpoint-274/trainer_state.json"
},
{
"rfilename": "checkpoint-274/training_args.bin"
},
{
"rfilename": "checkpoint-329/README.md"
},
{
"rfilename": "checkpoint-329/adapter_config.json"
},
{
"rfilename": "checkpoint-329/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-329/optimizer.pt"
},
{
"rfilename": "checkpoint-329/rng_state.pth"
},
{
"rfilename": "checkpoint-329/scheduler.pt"
},
{
"rfilename": "checkpoint-329/special_tokens_map.json"
},
{
"rfilename": "checkpoint-329/tokenizer.json"
},
{
"rfilename": "checkpoint-329/tokenizer.model"
},
{
"rfilename": "checkpoint-329/tokenizer_config.json"
},
{
"rfilename": "checkpoint-329/trainer_state.json"
},
{
"rfilename": "checkpoint-329/training_args.bin"
},
{
"rfilename": "checkpoint-384/README.md"
},
{
"rfilename": "checkpoint-384/adapter_config.json"
},
{
"rfilename": "checkpoint-384/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-384/optimizer.pt"
},
{
"rfilename": "checkpoint-384/rng_state.pth"
},
{
"rfilename": "checkpoint-384/scheduler.pt"
},
{
"rfilename": "checkpoint-384/special_tokens_map.json"
},
{
"rfilename": "checkpoint-384/tokenizer.json"
},
{
"rfilename": "checkpoint-384/tokenizer.model"
},
{
"rfilename": "checkpoint-384/tokenizer_config.json"
},
{
"rfilename": "checkpoint-384/trainer_state.json"
},
{
"rfilename": "checkpoint-384/training_args.bin"
},
{
"rfilename": "checkpoint-439/README.md"
},
{
"rfilename": "checkpoint-439/adapter_config.json"
},
{
"rfilename": "checkpoint-439/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-439/optimizer.pt"
},
{
"rfilename": "checkpoint-439/rng_state.pth"
},
{
"rfilename": "checkpoint-439/scheduler.pt"
},
{
"rfilename": "checkpoint-439/special_tokens_map.json"
},
{
"rfilename": "checkpoint-439/tokenizer.json"
},
{
"rfilename": "checkpoint-439/tokenizer.model"
},
{
"rfilename": "checkpoint-439/tokenizer_config.json"
},
{
"rfilename": "checkpoint-439/trainer_state.json"
},
{
"rfilename": "checkpoint-439/training_args.bin"
},
{
"rfilename": "checkpoint-493/README.md"
},
{
"rfilename": "checkpoint-493/adapter_config.json"
},
{
"rfilename": "checkpoint-493/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-493/optimizer.pt"
},
{
"rfilename": "checkpoint-493/rng_state.pth"
},
{
"rfilename": "checkpoint-493/scheduler.pt"
},
{
"rfilename": "checkpoint-493/special_tokens_map.json"
},
{
"rfilename": "checkpoint-493/tokenizer.json"
},
{
"rfilename": "checkpoint-493/tokenizer.model"
},
{
"rfilename": "checkpoint-493/tokenizer_config.json"
},
{
"rfilename": "checkpoint-493/trainer_state.json"
},
{
"rfilename": "checkpoint-493/training_args.bin"
},
{
"rfilename": "checkpoint-54/README.md"
},
{
"rfilename": "checkpoint-54/adapter_config.json"
},
{
"rfilename": "checkpoint-54/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-54/optimizer.pt"
},
{
"rfilename": "checkpoint-54/rng_state.pth"
},
{
"rfilename": "checkpoint-54/scheduler.pt"
},
{
"rfilename": "checkpoint-54/special_tokens_map.json"
},
{
"rfilename": "checkpoint-54/tokenizer.json"
},
{
"rfilename": "checkpoint-54/tokenizer.model"
},
{
"rfilename": "checkpoint-54/tokenizer_config.json"
},
{
"rfilename": "checkpoint-54/trainer_state.json"
},
{
"rfilename": "checkpoint-54/training_args.bin"
},
{
"rfilename": "checkpoint-540/README.md"
},
{
"rfilename": "checkpoint-540/adapter_config.json"
},
{
"rfilename": "checkpoint-540/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-540/optimizer.pt"
},
{
"rfilename": "checkpoint-540/rng_state.pth"
},
{
"rfilename": "checkpoint-540/scheduler.pt"
},
{
"rfilename": "checkpoint-540/special_tokens_map.json"
},
{
"rfilename": "checkpoint-540/tokenizer.json"
},
{
"rfilename": "checkpoint-540/tokenizer.model"
},
{
"rfilename": "checkpoint-540/tokenizer_config.json"
},
{
"rfilename": "checkpoint-540/trainer_state.json"
},
{
"rfilename": "checkpoint-540/training_args.bin"
},
{
"rfilename": "checkpoint-548/README.md"
},
{
"rfilename": "checkpoint-548/adapter_config.json"
},
{
"rfilename": "checkpoint-548/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-548/optimizer.pt"
},
{
"rfilename": "checkpoint-548/rng_state.pth"
},
{
"rfilename": "checkpoint-548/scheduler.pt"
},
{
"rfilename": "checkpoint-548/special_tokens_map.json"
},
{
"rfilename": "checkpoint-548/tokenizer.json"
},
{
"rfilename": "checkpoint-548/tokenizer.model"
},
{
"rfilename": "checkpoint-548/tokenizer_config.json"
},
{
"rfilename": "checkpoint-548/trainer_state.json"
},
{
"rfilename": "checkpoint-548/training_args.bin"
},
{
"rfilename": "checkpoint-603/README.md"
},
{
"rfilename": "checkpoint-603/adapter_config.json"
},
{
"rfilename": "checkpoint-603/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-603/optimizer.pt"
},
{
"rfilename": "checkpoint-603/rng_state.pth"
},
{
"rfilename": "checkpoint-603/scheduler.pt"
},
{
"rfilename": "checkpoint-603/special_tokens_map.json"
},
{
"rfilename": "checkpoint-603/tokenizer.json"
},
{
"rfilename": "checkpoint-603/tokenizer.model"
},
{
"rfilename": "checkpoint-603/tokenizer_config.json"
},
{
"rfilename": "checkpoint-603/trainer_state.json"
},
{
"rfilename": "checkpoint-603/training_args.bin"
},
{
"rfilename": "checkpoint-658/README.md"
},
{
"rfilename": "checkpoint-658/adapter_config.json"
},
{
"rfilename": "checkpoint-658/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-658/optimizer.pt"
},
{
"rfilename": "checkpoint-658/rng_state.pth"
},
{
"rfilename": "checkpoint-658/scheduler.pt"
},
{
"rfilename": "checkpoint-658/special_tokens_map.json"
},
{
"rfilename": "checkpoint-658/tokenizer.json"
},
{
"rfilename": "checkpoint-658/tokenizer.model"
},
{
"rfilename": "checkpoint-658/tokenizer_config.json"
},
{
"rfilename": "checkpoint-658/trainer_state.json"
},
{
"rfilename": "checkpoint-658/training_args.bin"
},
{
"rfilename": "checkpoint-713/README.md"
},
{
"rfilename": "checkpoint-713/adapter_config.json"
},
{
"rfilename": "checkpoint-713/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-713/optimizer.pt"
},
{
"rfilename": "checkpoint-713/rng_state.pth"
},
{
"rfilename": "checkpoint-713/scheduler.pt"
},
{
"rfilename": "checkpoint-713/special_tokens_map.json"
},
{
"rfilename": "checkpoint-713/tokenizer.json"
},
{
"rfilename": "checkpoint-713/tokenizer.model"
},
{
"rfilename": "checkpoint-713/tokenizer_config.json"
},
{
"rfilename": "checkpoint-713/trainer_state.json"
},
{
"rfilename": "checkpoint-713/training_args.bin"
},
{
"rfilename": "checkpoint-768/README.md"
},
{
"rfilename": "checkpoint-768/adapter_config.json"
},
{
"rfilename": "checkpoint-768/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-768/optimizer.pt"
},
{
"rfilename": "checkpoint-768/rng_state.pth"
},
{
"rfilename": "checkpoint-768/scheduler.pt"
},
{
"rfilename": "checkpoint-768/special_tokens_map.json"
},
{
"rfilename": "checkpoint-768/tokenizer.json"
},
{
"rfilename": "checkpoint-768/tokenizer.model"
},
{
"rfilename": "checkpoint-768/tokenizer_config.json"
},
{
"rfilename": "checkpoint-768/trainer_state.json"
},
{
"rfilename": "checkpoint-768/training_args.bin"
},
{
"rfilename": "checkpoint-823/README.md"
},
{
"rfilename": "checkpoint-823/adapter_config.json"
},
{
"rfilename": "checkpoint-823/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-823/optimizer.pt"
},
{
"rfilename": "checkpoint-823/rng_state.pth"
},
{
"rfilename": "checkpoint-823/scheduler.pt"
},
{
"rfilename": "checkpoint-823/special_tokens_map.json"
},
{
"rfilename": "checkpoint-823/tokenizer.json"
},
{
"rfilename": "checkpoint-823/tokenizer.model"
},
{
"rfilename": "checkpoint-823/tokenizer_config.json"
},
{
"rfilename": "checkpoint-823/trainer_state.json"
},
{
"rfilename": "checkpoint-823/training_args.bin"
},
{
"rfilename": "checkpoint-878/README.md"
},
{
"rfilename": "checkpoint-878/adapter_config.json"
},
{
"rfilename": "checkpoint-878/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-878/optimizer.pt"
},
{
"rfilename": "checkpoint-878/rng_state.pth"
},
{
"rfilename": "checkpoint-878/scheduler.pt"
},
{
"rfilename": "checkpoint-878/special_tokens_map.json"
},
{
"rfilename": "checkpoint-878/tokenizer.json"
},
{
"rfilename": "checkpoint-878/tokenizer.model"
},
{
"rfilename": "checkpoint-878/tokenizer_config.json"
},
{
"rfilename": "checkpoint-878/trainer_state.json"
},
{
"rfilename": "checkpoint-878/training_args.bin"
},
{
"rfilename": "checkpoint-918/README.md"
},
{
"rfilename": "checkpoint-918/adapter_config.json"
},
{
"rfilename": "checkpoint-918/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-918/optimizer.pt"
},
{
"rfilename": "checkpoint-918/rng_state.pth"
},
{
"rfilename": "checkpoint-918/scheduler.pt"
},
{
"rfilename": "checkpoint-918/special_tokens_map.json"
},
{
"rfilename": "checkpoint-918/tokenizer.json"
},
{
"rfilename": "checkpoint-918/tokenizer.model"
},
{
"rfilename": "checkpoint-918/tokenizer_config.json"
},
{
"rfilename": "checkpoint-918/trainer_state.json"
},
{
"rfilename": "checkpoint-918/training_args.bin"
},
{
"rfilename": "checkpoint-932/README.md"
},
{
"rfilename": "checkpoint-932/adapter_config.json"
},
{
"rfilename": "checkpoint-932/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-932/optimizer.pt"
},
{
"rfilename": "checkpoint-932/rng_state.pth"
},
{
"rfilename": "checkpoint-932/scheduler.pt"
},
{
"rfilename": "checkpoint-932/special_tokens_map.json"
},
{
"rfilename": "checkpoint-932/tokenizer.json"
},
{
"rfilename": "checkpoint-932/tokenizer.model"
},
{
"rfilename": "checkpoint-932/tokenizer_config.json"
},
{
"rfilename": "checkpoint-932/trainer_state.json"
},
{
"rfilename": "checkpoint-932/training_args.bin"
},
{
"rfilename": "checkpoint-987/README.md"
},
{
"rfilename": "checkpoint-987/adapter_config.json"
},
{
"rfilename": "checkpoint-987/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-987/optimizer.pt"
},
{
"rfilename": "checkpoint-987/rng_state.pth"
},
{
"rfilename": "checkpoint-987/scheduler.pt"
},
{
"rfilename": "checkpoint-987/special_tokens_map.json"
},
{
"rfilename": "checkpoint-987/tokenizer.json"
},
{
"rfilename": "checkpoint-987/tokenizer.model"
},
{
"rfilename": "checkpoint-987/tokenizer_config.json"
},
{
"rfilename": "checkpoint-987/trainer_state.json"
},
{
"rfilename": "checkpoint-987/training_args.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 932,647 |
66353871a4911a4570735dc4
|
umarigan/llama-3.1-openhermes-tr
|
umarigan
|
False
|
not-popular-enough
|
2024-09-12T06:49:56.000Z
| 0 | false |
ce20413536cc7ddacf1248725586ed5e07939545
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|finetune_right_pad_id|>"}}
| 4,297 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"question-answering",
"en",
"tr",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
question-answering
|
transformers
|
2024-05-03T19:18:09.000Z
|
umarigan/llama-3.1-openhermes-tr
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
unsloth/llama-3-8b-bnb-4bit
| 3,018 |
text-generation
| 3,018 | 0 | 0 | 0 | 1 | 619,416 |
667a279ab3a6b45bb28116f0
|
cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
|
cognitivecomputations
|
False
|
not-popular-enough
|
2024-07-06T02:23:30.000Z
| 43 | false |
311c04607bf1434fc8294cd7ff4aa7a4ac968044
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 4,288 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:finetune:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-25T02:12:42.000Z
|
cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.3
| 338 |
text-generation
| 338 | 0 | 0 | 6 | 8 | 719,844 |
66ab3d61fc35e079a96fa448
|
MaziyarPanahi/gemma-2-2b-it-GGUF
|
MaziyarPanahi
|
False
|
explicit-opt-out
|
2024-08-01T08:01:55.000Z
| 4 | false |
bd652eddf75b473fe86639b6b927e06972277d1a
|
{"model_type": "mistral"}
| 4,233 |
[
"transformers",
"gguf",
"mistral",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"GGUF",
"text-generation",
"base_model:google/gemma-2-2b-it",
"base_model:quantized:google/gemma-2-2b-it",
"text-generation-inference",
"region:us"
] |
text-generation
|
transformers
|
2024-08-01T07:46:41.000Z
|
MaziyarPanahi/gemma-2-2b-it-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "gemma-2-2b-it-GGUF_imatrix.dat"
},
{
"rfilename": "gemma-2-2b-it.IQ1_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ1_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ2_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ3_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q2_K.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q4_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q4_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q5_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q5_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q6_K.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q8_0.gguf"
},
{
"rfilename": "gemma-2-2b-it.fp16.gguf"
}
] |
mistral
|
google/gemma-2-2b-it
| 265 |
text-generation
| 265 | 0 | 0 | 0 | 0 | 792,685 |
66a07a525d5f5950b2f28975
|
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16
|
neuralmagic
|
False
|
not-popular-enough
|
2024-09-27T17:20:00.000Z
| 7 | false |
91597e9dae7e037f75d84d1b6be72ab85d054c9f
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}}
| 4,212 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"int8",
"vllm",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2210.17323",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-07-24T03:51:46.000Z
|
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "recipe.yaml"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-3.1-8B-Instruct
| 695 |
text-generation
| 695 | 0 | 0 | 0 | 0 | 773,942 |
66501bac87d9d7946416180d
|
cognitivecomputations/dolphin-2.9.2-qwen2-7b
|
cognitivecomputations
|
False
|
not-popular-enough
|
2024-06-18T06:39:05.000Z
| 60 | false |
c443c4eb5138ed746ac49ed98bf3c183dc5380ac
|
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
| 4,126 |
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-24T04:46:36.000Z
|
cognitivecomputations/dolphin-2.9.2-qwen2-7b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] |
qwen2
|
Qwen/Qwen2-7B
| 266 |
text-generation
| 266 | 0 | 3 | 8 | 8 | 656,059 |
66d53d2c18b92f31af998f31
|
nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843
|
nekokiku
|
False
|
pipeline-not-detected
|
2024-09-02T04:21:26.000Z
| 0 | false |
4332231b6813c074c01b23a576b4393a281d18f0
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
| 4,073 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null |
peft
|
2024-09-02T04:21:00.000Z
|
nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B-Chat
| 134 |
text-generation
| 134 | 0 | 0 | 0 | 0 | 886,934 |
66d32c311ba71ac4c0a48367
|
xueyj/google-gemma-2b-1725115441
|
xueyj
|
False
|
pipeline-not-detected
|
2024-08-31T14:46:57.000Z
| 0 | false |
249a27f4d333bf2f9803420784a6011dcef362db
|
{"tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "google/gemma-7b", "task_type": "CAUSAL_LM"}}
| 4,010 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-7b",
"base_model:adapter:google/gemma-7b",
"region:us"
] | null |
peft
|
2024-08-31T14:44:01.000Z
|
xueyj/google-gemma-2b-1725115441
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
google/gemma-7b
| 9,278 |
text-generation
| 9,278 | 0 | 0 | 0 | 0 | 881,236 |
66df0455fd364d10756135fd
|
silent666/01-ai-Yi-1.5-9B-1725891669
|
silent666
|
False
|
pipeline-not-detected
|
2024-09-09T14:21:28.000Z
| 0 | false |
1a7c37ab49cf39f6fe9348dc4f8b89567bfa9138
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message + '\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '\nAssistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '<|endoftext|>' + '\n' }}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 3,987 |
[
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"license:other",
"region:us"
] | null |
peft
|
2024-09-09T14:21:09.000Z
|
silent666/01-ai-Yi-1.5-9B-1725891669
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 920,419 |
6631f509060ab1f6667ec024
|
rinna/llama-3-youko-8b
|
rinna
|
False
|
explicit-opt-out
|
2024-08-31T15:07:48.000Z
| 55 | false |
c11ca0b15509262cb443045c57f7f75bc7e7d9d9
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
| 3,894 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"ja",
"en",
"dataset:mc4",
"dataset:wikipedia",
"dataset:EleutherAI/pile",
"dataset:oscar-corpus/colossal-oscar-1.0",
"dataset:cc100",
"arxiv:2404.01657",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] |
text-generation
|
transformers
|
2024-05-01T07:53:45.000Z
|
rinna/llama-3-youko-8b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "rinna.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 5 | 1 | 8 | 6 | 614,533 |
6648db50f7e6c2f290292ac2
|
cognitivecomputations/dolphin-2.9.1-yi-1.5-9b
|
cognitivecomputations
|
False
|
not-popular-enough
|
2024-05-20T14:35:39.000Z
| 26 | false |
91f0a521e3e2a0675a3549aa5d3f40717068de94
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,750 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:01-ai/Yi-1.5-9B",
"base_model:finetune:01-ai/Yi-1.5-9B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-18T16:46:08.000Z
|
cognitivecomputations/dolphin-2.9.1-yi-1.5-9b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 2 | 2 | 5 | 9 | 645,594 |
66840a15f9b20b3c916b2584
|
lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF
|
lmstudio-community
|
False
|
library-not-detected
|
2024-08-01T13:09:44.000Z
| 19 | false |
7b8efa410c799e3254c1f1e2d71a49e7a985732e
| null | 3,731 |
[
"gguf",
"nlp",
"code",
"text-generation",
"en",
"arxiv:2404.14219",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"region:us"
] |
text-generation
| null |
2024-07-02T14:09:25.000Z
|
lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null |
microsoft/Phi-3-mini-4k-instruct
| 470 |
text-generation
| 470 | 0 | 0 | 0 | 0 | 733,295 |
6679d08b77c4b63bb8f9db79
|
Weyaxi/Einstein-v7-Qwen2-7B
|
Weyaxi
|
False
|
not-popular-enough
|
2024-07-23T21:10:17.000Z
| 33 | false |
e5046fb8d078d2028b1743b309c734e5d94138c1
|
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>", "unk_token": null}}
| 3,728 |
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"axolotl",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"science",
"physics",
"chemistry",
"biology",
"math",
"qwen",
"conversational",
"en",
"dataset:allenai/ai2_arc",
"dataset:camel-ai/physics",
"dataset:camel-ai/chemistry",
"dataset:camel-ai/biology",
"dataset:camel-ai/math",
"dataset:metaeval/reclor",
"dataset:openbookqa",
"dataset:mandyyyyii/scibench",
"dataset:derek-thomas/ScienceQA",
"dataset:TIGER-Lab/ScienceEval",
"dataset:jondurbin/airoboros-3.2",
"dataset:LDJnr/Capybara",
"dataset:Cot-Alpaca-GPT4-From-OpenHermes-2.5",
"dataset:STEM-AI-mtl/Electrical-engineering",
"dataset:knowrohit07/saraswati-stem",
"dataset:sablo/oasst2_curated",
"dataset:lmsys/lmsys-chat-1m",
"dataset:TIGER-Lab/MathInstruct",
"dataset:bigbio/med_qa",
"dataset:meta-math/MetaMathQA-40K",
"dataset:piqa",
"dataset:scibench",
"dataset:sciq",
"dataset:Open-Orca/SlimOrca",
"dataset:migtissera/Synthia-v1.3",
"dataset:allenai/WildChat",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:openchat/openchat_sharegpt4_dataset",
"dataset:teknium/GPTeacher-General-Instruct",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:totally-not-an-llm/EverythingLM-data-V3",
"dataset:HuggingFaceH4/no_robots",
"dataset:OpenAssistant/oasst_top1_2023-08-25",
"dataset:WizardLM/WizardLM_evol_instruct_70k",
"dataset:abacusai/SystemChat-1.1",
"dataset:H-D-T/Buzz-V1.2",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:other",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-24T20:01:15.000Z
|
Weyaxi/Einstein-v7-Qwen2-7B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "checkpoint-500/added_tokens.json"
},
{
"rfilename": "checkpoint-500/config.json"
},
{
"rfilename": "checkpoint-500/generation_config.json"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_0_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_1_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_2_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_3_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_4_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_5_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_6_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_7_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/latest"
},
{
"rfilename": "checkpoint-500/merges.txt"
},
{
"rfilename": "checkpoint-500/model-00001-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00002-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00003-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00004-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model.safetensors.index.json"
},
{
"rfilename": "checkpoint-500/rng_state_0.pth"
},
{
"rfilename": "checkpoint-500/rng_state_1.pth"
},
{
"rfilename": "checkpoint-500/rng_state_2.pth"
},
{
"rfilename": "checkpoint-500/rng_state_3.pth"
},
{
"rfilename": "checkpoint-500/rng_state_4.pth"
},
{
"rfilename": "checkpoint-500/rng_state_5.pth"
},
{
"rfilename": "checkpoint-500/rng_state_6.pth"
},
{
"rfilename": "checkpoint-500/rng_state_7.pth"
},
{
"rfilename": "checkpoint-500/scheduler.pt"
},
{
"rfilename": "checkpoint-500/special_tokens_map.json"
},
{
"rfilename": "checkpoint-500/tokenizer.json"
},
{
"rfilename": "checkpoint-500/tokenizer_config.json"
},
{
"rfilename": "checkpoint-500/trainer_state.json"
},
{
"rfilename": "checkpoint-500/training_args.bin"
},
{
"rfilename": "checkpoint-500/vocab.json"
},
{
"rfilename": "checkpoint-500/zero_to_fp32.py"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] |
qwen2
|
Qwen/Qwen2-7B
| 266 |
text-generation
| 266 | 0 | 4 | 6 | 7 | 719,380 |
666e27513b570f44d7862076
|
macadeliccc/Samantha-Qwen-2-7B
|
macadeliccc
|
False
|
not-popular-enough
|
2024-06-17T18:24:56.000Z
| 2 | false |
59058972fa9b56d132d04589eb17cbba277c2826
|
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}}
| 3,695 |
[
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"zh",
"dataset:macadeliccc/opus_samantha",
"dataset:HuggingfaceH4/ultrachat_200k",
"dataset:teknium/OpenHermes-2.5",
"dataset:Sao10K/Claude-3-Opus-Instruct-15K",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-15T23:44:17.000Z
|
macadeliccc/Samantha-Qwen-2-7B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] |
qwen2
|
Qwen/Qwen2-7B
| 266 |
text-generation
| 266 | 0 | 2 | 0 | 3 | 702,775 |
6622940af5c285535ccd6658
|
astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit
|
astronomer
|
False
|
explicit-opt-out
|
2024-04-22T01:34:29.000Z
| 24 | false |
82ea696da54c1ef4f56b9fd1dadaa7c1224bfd5b
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
| 3,672 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"facebook",
"meta",
"astronomer",
"gptq",
"pretrained",
"quantized",
"finetuned",
"autotrain_compatible",
"endpoints_compatible",
"conversational",
"dataset:wikitext",
"arxiv:2210.17323",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"text-generation-inference",
"4-bit",
"region:us"
] |
text-generation
|
transformers
|
2024-04-19T15:55:54.000Z
|
astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 2 | 0 | 0 | 0 | 594,138 |
668119f87ad6fe37b1350d3a
|
shenzhi-wang/Gemma-2-9B-Chinese-Chat
|
shenzhi-wang
|
False
|
not-popular-enough
|
2024-07-04T10:00:18.000Z
| 61 | false |
6265d59abe241f65b593896d217743d9d8f63184
|
{"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{{ '<bos>' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<start_of_turn>user\\n' + content + '<end_of_turn>\\n<start_of_turn>model\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<end_of_turn>\\n' }}{% endif %}{% endfor %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,668 |
[
"transformers",
"safetensors",
"gguf",
"gemma2",
"text-generation",
"llama-factory",
"orpo",
"conversational",
"en",
"zh",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"doi:10.57967/hf/2667",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-30T08:40:24.000Z
|
shenzhi-wang/Gemma-2-9B-Chinese-Chat
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_f16.gguf"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_q4_k_m.gguf"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_q8_0.gguf"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
gemma2
|
google/gemma-2-9b-it
| 131 |
text-generation
| 131 | 0 | 0 | 0 | 8 | 729,536 |
655ba1a6419819fec2aab890
|
typeof/zephyr-7b-beta-lora
|
typeof
|
False
|
not-popular-enough
|
2024-05-25T14:16:49.000Z
| 5 | false |
0af61c575d6f3fd34cfe2e269becfd232312a0b5
|
{"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}}
| 3,641 |
[
"peft",
"safetensors",
"Δ",
"LoRA",
"text-generation",
"en",
"arxiv:2305.18290",
"arxiv:2310.16944",
"arxiv:2305.14314",
"arxiv:2106.09685",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:adapter:mistralai/Mistral-7B-v0.1",
"region:us"
] |
text-generation
|
peft
|
2023-11-20T18:12:54.000Z
|
typeof/zephyr-7b-beta-lora
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null |
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 0 | 0 | 0 | 386,832 |
6621de3a2e2575aa8c884f3c
|
instructlab/merlinite-7b-lab
|
instructlab
|
False
|
not-popular-enough
|
2024-04-19T15:35:58.000Z
| 19 | false |
fe82f56b265c674e3cedb4000529c5bad138575a
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>'+ '\n' + message['content'] + '\n'}}{% elif message['role'] == 'user' %}{{'<|user|>' + '\n' + message['content'] + '\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>' + '\n' + message['content'] + '<|endoftext|>' + ('' if loop.last else '\n')}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|pad|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,615 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"merlinite",
"ibm",
"lab",
"labrador",
"labradorite",
"conversational",
"en",
"arxiv:2403.01081",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-19T03:00:10.000Z
|
instructlab/merlinite-7b-lab
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Screenshot_2024-02-22_at_11.26.13_AM.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 1.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 2.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_intuition.png"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "paper.pdf"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 2 | 2 | 6 | 4 | 593,150 |
66cf0d303ca43793249a1ba7
|
silent666/01-ai-Yi-1.5-9B-Chat-1724845360
|
silent666
|
False
|
pipeline-not-detected
|
2024-08-28T11:42:59.000Z
| 0 | false |
59cba4e6b4520453afe3396ca31224f4593fa225
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
| 3,564 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null |
peft
|
2024-08-28T11:42:40.000Z
|
silent666/01-ai-Yi-1.5-9B-Chat-1724845360
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B-Chat
| 134 |
text-generation
| 134 | 0 | 0 | 0 | 0 | 869,386 |
66d9209310d9daffc37cc28e
|
bill0919g/Qwen-Qwen1.5-1.8B-1725505685
|
bill0919g
|
False
|
pipeline-not-detected
|
2024-09-05T03:08:07.000Z
| 0 | false |
c27fa49773fd15be1f639649b2c32d5f2a58ff30
|
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-1.8B", "task_type": "CAUSAL_LM"}}
| 3,560 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"base_model:adapter:Qwen/Qwen1.5-1.8B",
"region:us"
] | null |
peft
|
2024-09-05T03:08:03.000Z
|
bill0919g/Qwen-Qwen1.5-1.8B-1725505685
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null |
Qwen/Qwen1.5-1.8B
| 26,473 |
text-generation
| 26,473 | 0 | 0 | 0 | 0 | 899,882 |
66a0164c2e7ac0dc7be40925
|
mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC
|
mlc-ai
|
False
|
pipeline-not-detected
|
2024-09-19T04:06:13.000Z
| 2 | false |
f7f71c154986e38428bb2696061b913f2b6f9c28
|
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}", "eos_token": "<|eot_id|>"}}
| 3,557 |
[
"mlc-llm",
"web-llm",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"region:us"
] | null |
mlc-llm
|
2024-07-23T20:45:00.000Z
|
mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache-b16.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_100.bin"
},
{
"rfilename": "params_shard_101.bin"
},
{
"rfilename": "params_shard_102.bin"
},
{
"rfilename": "params_shard_103.bin"
},
{
"rfilename": "params_shard_104.bin"
},
{
"rfilename": "params_shard_105.bin"
},
{
"rfilename": "params_shard_106.bin"
},
{
"rfilename": "params_shard_107.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_38.bin"
},
{
"rfilename": "params_shard_39.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_40.bin"
},
{
"rfilename": "params_shard_41.bin"
},
{
"rfilename": "params_shard_42.bin"
},
{
"rfilename": "params_shard_43.bin"
},
{
"rfilename": "params_shard_44.bin"
},
{
"rfilename": "params_shard_45.bin"
},
{
"rfilename": "params_shard_46.bin"
},
{
"rfilename": "params_shard_47.bin"
},
{
"rfilename": "params_shard_48.bin"
},
{
"rfilename": "params_shard_49.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_50.bin"
},
{
"rfilename": "params_shard_51.bin"
},
{
"rfilename": "params_shard_52.bin"
},
{
"rfilename": "params_shard_53.bin"
},
{
"rfilename": "params_shard_54.bin"
},
{
"rfilename": "params_shard_55.bin"
},
{
"rfilename": "params_shard_56.bin"
},
{
"rfilename": "params_shard_57.bin"
},
{
"rfilename": "params_shard_58.bin"
},
{
"rfilename": "params_shard_59.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_60.bin"
},
{
"rfilename": "params_shard_61.bin"
},
{
"rfilename": "params_shard_62.bin"
},
{
"rfilename": "params_shard_63.bin"
},
{
"rfilename": "params_shard_64.bin"
},
{
"rfilename": "params_shard_65.bin"
},
{
"rfilename": "params_shard_66.bin"
},
{
"rfilename": "params_shard_67.bin"
},
{
"rfilename": "params_shard_68.bin"
},
{
"rfilename": "params_shard_69.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_70.bin"
},
{
"rfilename": "params_shard_71.bin"
},
{
"rfilename": "params_shard_72.bin"
},
{
"rfilename": "params_shard_73.bin"
},
{
"rfilename": "params_shard_74.bin"
},
{
"rfilename": "params_shard_75.bin"
},
{
"rfilename": "params_shard_76.bin"
},
{
"rfilename": "params_shard_77.bin"
},
{
"rfilename": "params_shard_78.bin"
},
{
"rfilename": "params_shard_79.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_80.bin"
},
{
"rfilename": "params_shard_81.bin"
},
{
"rfilename": "params_shard_82.bin"
},
{
"rfilename": "params_shard_83.bin"
},
{
"rfilename": "params_shard_84.bin"
},
{
"rfilename": "params_shard_85.bin"
},
{
"rfilename": "params_shard_86.bin"
},
{
"rfilename": "params_shard_87.bin"
},
{
"rfilename": "params_shard_88.bin"
},
{
"rfilename": "params_shard_89.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "params_shard_90.bin"
},
{
"rfilename": "params_shard_91.bin"
},
{
"rfilename": "params_shard_92.bin"
},
{
"rfilename": "params_shard_93.bin"
},
{
"rfilename": "params_shard_94.bin"
},
{
"rfilename": "params_shard_95.bin"
},
{
"rfilename": "params_shard_96.bin"
},
{
"rfilename": "params_shard_97.bin"
},
{
"rfilename": "params_shard_98.bin"
},
{
"rfilename": "params_shard_99.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | null |
meta-llama/Llama-3.1-8B-Instruct
| 695 |
text-generation
| 695 | 0 | 0 | 0 | 0 | 773,301 |
64e78d117acd8971f2c6bbe6
|
TheBloke/CodeLlama-7B-GGUF
|
TheBloke
|
False
|
explicit-opt-out
|
2023-09-27T12:46:03.000Z
| 105 | false |
98596f7f6c318118824bcbee4b0e20010ec510ec
|
{"model_type": "llama"}
| 3,511 |
[
"transformers",
"gguf",
"llama",
"llama-2",
"text-generation",
"code",
"arxiv:2308.12950",
"base_model:codellama/CodeLlama-7b-hf",
"base_model:quantized:codellama/CodeLlama-7b-hf",
"license:llama2",
"text-generation-inference",
"region:us"
] |
text-generation
|
transformers
|
2023-08-24T17:02:09.000Z
|
TheBloke/CodeLlama-7B-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "codellama-7b.Q2_K.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_L.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q4_0.gguf"
},
{
"rfilename": "codellama-7b.Q4_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q4_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q5_0.gguf"
},
{
"rfilename": "codellama-7b.Q5_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q5_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q6_K.gguf"
},
{
"rfilename": "codellama-7b.Q8_0.gguf"
},
{
"rfilename": "config.json"
}
] |
llama
|
codellama/CodeLlama-7b-hf
| 280 |
text-generation
| 280 | 0 | 0 | 0 | 0 | 290,016 |
65f47f7b65cb2414e060f5ba
|
Commencis/Commencis-LLM
|
Commencis
|
False
|
not-popular-enough
|
2024-03-19T14:12:59.000Z
| 12 | false |
7d7f5a8d47767e71af4c1867fa612c468b59417f
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,511 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"tr",
"en",
"dataset:uonlp/CulturaX",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"base_model:finetune:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-15T17:03:55.000Z
|
Commencis/Commencis-LLM
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-Instruct-v0.1
| 500 |
text-generation
| 500 | 0 | 0 | 0 | 1 | 537,537 |
658dbb208965a503497f9cb2
|
Mozilla/Mistral-7B-Instruct-v0.2-llamafile
|
Mozilla
|
False
|
explicit-opt-out
|
2024-05-25T10:47:13.000Z
| 25 | false |
772507f9981f04bca0b745027b9a9cf585cb490c
|
{"model_type": "mistral"}
| 3,496 |
[
"transformers",
"gguf",
"llamafile",
"mistral",
"finetuned",
"text-generation",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] |
text-generation
|
transformers
|
2023-12-28T18:14:56.000Z
|
Mozilla/Mistral-7B-Instruct-v0.2-llamafile
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mistral-7b-instruct-v0.2.BF16.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.2.BF16.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.F16.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q2_K.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_L.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_0.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_1.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_0.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_1.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q6_K.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q8_0.llamafile"
}
] |
mistral
|
mistralai/Mistral-7B-Instruct-v0.2
| 1,321 |
text-generation
| 1,321 | 0 | 0 | 0 | 0 | 434,304 |
6645102c8b689e8ac99b5187
|
bartowski/Hermes-2-Theta-Llama-3-8B-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-05-18T16:40:16.000Z
| 14 | false |
7c3f5a92241cdbf4d36d88c68ca70f4a1310ddd9
| null | 3,466 |
[
"gguf",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"merges",
"text-generation",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B",
"region:us"
] |
text-generation
| null |
2024-05-15T19:42:36.000Z
|
bartowski/Hermes-2-Theta-Llama-3-8B-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XXS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XXS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ4_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q2_K.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_L.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q6_K.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q8_0.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-f32.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
NousResearch/Hermes-2-Pro-Llama-3-8B
| 131 |
text-generation
| 131 | 0 | 0 | 0 | 0 | 640,694 |
65f45acc243549de57ec81ff
|
NovusResearch/Thestral-0.1-tr-chat-7B
|
NovusResearch
|
False
|
not-popular-enough
|
2024-03-21T08:29:39.000Z
| 3 | false |
9f23ce6c9b8c651fac66ce4ee7cbfe8db250df50
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,463 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-15T14:27:24.000Z
|
NovusResearch/Thestral-0.1-tr-chat-7B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00006.safetensors"
},
{
"rfilename": "model-00002-of-00006.safetensors"
},
{
"rfilename": "model-00003-of-00006.safetensors"
},
{
"rfilename": "model-00004-of-00006.safetensors"
},
{
"rfilename": "model-00005-of-00006.safetensors"
},
{
"rfilename": "model-00006-of-00006.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 0 | 0 | 0 | 537,325 |
6634f79059085933a424818c
|
e-palmisano/Phi3-ITA-mini-4K-instruct
|
e-palmisano
|
False
|
not-popular-enough
|
2024-05-09T07:28:07.000Z
| 1 | false |
26ed09e1cec43e20f73eb563ddbd5af25a266f7e
|
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,428 |
[
"transformers",
"safetensors",
"phi3",
"text-generation",
"text-generation-inference",
"trl",
"sft",
"phi-3",
"phi-3-mini",
"italian",
"conversational",
"custom_code",
"it",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-03T14:41:20.000Z
|
e-palmisano/Phi3-ITA-mini-4K-instruct
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
phi3
|
microsoft/Phi-3-mini-4k-instruct
| 470 |
text-generation
| 470 | 0 | 0 | 0 | 0 | 619,012 |
6640e19023dcace65d05f3e2
|
manupande21/GPT2_PMC
|
manupande21
|
False
|
not-popular-enough
|
2024-05-13T05:43:05.000Z
| 0 | false |
ffee30e670e07dd0f20cfbce80576e0b3be9db5d
|
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": null, "unk_token": "<|endoftext|>"}}
| 3,420 |
[
"transformers",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-12T15:34:40.000Z
|
manupande21/GPT2_PMC
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] |
gpt2
|
openai-community/gpt2
| 2,686 |
text-generation
| 2,686 | 0 | 0 | 0 | 0 | 634,800 |
653fc7d168d8f1436b74b332
|
rinna/youri-7b
|
rinna
|
False
|
explicit-opt-out
|
2024-08-31T14:37:37.000Z
| 22 | false |
5931fb8abebd035d66601d965bd34b7614c816a5
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
| 3,413 |
[
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"ja",
"en",
"dataset:mc4",
"dataset:wikipedia",
"dataset:EleutherAI/pile",
"dataset:oscar-corpus/colossal-oscar-1.0",
"dataset:cc100",
"arxiv:2307.09288",
"arxiv:2404.01657",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"license:llama2",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] |
text-generation
|
transformers
|
2023-10-30T15:12:17.000Z
|
rinna/youri-7b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "rinna.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_checklist.chk"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-2-7b-hf
| 1,595 |
text-generation
| 1,595 | 0 | 0 | 6 | 4 | 361,563 |
66267ef88aa87635c22ce4ec
|
johnsnowlabs/JSL-Med-Sft-Llama-3-8B
|
johnsnowlabs
|
False
|
not-popular-enough
|
2024-04-22T15:31:48.000Z
| 4 | false |
0ea5b1d1f30f962c45ee133c4474ffb184e91c4e
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
| 3,401 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3-8b",
"sft",
"medical",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:cc-by-nc-nd-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-22T15:15:04.000Z
|
johnsnowlabs/JSL-Med-Sft-Llama-3-8B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 0 | 599,138 |
65c7328dc718b4df6907b83c
|
Locutusque/Hercules-2.5-Mistral-7B
|
Locutusque
|
False
|
not-popular-enough
|
2024-02-12T16:59:28.000Z
| 6 | false |
2fb135d4d47e97f7674e15b3776a2dc9371b5858
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,364 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"not-for-all-audiences",
"chemistry",
"math",
"code",
"physics",
"dataset:Locutusque/hercules-v2.0",
"dataset:Locutusque/hercules-v2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-02-10T08:23:41.000Z
|
Locutusque/Hercules-2.5-Mistral-7B
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 10 | 1 | 1 | 487,266 |
662522f22027f3a169d34a5e
|
skfrost19/BioMistralMerged
|
skfrost19
|
False
|
not-popular-enough
|
2024-05-10T03:56:42.000Z
| 0 | false |
43d05fe8c3bb6633b3790773c96ba7572188539d
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,359 |
[
"transformers",
"safetensors",
"gguf",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:quantized:BioMistral/BioMistral-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-21T14:30:10.000Z
|
skfrost19/BioMistralMerged
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "biomistral-merged-v0.1.gguf"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
BioMistral/BioMistral-7B
| 124 |
text-generation
| 124 | 0 | 0 | 0 | 0 | 597,255 |
664f4a93b96b937de17914e4
|
ytu-ce-cosmos/Turkish-Llama-8b-v0.1
|
ytu-ce-cosmos
|
False
|
not-popular-enough
|
2024-05-23T18:30:58.000Z
| 48 | false |
a6fc8bdfa1e4a00ac5b5495131e0204009703ad4
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
| 3,326 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"Turkish",
"turkish",
"Llama",
"Llama3",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-23T13:54:27.000Z
|
ytu-ce-cosmos/Turkish-Llama-8b-v0.1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "CosmosLlaMa.png"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 1 | 0 | 0 | 1 | 655,023 |
65770d54b2b935d83f6ea4d1
|
TheBloke/Mixtral-8x7B-v0.1-GGUF
|
TheBloke
|
False
|
explicit-opt-out
|
2023-12-14T14:30:53.000Z
| 421 | false |
38762deaceb8f976428ab44e4c3ccf927a91132a
|
{"model_type": "mixtral"}
| 3,311 |
[
"transformers",
"gguf",
"mixtral",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | null |
transformers
|
2023-12-11T13:23:32.000Z
|
TheBloke/Mixtral-8x7B-v0.1-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q2_K.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q3_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q4_0.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q4_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q5_0.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q5_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q6_K.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q8_0.gguf"
}
] |
mixtral
|
mistralai/Mixtral-8x7B-v0.1
| 182 |
text-generation
| 182 | 0 | 0 | 0 | 0 | 414,247 |
667aecf5d361e5f642649ee4
|
bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-06-25T16:34:45.000Z
| 7 | false |
740ce4567b3392bd065637d2ac29127ca417cc45
| null | 3,286 |
[
"gguf",
"generated_from_trainer",
"axolotl",
"text-generation",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:quantized:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"region:us"
] |
text-generation
| null |
2024-06-25T16:14:45.000Z
|
bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XXS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ4_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q2_K.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-f32.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k.imatrix"
}
] | null |
mistralai/Mistral-7B-v0.3
| 338 |
text-generation
| 338 | 0 | 0 | 0 | 0 | 721,246 |
66599e09e71d3742325471bf
|
numind/NuExtract
|
numind
|
False
|
not-popular-enough
|
2024-08-22T16:07:13.000Z
| 181 | false |
b27076155936986acf17ef509fae526d262d8bdb
|
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|end-output|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,256 |
[
"transformers",
"safetensors",
"phi3",
"text-generation",
"conversational",
"custom_code",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-31T09:53:13.000Z
|
numind/NuExtract
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "result.png"
},
{
"rfilename": "result_ft.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
phi3
|
microsoft/Phi-3-mini-4k-instruct
| 470 |
text-generation
| 470 | 0 | 0 | 7 | 3 | 672,123 |
66f36be6af8609e9ef8198e5
|
allenai/Molmo-7B-D-0924
|
allenai
|
False
|
custom-code
|
2024-09-27T15:19:57.000Z
| 162 | false |
90426556d5eb7c123eb4368dd1768e8e77f624af
|
{"architectures": ["MolmoForCausalLM"], "auto_map": {"AutoConfig": "config_molmo.MolmoConfig", "AutoModelForCausalLM": "modeling_molmo.MolmoForCausalLM"}, "model_type": "molmo", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}}
| 3,207 |
[
"transformers",
"safetensors",
"molmo",
"text-generation",
"multimodal",
"olmo",
"pixmo",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] |
image-text-to-text
|
transformers
|
2024-09-25T01:48:22.000Z
|
allenai/Molmo-7B-D-0924
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "config_molmo.py"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "image_preprocessing_molmo.py"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00007.safetensors"
},
{
"rfilename": "model-00002-of-00007.safetensors"
},
{
"rfilename": "model-00003-of-00007.safetensors"
},
{
"rfilename": "model-00004-of-00007.safetensors"
},
{
"rfilename": "model-00005-of-00007.safetensors"
},
{
"rfilename": "model-00006-of-00007.safetensors"
},
{
"rfilename": "model-00007-of-00007.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "modeling_molmo.py"
},
{
"rfilename": "molmo_logo.png"
},
{
"rfilename": "preprocessing_molmo.py"
},
{
"rfilename": "preprocessor_config.json"
},
{
"rfilename": "processor_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] |
molmo
|
Qwen/Qwen2-7B
| 266 |
text-generation
| 266 | 0 | 0 | 0 | 1 | 991,876 |
667403ab0c3379a2a436805c
|
ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1
|
ytu-ce-cosmos
|
False
|
not-popular-enough
|
2024-07-02T15:46:04.000Z
| 20 | false |
e4838d5188d00c3e72f9a8fa2cd85cee35f61c80
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
| 3,204 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"Turkish",
"turkish",
"Llama",
"Llama3",
"conversational",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-06-20T10:25:47.000Z
|
ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "cosmosLLaMa2_r2.png"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 6 | 710,949 |
66e51df977230586321100bc
|
beiyemu/01-ai-Yi-1.5-9B-1726291449
|
beiyemu
|
False
|
pipeline-not-detected
|
2024-09-14T05:24:16.000Z
| 0 | false |
fbc3200a2f85db71366ed6350e64a81fe4c531b4
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 3,175 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null |
peft
|
2024-09-14T05:24:09.000Z
|
beiyemu/01-ai-Yi-1.5-9B-1726291449
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 943,178 |
66244c6c251ee52d52247964
|
cognitivecomputations/dolphin-2.9-llama3-8b
|
cognitivecomputations
|
False
|
not-popular-enough
|
2024-05-20T14:42:32.000Z
| 407 | false |
5aeb036f9215c558b483a654a8c6e1cc22e841bf
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}}
| 3,149 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:abacusai/SystemChat-1.1",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-20T23:14:52.000Z
|
cognitivecomputations/dolphin-2.9-llama3-8b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "latest"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "scheduler.pt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 4 | 42 | 23 | 13 | 596,233 |
66d8abc0a0a8dc8c0cdf40f2
|
modularai/llama-3
|
modularai
|
False
|
explicit-opt-out
|
2024-09-04T22:46:52.000Z
| 0 | false |
04d5a76d896e71429f2d042bdb2e5956c649cb41
| null | 3,146 |
[
"gguf",
"facebook",
"meta",
"pytorch",
"llama",
"llama-3",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] |
text-generation
| null |
2024-09-04T18:49:36.000Z
|
modularai/llama-3
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "llama-3-8b-f32.gguf"
},
{
"rfilename": "llama-3-8b-instruct-bf16.gguf"
},
{
"rfilename": "llama-3-8b-instruct-q4_k_m.gguf"
},
{
"rfilename": "llama-3-8b-instruct-q6_k.gguf"
},
{
"rfilename": "llama-3-8b-instruct.q4_0.gguf"
}
] | null |
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 0 | 0 | 0 | 898,574 |
65e7d6dfc7a0617cc712c5a5
|
abacusai/bigstral-12b-32k
|
abacusai
|
False
|
not-popular-enough
|
2024-09-09T14:56:57.000Z
| 43 | false |
cf85301bc4deed33852c292cc6d6ae8ff946c129
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 3,145 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:finetune:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-06T02:37:19.000Z
|
abacusai/bigstral-12b-32k
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-Instruct-v0.2
| 1,321 |
text-generation
| 1,321 | 0 | 0 | 0 | 2 | 522,188 |
66454038994cdc7d34c90922
|
elinas/Llama-3-15B-Instruct-zeroed
|
elinas
|
False
|
not-popular-enough
|
2024-05-15T23:28:20.000Z
| 2 | false |
da513b18459f6adc8d5f583a15161c6a273243b0
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
| 3,141 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-05-15T23:07:36.000Z
|
elinas/Llama-3-15B-Instruct-zeroed
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 3 | 3 | 6 | 640,973 |
66744a6fab975c85911e9818
|
swapnilbp/llama_tweet_ptune
|
swapnilbp
|
False
|
pipeline-not-detected
|
2024-06-20T15:28:25.000Z
| 0 | false |
ad7faa5e642dd3f4311c927b0c423402444cd7a8
|
{"peft": {"base_model_name_or_path": "meta-llama/Llama-2-7b-hf", "task_type": "CAUSAL_LM"}}
| 3,109 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:adapter:meta-llama/Llama-2-7b-hf",
"region:us"
] | null |
peft
|
2024-06-20T15:27:43.000Z
|
swapnilbp/llama_tweet_ptune
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null |
meta-llama/Llama-2-7b-hf
| 1,595 |
text-generation
| 1,595 | 0 | 0 | 0 | 0 | 711,457 |
66dbdb05becd5c1c0c4bde46
|
DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485
|
DreamGallery
|
False
|
pipeline-not-detected
|
2024-09-07T04:48:23.000Z
| 0 | false |
f4868625dbdb11db4df32c7c46179df3a73bbc03
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
| 3,096 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null |
peft
|
2024-09-07T04:48:05.000Z
|
DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B-Chat
| 134 |
text-generation
| 134 | 0 | 0 | 0 | 0 | 908,877 |
659cd276c80023a02e291c82
|
billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF
|
billborkowski
|
False
|
not-popular-enough
|
2024-01-09T07:49:49.000Z
| 20 | false |
fdebcd5b42cff2a44a54174f3cd05b0eae909e8c
|
{"architectures": ["LlavaMistralForCausalLM"], "model_type": "llava_mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": true}}
| 3,023 |
[
"transformers",
"pytorch",
"gguf",
"llava_mistral",
"text-generation",
"mistral",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"multimodal",
"llava",
"conversational",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:quantized:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-01-09T04:58:30.000Z
|
billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "NousResearch_Nous-Hermes-2-Vision-GGUF_Q4_0.gguf"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "mm_projector.bin"
},
{
"rfilename": "mmproj-model-f16.gguf"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] |
llava_mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 0 | 0 | 0 | 445,992 |
6551a45d1e13b38aece0c99d
|
allenai/tulu-v1-llama2-7b
|
allenai
|
False
|
not-popular-enough
|
2023-11-20T02:35:43.000Z
| 0 | false |
4149a2925ddfaef659573740791d5524920f40d4
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
| 2,992 |
[
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"dataset:allenai/tulu-v1-sft-mixture",
"arxiv:2311.10702",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2023-11-13T04:21:49.000Z
|
allenai/tulu-v1-llama2-7b
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.md"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "llama2_license.txt"
},
{
"rfilename": "notice.txt"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Llama-2-7b-hf
| 1,595 |
text-generation
| 1,595 | 0 | 0 | 0 | 0 | 376,644 |
662bc15de85a9207066742ba
|
bartowski/OpenBioLLM-Llama3-8B-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-04-26T15:21:01.000Z
| 4 | false |
0897aa2f9f07b5d390df9e9544934c006e11d4f9
| null | 2,974 |
[
"gguf",
"llama-3",
"llama",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"region:us"
] |
text-generation
| null |
2024-04-26T14:59:41.000Z
|
bartowski/OpenBioLLM-Llama3-8B-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ1_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ1_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_XXS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_XXS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ4_NL.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ4_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q2_K.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_L.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q4_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q4_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q5_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q5_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q6_K.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q8_0.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 0 | 606,963 |
666272569ced3e1387863564
|
mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC
|
mlc-ai
|
False
|
pipeline-not-detected
|
2024-08-08T21:23:32.000Z
| 0 | false |
b659e1b7b27d761da798392c6a7b63862660982f
|
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
| 2,954 |
[
"mlc-llm",
"web-llm",
"base_model:Qwen/Qwen2-1.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-1.5B-Instruct",
"region:us"
] | null |
mlc-llm
|
2024-06-07T02:37:10.000Z
|
mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "logs.txt"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | null |
Qwen/Qwen2-1.5B-Instruct
| 495 |
text-generation
| 495 | 0 | 0 | 0 | 0 | 686,875 |
66dfdeddb50c9046562600cd
|
yizhujiao/llama3-8b-sft-medmcqa
|
yizhujiao
|
False
|
pipeline-not-detected
|
2024-09-27T22:17:28.000Z
| 0 | false |
d71d5e7a5bbf2f7e42430416a467b45ba2e8091a
|
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}}
| 2,867 |
[
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | null |
peft
|
2024-09-10T05:53:33.000Z
|
yizhujiao/llama3-8b-sft-medmcqa
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 0 | 0 | 0 | 923,742 |
662607cad7b6076f66e28caa
|
lightblue/suzume-llama-3-8B-japanese
|
lightblue
|
False
|
not-popular-enough
|
2024-06-02T02:14:36.000Z
| 22 | false |
3306e6bac49e1bd80bfe54001994a828ceeb67fa
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
| 2,865 |
[
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"arxiv:2405.12612",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-22T06:46:34.000Z
|
lightblue/suzume-llama-3-8B-japanese
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 1 | 0 | 5 | 1 | 598,612 |
6676e08c9f2810b0090763a6
|
chenjoya/videollm-online-8b-v1plus
|
chenjoya
|
auto
|
pipeline-library-pair-not-supported
|
2024-07-13T16:24:08.000Z
| 9 | false |
b6541f5208f887690856ebb705e2c6c3b71d0095
|
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ bos_token + messages[0]['content'] + '\n' }}{% set messages = messages[1:] %}{% endif %}{% for message in messages %}{% if message['role'] == 'user' %}{% if add_stream_query_prompt %}{{ ']\nUser: ' + message['content'] }}{% else %}{{ '\nUser: ' + message['content'] }}{% endif %}{% elif message['role'] == 'assistant' %}{{ '\nAssistant: ' + message['content'] + eos_token }}{% elif message['role'] == 'stream' and message['num_frames'] > 0: %}{{ '\n[' + ','.join([10 * '<v>'] * message['num_frames']) + ']' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '\nAssistant:' }}{% elif add_stream_prompt %}{{ '\n[' }}{% elif add_stream_generation_prompt %}{{ ']\nAssistant:' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}}
| 2,859 |
[
"peft",
"safetensors",
"llama",
"llama-3",
"multimodal",
"llm",
"video stream",
"online video understanding",
"video understanding",
"video-text-to-text",
"en",
"dataset:chenjoya/videollm-online-chat-ego4d-134k",
"arxiv:2406.11816",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct",
"license:mit",
"region:us"
] |
video-text-to-text
|
peft
|
2024-06-22T14:32:44.000Z
|
chenjoya/videollm-online-8b-v1plus
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
meta-llama/Meta-Llama-3-8B-Instruct
| 1,076 |
text-generation
| 1,076 | 0 | 0 | 0 | 0 | 715,206 |
6625c8bc11e3eb62194e0d63
|
12thD/ko-Llama-3-8B-sft-v0.1
|
12thD
|
False
|
not-popular-enough
|
2024-04-22T02:50:36.000Z
| 0 | false |
5aca578ff0479831b5417ce031693c1f97899620
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
| 2,791 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"korean",
"gemma",
"conversational",
"ko",
"en",
"arxiv:1910.09700",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-22T02:17:32.000Z
|
12thD/ko-Llama-3-8B-sft-v0.1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 0 | 598,276 |
66df283f731c786526e37ad5
|
silent666/01-ai-Yi-1.5-9B-1725900863
|
silent666
|
False
|
pipeline-not-detected
|
2024-09-09T16:55:06.000Z
| 0 | false |
b0153c783bd1532ea8d7028f657a6a796d2c054f
|
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
| 2,681 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null |
peft
|
2024-09-09T16:54:23.000Z
|
silent666/01-ai-Yi-1.5-9B-1725900863
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null |
01-ai/Yi-1.5-9B
| 225 |
text-generation
| 225 | 0 | 0 | 0 | 0 | 921,094 |
65e4aa71c368fce1336ffe4f
|
INSAIT-Institute/BgGPT-7B-Instruct-v0.2
|
INSAIT-Institute
|
False
|
not-popular-enough
|
2024-03-06T12:01:16.000Z
| 22 | false |
27d7711b0afb0aa2d5edb6a0b5780aff371b662a
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
| 2,629 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"bggpt",
"insait",
"conversational",
"bg",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-03T16:50:57.000Z
|
INSAIT-Institute/BgGPT-7B-Instruct-v0.2
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 1 | 1 | 0 | 2 | 517,981 |
6631258e9fcb8095d5ccbce2
|
johnsnowlabs/JSL-MedLlama-3-8B-v2.0
|
johnsnowlabs
|
False
|
not-popular-enough
|
2024-04-30T18:06:24.000Z
| 22 | false |
f3d41874ba4998c31727b8a18fae3e680f440f2f
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
| 2,619 |
[
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3-8b",
"sft",
"medical",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:cc-by-nc-nd-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-04-30T17:08:30.000Z
|
johnsnowlabs/JSL-MedLlama-3-8B-v2.0
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 7 | 0 | 1 | 613,582 |
66248a1963788fa8b4747e07
|
cognitivecomputations/dolphin-2.9-llama3-8b-gguf
|
cognitivecomputations
|
False
|
library-not-detected
|
2024-05-20T14:43:19.000Z
| 83 | false |
5b5119bdd9fb1fa7f52470a2bbe093c9b89d61e1
| null | 2,575 |
[
"gguf",
"generated_from_trainer",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:abacusai/SystemChat-1.1",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"license:other",
"region:us"
] | null | null |
2024-04-21T03:38:01.000Z
|
cognitivecomputations/dolphin-2.9-llama3-8b-gguf
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q3_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q4_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q5_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q6_K.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q8_0.gguf"
}
] | null |
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 2 | 0 | 596,517 |
66e3a9e88bcd43ff6cbc762e
|
beiyemu/Qwen-Qwen1.5-7B-1726196200
|
beiyemu
|
False
|
pipeline-not-detected
|
2024-09-13T02:56:44.000Z
| 0 | false |
ae7e98d41de394f8b08d613b9c64811930b492c6
|
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}}
| 2,521 |
[
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null |
peft
|
2024-09-13T02:56:40.000Z
|
beiyemu/Qwen-Qwen1.5-7B-1726196200
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null |
Qwen/Qwen1.5-7B
| 6,515 |
text-generation
| 6,515 | 0 | 0 | 0 | 0 | 937,185 |
6661f241b7c705358e201cfe
|
bartowski/Qwen2-7B-Instruct-GGUF
|
bartowski
|
False
|
library-not-detected
|
2024-06-07T01:07:13.000Z
| 15 | false |
f1c7bba77b35fa3f9bc1d470646abdc0c8a9ceef
| null | 2,509 |
[
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"region:us"
] |
text-generation
| null |
2024-06-06T17:30:41.000Z
|
bartowski/Qwen2-7B-Instruct-GGUF
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_XXS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ4_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q2_K.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q4_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q4_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q5_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q5_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q6_K.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q8_0.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-bf16.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-f32.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null |
Qwen/Qwen2-7B-Instruct
| 161 |
text-generation
| 161 | 0 | 0 | 0 | 0 | 686,239 |
65ec08424f2eb01585daf53a
|
Kabster/BioMistral-Zephyr-Beta-SLERP
|
Kabster
|
False
|
not-popular-enough
|
2024-03-09T07:04:51.000Z
| 1 | false |
b6f6be7fa65ed209721e55c6545cb332113a6bd5
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 2,503 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:merge:BioMistral/BioMistral-7B",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:merge:HuggingFaceH4/zephyr-7b-beta",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-09T06:57:06.000Z
|
Kabster/BioMistral-Zephyr-Beta-SLERP
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
BioMistral/BioMistral-7B
| 124 |
text-generation
| 124 | 0 | 0 | 0 | 0 | 526,916 |
65e6b87177658add76a94593
|
Trendyol/Trendyol-LLM-7b-base-v1.0
|
Trendyol
|
False
|
not-popular-enough
|
2024-03-11T06:37:48.000Z
| 14 | false |
327cfb054dd7613371c759456475c97eda59ab1a
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 2,502 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"tr",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-05T06:15:13.000Z
|
Trendyol/Trendyol-LLM-7b-base-v1.0
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trendyol-llm-mistral.jpg"
}
] |
mistral
|
mistralai/Mistral-7B-v0.1
| 2,038 |
text-generation
| 2,038 | 0 | 1 | 2 | 2 | 520,578 |
66d8c989138cd2156c27f39c
|
ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1
|
ytu-ce-cosmos
|
False
|
library-not-detected
|
2024-09-09T10:57:09.000Z
| 24 | false |
5d986d00dec8811e1dfa5eb60d09cc1c36c92d29
|
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
| 2,470 |
[
"safetensors",
"llama",
"Turkish",
"turkish",
"Llama",
"Llama3",
"text-generation",
"conversational",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"region:us"
] |
text-generation
| null |
2024-09-04T20:56:41.000Z
|
ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "cosmosLLaMa2_r2.png"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] |
llama
|
meta-llama/Meta-Llama-3-8B
| 976 |
text-generation
| 976 | 0 | 0 | 0 | 4 | 898,976 |
65ebfaa9e34f51915193ac90
|
Kabster/Bio-Mistralv2-Squared
|
Kabster
|
False
|
not-popular-enough
|
2024-03-09T07:04:57.000Z
| 1 | false |
77aa25e74e78c21e5ede5411d38e819d70d5ba9d
|
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
| 2,458 |
[
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:merge:BioMistral/BioMistral-7B",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:merge:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] |
text-generation
|
transformers
|
2024-03-09T05:59:05.000Z
|
Kabster/Bio-Mistralv2-Squared
|
[
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] |
mistral
|
BioMistral/BioMistral-7B
| 124 |
text-generation
| 124 | 0 | 0 | 0 | 0 | 526,863 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.