Update config.json
Browse files- config.json +10 -0
config.json
CHANGED
@@ -28,6 +28,16 @@
|
|
28 |
"transformers_version": "4.41.2",
|
29 |
"use_cache": true,
|
30 |
"vocab_size": 32000,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
|
32 |
"test_set": [
|
33 |
{
|
|
|
28 |
"transformers_version": "4.41.2",
|
29 |
"use_cache": true,
|
30 |
"vocab_size": 32000,
|
31 |
+
"prompt_wrapper": "human_bot",
|
32 |
+
"prompt_format": "<human> {context_passage} \n {question} \n<bot>:",
|
33 |
+
"prompt_format_dict": {"main_start": "<human>: ", "main_stop": "\n", "start_llm_response": "<bot>:"},
|
34 |
+
"tokenizer_local": "tokenizer_tl.json",
|
35 |
+
"tokenizer_config": {"bos_id":[1], "bos_token":["<s>"], "eos_id":[2],"eos_token":["</s>"]},
|
36 |
+
"model_parent": "llmware/slim-sql-v0",
|
37 |
+
"description": "Fact-based question-answering model from llmware - finetuned on tiny-llama - 1.1 parameter base",
|
38 |
+
"quantization": "int4",
|
39 |
+
"model_family": "OVGenerativeModel",
|
40 |
+
"parameters": 1.1,
|
41 |
|
42 |
"test_set": [
|
43 |
{
|