|
{ |
|
"Model": "CogniFusion-XTTS-slerp", |
|
"Description": "CogniFusion-XTTS-slerp is a merge of the following models using LazyMergekit: Or4cl3-1/cognitive-agent-xtts-optimized and Or4cl3-1/multimodal-fusion-optimized", |
|
"Configuration": { |
|
"slices": [ |
|
{ |
|
"sources": [ |
|
{ |
|
"model": "Or4cl3-1/cognitive-agent-xtts-optimized", |
|
"layer_range": [0, 32] |
|
}, |
|
{ |
|
"model": "Or4cl3-1/multimodal-fusion-optimized", |
|
"layer_range": [0, 32] |
|
} |
|
] |
|
} |
|
], |
|
"merge_method": "slerp", |
|
"base_model": "Or4cl3-1/cognitive-agent-xtts-optimized", |
|
"parameters": { |
|
"t": [ |
|
{ |
|
"filter": "self_attn", |
|
"value": [0, 0.5, 0.3, 0.7, 1] |
|
}, |
|
{ |
|
"filter": "mlp", |
|
"value": [1, 0.5, 0.7, 0.3, 0] |
|
}, |
|
{ |
|
"value": 0.5 |
|
} |
|
] |
|
}, |
|
"dtype": "bfloat16" |
|
}, |
|
"Ethical Considerations": "Add ethical considerations and any additional optimization parameters here", |
|
"Usage": { |
|
"Installation": "!pip install -qU transformers accelerate", |
|
"Code": [ |
|
"from transformers import AutoTokenizer", |
|
"import transformers", |
|
"import torch", |
|
"model = \"Or4cl3-1/CogniFusion-XTTS-slerp\"", |
|
"messages = [{\"role\": \"user\", \"content\": \"What is a large language model?\"}]", |
|
"tokenizer = AutoTokenizer.from_pretrained(model)", |
|
"prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)", |
|
"pipeline = transformers.pipeline(", |
|
" \"text-generation\",", |
|
" model=model,", |
|
" torch_dtype=torch.float16,", |
|
" device_map=\"auto\",", |
|
")", |
|
"outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)", |
|
"print(outputs[0][\"generated_text\"])" |
|
] |
|
} |
|
} |