--- license: apache-2.0 library_name: transformers tags: - mergekit - merge base_model: - Krystalan/DRT-o1-14B - v000000/Qwen2.5-Lumen-14B model-index: - name: Herodotos-14B results: - task: type: text-generation name: Text Generation dataset: name: IFEval (0-Shot) type: HuggingFaceH4/ifeval args: num_few_shot: 0 metrics: - type: inst_level_strict_acc and prompt_level_strict_acc value: 46.67 name: strict accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: BBH (3-Shot) type: BBH args: num_few_shot: 3 metrics: - type: acc_norm value: 48.91 name: normalized accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MATH Lvl 5 (4-Shot) type: hendrycks/competition_math args: num_few_shot: 4 metrics: - type: exact_match value: 45.77 name: exact match source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GPQA (0-shot) type: Idavidrein/gpqa args: num_few_shot: 0 metrics: - type: acc_norm value: 16.44 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MuSR (0-shot) type: TAUR-Lab/MuSR args: num_few_shot: 0 metrics: - type: acc_norm value: 19.88 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU-PRO (5-shot) type: TIGER-Lab/MMLU-Pro config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 47.67 name: accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Triangle104/Herodotos-14B name: Open LLM Leaderboard --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ![image/webp](https://cdn-uploads.huggingface.co/production/uploads/66c1cc08453a7ef6c5fe657a/JZgPH1YpPt8YRjb61eglK.webp) *We're all stories in the end* ### Merge Method This model was merged using the SLERP merge method. ### Models Merged The following models were included in the merge: * [Krystalan/DRT-o1-14B](https://huggingface.co/Krystalan/DRT-o1-14B) * [v000000/Qwen2.5-Lumen-14B](https://huggingface.co/v000000/Qwen2.5-Lumen-14B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: Krystalan/DRT-o1-14B - model: v000000/Qwen2.5-Lumen-14B merge_method: slerp base_model: v000000/Qwen2.5-Lumen-14B dtype: bfloat16 parameters: t: [0, 0.5, 1, 0.5, 0] ``` # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/Triangle104__Herodotos-14B-details) | Metric |Value| |-------------------|----:| |Avg. |37.56| |IFEval (0-Shot) |46.67| |BBH (3-Shot) |48.91| |MATH Lvl 5 (4-Shot)|45.77| |GPQA (0-shot) |16.44| |MuSR (0-shot) |19.88| |MMLU-PRO (5-shot) |47.67|