|
base_model: SillyTilly/mistralai_Mistral-Nemo-Base-2407 |
|
dtype: float32 |
|
merge_method: model_stock |
|
slices: |
|
- sources: |
|
- layer_range: [0, 40] |
|
model: nbeerbower/Lyra4-Gutenberg-12B |
|
- layer_range: [0, 40] |
|
model: nbeerbower/mistral-nemo-gutenberg-12B-v4 |
|
- layer_range: [0, 40] |
|
model: elinas/Chronos-Gold-12B-1.0 |
|
- layer_range: [0, 40] |
|
model: UsernameJustAnother/Nemo-12B-Marlin-v8 |
|
- layer_range: [0, 40] |
|
model: TheDrummer/Rocinante-12B-v1.1 |
|
- layer_range: [0, 40] |
|
model: Epiculous/Azure_Dusk-v0.2 |
|
- layer_range: [0, 40] |
|
model: Epiculous/Crimson_Dawn-v0.2 |
|
- layer_range: [0, 40] |
|
model: TheDrummer/Rocinante-12B-v1+jtatman/mistral_nemo_12b_reasoning_psychology_lora |
|
- layer_range: [0, 40] |
|
model: nbeerbower/mistral-nemo-wissenschaft-12B |
|
- layer_range: [0, 40] |
|
model: nbeerbower/mistral-nemo-bophades-12B |
|
- layer_range: [0, 40] |
|
model: anthracite-org/magnum-v2.5-12b-kto+mpasila/Mistral-freeLiPPA-LoRA-12B |
|
- layer_range: [0, 40] |
|
model: nbeerbower/mistral-nemo-cc-12B |
|
- layer_range: [0, 40] |
|
model: anthracite-org/magnum-v2-12b |
|
- layer_range: [0, 40] |
|
model: anthracite-org/magnum-v2.5-12b-kto+jeiku/Aura-NeMo-12B |
|
- layer_range: [0, 40] |
|
model: SillyTilly/mistralai_Mistral-Nemo-Base-2407 |
|
tokenizer_source: unsloth/Mistral-Nemo-Base-2407 |