
sometimesanotion PRO
AI & ML interests
Recent Activity
Organizations
sometimesanotion's activity
Censored
It's been a wild ride, folks :) (end of the Open LLM Leaderboard)


@Inschrift-Spruch-Raum ,I am looking through recent PRs to mergekit, and I am optimistic that Lamarck's recipes will be working again soon!
When that happens, there will be two efforts: one to make a compelling non-CoT model, and another to blend CoT in right amounts.
Lamarck's multilingual capabilities improved noticeably from light influence of Krystalan/DRT-14B in v0.6, and merging from other CoT models like DeepSeek R1 is a matter of careful moderation. I will always put the overall apparent quality of translation, prose, and reasoning first.

No worries! See, I agree, the recipe behind Lamarck is pretty good, and there's a lot more to get out of it. It'll likely depend on getting multiple mergekit versions working on the pipeline. The new mergekit's fusion and sce merges offer some interesting potential, but I use fine-grained sliced merges to control the mix of branches, which last I checked, work only with older mergekit and bitsnbytes.
By now there are ample upgrades to try. I did feel Lamarck v0.7 was a proof-of-concept and had plenty of headroom to grow!

Yes, there's a reason for that! I ran into a mergekit bug with the use of slices and della_linear merge methods which are key to intentionally crafting Lamarck's releases. @Crystalcareai , @Arcee , are there fixes in the queue?
https://huggingface.co/posts/sometimesanotion/507492798113402
Fusion vs. SLERP?


Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v8

TimeLordRaps/DS-R1-Lamarckvergence-14B-1M-test3

microsoft/Phi-4-mini-instruct
YOYO-AI/Qwen2.5-14B-YOYO-V4-p2

Lunzima/NQLSG-Qwen2.5-14B-OriginalFusion
