agentica-org/DeepScaleR-1.5B-Preview Text Generation โข 2B โข Updated Apr 9, 2025 โข 45.3k โข 577
view reply Very cool thanks! I think OpenAI already hate Open Source :)))))Products that are trying so hard to monetize are created in one day.
view post Post 4954 ๐DeepSeek ๐ is the real OpenAI ๐ฏ See translation 6 replies ยท ๐ฅ 28 28 + Reply
view reply Yes, it's been tested, and it's false. It's even worse than the regular LLAMA 3.1 70b. It's even funny to compare it to Claude.https://www.reddit.com/r/LocalLLaMA/s/BH5A2ngyui
imone/Llama-3-8B-fixed-special-embedding Text Generation โข 8B โข Updated Apr 25, 2024 โข 6 โข 17
view reply Yeah, I was thinking the same thing. A large vocabulary does improve the performance of smaller LLMs and judging by the GPT-4o the same is true for larger LLM. Give it a try. I'm just doing this for small size models up to 3B parameters.