--- base_model: LumiOpen/Viking-7B language: - en - fi - sv - 'no' - da - is - nn license: apache-2.0 tags: - text-generation-inference - transformers - unsloth - llama - trl datasets: - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned - mpasila/Sonnet3.5-SlimOrcaDedupCleaned-4k-context library_name: peft --- This is the fully trained version (with fixed formatting!!). Dataset used: [Gryphe/Sonnet3.5-SlimOrcaDedupCleaned](https://huggingface.co/datasets/Gryphe/Sonnet3.5-SlimOrcaDedupCleaned) which was further [filtered](https://huggingface.co/datasets/mpasila/Sonnet3.5-SlimOrcaDedupCleaned-4k-context) to remove prompts/examples that are longer than 4076 tokens (removed about 385 examples). Prompt format is: ChatML Merged model: [mpasila/Viking-SlimSonnet-v1-7B](https://huggingface.co/mpasila/Viking-SlimSonnet-v1-7B) Trained with regular LoRA (not quantized/QLoRA) and LoRA rank was 128 and Alpha set to 32. Trained for 1 epoch using A40 for about 23 hours. # Uploaded model - **Developed by:** mpasila - **License:** apache-2.0 - **Finetuned from model :** LumiOpen/Viking-7B This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [](https://github.com/unslothai/unsloth)