Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? 😃
Here's a cheat sheet for devs (but open to anyone!)
---
TL;DR
- Full Fine-Tuning: Max performance, high resource needs, best reliability. - PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML. - Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT. - RAFT: Best for fact-grounded models with dynamic retrieval. - RLHF: Produces ethical, high-quality conversational AI, but expensive.
Choose wisely and match your approach to your task, budget, and deployment constraints.
I just posted the full extended article here if you want to continue reading >>>
_______________________ Second: Quazim0t0/TXTAgent Created an Agent that converts a .txt file into a CSV file, then you can ask about the data and also download the CSV file that was generated.
_______________________ Third: Quazim0t0/ReportAgent Upload Multiple TXT/DOC files to then generate a report from those files.
_______________________ Lastly: Quazim0t0/qResearch A Research tool that uses DuckDuckGo for Web Searches, Wikipedia and tries to refine the answers in MLA Format.
This dataset was collected in roughly 4 hours using the Rapidata Python API, showcasing how quickly large-scale annotations can be performed with the right tooling!
All that at less than the cost of a single hour of a typical ML engineer in Zurich!
The new dataset of ~22,000 human annotations evaluating AI-generated videos based on different dimensions, such as Prompt-Video Alignment, Word for Word Prompt Alignment, Style, Speed of Time flow and Quality of Physics.