AI & ML interests

None defined yet.

Recent Activity

gnr8's activity

abhishekย 
posted an update 26 days ago
view post
Post
1617
๐ŸŽ‰ SUPER BLACK FRIDAY DEAL ๐ŸŽ‰

Train almost any model on a variety of tasks such as llm finetuning, text classification/regression, summarization, question answering, image classification/regression, object detection, tabular data, etc for FREE using AutoTrain locally. ๐Ÿ”ฅ
https://github.com/huggingface/autotrain-advanced
abhishekย 
posted an update about 2 months ago
view post
Post
5410
INTRODUCING Hugging Face AutoTrain Client ๐Ÿ”ฅ
Fine-tuning models got even easier!!!!
Now you can fine-tune SOTA models on all compatible dataset-model pairs on Hugging Face Hub using Python on Hugging Face Servers. Choose from a number of GPU flavors, millions of models and dataset pairs and 10+ tasks ๐Ÿค—

To try, install autotrain-advanced using pip. You can ignore dependencies and install without --no-deps and then you'd need to install some dependencies by hand.

"pip install autotrain-advanced"

Github repo: https://github.com/huggingface/autotrain-advanced
  • 6 replies
ยท
abhishekย 
posted an update 2 months ago
abhishekย 
posted an update 4 months ago
abhishekย 
posted an update 4 months ago
view post
Post
1854
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
Extractive Question Answering: because sometimes generative is not all you need ๐Ÿ˜‰
AutoTrain is the only open-source, no code solution to offer so many tasks across different modalities. Current task count: 23 ๐Ÿš€
Check out the blog post on getting started with this task: https://huggingface.co/blog/abhishek/extractive-qa-autotrain
abhishekย 
posted an update 7 months ago
abhishekย 
posted an update 8 months ago
view post
Post
2935
๐Ÿšจ NEW TASK ALERT ๐Ÿšจ
๐ŸŽ‰ AutoTrain now supports Object Detection! ๐ŸŽ‰
Transform your projects with these powerful new features:
๐Ÿ”น Fine-tune any supported model from the Hugging Face Hub
๐Ÿ”น Seamless logging with TensorBoard or W&B
๐Ÿ”น Support for local and hub datasets
๐Ÿ”น Configurable training for tailored results
๐Ÿ”น Train locally or leverage Hugging Face Spaces
๐Ÿ”น Deployment-ready with API inference or Hugging Face endpoints
AutoTrain: https://hf.co/autotrain
abhishekย 
posted an update 8 months ago
view post
Post
3062
๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€ Introducing AutoTrain Configs! ๐Ÿš€๐Ÿš€๐Ÿš€๐Ÿš€
Now you can train models using yaml config files! ๐Ÿ’ฅ These configs are easy to understand and are not at all overwhelming. So, even a person with almost zero knowledge of machine learning can train state of the art models without writing any code. Check out example configs in the config directory of autotrain-advanced github repo and feel free to share configs by creating a pull request ๐Ÿค—
Github repo: https://github.com/huggingface/autotrain-advanced
  • 2 replies
ยท
abhishekย 
posted an update 8 months ago
abhishekย 
posted an update 8 months ago
view post
Post
2371
Trained another version of llama3-8b-instruct which beats the base model. This time without losing too many points on gsm8k benchmark. Again, using AutoTrain ๐Ÿ’ฅ pip install autotrain-advanced
Trained model: abhishek/autotrain-llama3-orpo-v2
  • 1 reply
ยท
abhishekย 
posted an update 8 months ago
view post
Post
3476
With AutoTrain, you can already finetune the latest llama3 models without writing a single line of code. Here's an example finetune of llama3 8b model: abhishek/autotrain-llama3-no-robots
  • 2 replies
ยท
philschmidย 
posted an update 9 months ago
view post
Post
6895
New state-of-the-art open LLM! ๐Ÿš€ Databricks just released DBRX, a 132B MoE trained on 12T tokens. Claiming to surpass OpenAI GPT-3.5 and is competitive with Google Gemini 1.0 Pro. ๐Ÿคฏ

TL;DR
๐Ÿงฎ 132B MoE with 16 experts with 4 active in generation
๐ŸชŸ 32 000 context window
๐Ÿ“ˆ Outperforms open LLMs on common benchmarks, including MMLU
๐Ÿš€ Up to 2x faster inference than Llama 2 70B
๐Ÿ’ป Trained on 12T tokens
๐Ÿ”ก Uses the GPT-4 tokenizer
๐Ÿ“œ Custom License, commercially useable

Collection: databricks/dbrx-6601c0852a0cdd3c59f71962
Demo: databricks/dbrx-instruct

Kudos to the Team at Databricks and MosaicML for this strong release in the open community! ๐Ÿค—
ยท
philschmidย 
posted an update 11 months ago
view post
Post
What's the best way to fine-tune open LLMs in 2024? Look no further! ๐Ÿ‘€ย I am excited to share โ€œHow to Fine-Tune LLMs in 2024 with Hugging Faceโ€ using the latest research techniques, including Flash Attention, Q-LoRA, OpenAI dataset formats (messages), ChatML, Packing, all built with Hugging Face TRL. ๐Ÿš€

It is created for consumer-size GPUs (24GB) covering the full end-to-end lifecycle with:
๐Ÿ’กDefine and understand use cases for fine-tuning
๐Ÿง‘๐Ÿปโ€๐Ÿ’ปย Setup of the development environment
๐Ÿงฎย Create and prepare dataset (OpenAI format)
๐Ÿ‹๏ธโ€โ™€๏ธย Fine-tune LLM using TRL and the SFTTrainer
๐Ÿฅ‡ย Test and evaluate the LLM
๐Ÿš€ย Deploy for production with TGI

๐Ÿ‘‰ย  https://www.philschmid.de/fine-tune-llms-in-2024-with-trl

Coming soon: Advanced Guides for multi-GPU/multi-Node full fine-tuning and alignment using DPO & KTO. ๐Ÿ”œ
ยท
abhishekย 
posted an update 11 months ago
view post
Post
Happy to announce, brand new, open-source Hugging Face Competitions platform ๐Ÿš€ Now, create a machine learning competition for your friends, colleagues or the world for FREE* and host it on Hugging Face: the AI community building the future. Creating a competition requires only two steps: pip install competitions, then run competitions create and create competition by answering a few questions ๐Ÿ’ฅ Checkout the github repo: https://github.com/huggingface/competitions and docs: https://hf.co/docs/competitions
ยท
abhishekย 
posted an update 12 months ago
abhishekย 
posted an update 12 months ago
view post
Post
Hello Huggers! ๐Ÿค—
ยท