AutoTrain Projects

Activity Feed

AI & ML interests

Official projects from Hugging Face AutoTrain

Recent Activity

autotrain-projects's activity

abhishekΒ 
posted an update 25 days ago
view post
Post
1615
πŸŽ‰ SUPER BLACK FRIDAY DEAL πŸŽ‰

Train almost any model on a variety of tasks such as llm finetuning, text classification/regression, summarization, question answering, image classification/regression, object detection, tabular data, etc for FREE using AutoTrain locally. πŸ”₯
https://github.com/huggingface/autotrain-advanced
abhishekΒ 
posted an update about 2 months ago
view post
Post
5406
INTRODUCING Hugging Face AutoTrain Client πŸ”₯
Fine-tuning models got even easier!!!!
Now you can fine-tune SOTA models on all compatible dataset-model pairs on Hugging Face Hub using Python on Hugging Face Servers. Choose from a number of GPU flavors, millions of models and dataset pairs and 10+ tasks πŸ€—

To try, install autotrain-advanced using pip. You can ignore dependencies and install without --no-deps and then you'd need to install some dependencies by hand.

"pip install autotrain-advanced"

Github repo: https://github.com/huggingface/autotrain-advanced
  • 6 replies
Β·
abhishekΒ 
posted an update 2 months ago
abhishekΒ 
posted an update 4 months ago
abhishekΒ 
posted an update 4 months ago
view post
Post
1854
🚨 NEW TASK ALERT 🚨
Extractive Question Answering: because sometimes generative is not all you need πŸ˜‰
AutoTrain is the only open-source, no code solution to offer so many tasks across different modalities. Current task count: 23 πŸš€
Check out the blog post on getting started with this task: https://huggingface.co/blog/abhishek/extractive-qa-autotrain
multimodalartΒ 
posted an update 5 months ago
abhishekΒ 
posted an update 7 months ago
abhishekΒ 
posted an update 7 months ago
view post
Post
2935
🚨 NEW TASK ALERT 🚨
πŸŽ‰ AutoTrain now supports Object Detection! πŸŽ‰
Transform your projects with these powerful new features:
πŸ”Ή Fine-tune any supported model from the Hugging Face Hub
πŸ”Ή Seamless logging with TensorBoard or W&B
πŸ”Ή Support for local and hub datasets
πŸ”Ή Configurable training for tailored results
πŸ”Ή Train locally or leverage Hugging Face Spaces
πŸ”Ή Deployment-ready with API inference or Hugging Face endpoints
AutoTrain: https://hf.co/autotrain
multimodalartΒ 
posted an update 7 months ago
view post
Post
25181
The first open Stable Diffusion 3-like architecture model is JUST out πŸ’£ - but it is not SD3! πŸ€”

It is Tencent-Hunyuan/HunyuanDiT by Tencent, a 1.5B parameter DiT (diffusion transformer) text-to-image model πŸ–ΌοΈβœ¨, trained with multi-lingual CLIP + multi-lingual T5 text-encoders for english 🀝 chinese understanding

Try it out by yourself here ▢️ https://huggingface.co/spaces/multimodalart/HunyuanDiT
(a bit too slow as the model is chunky and the research code isn't super optimized for inference speed yet)

In the paper they claim to be SOTA open source based on human preference evaluation!
abhishekΒ 
posted an update 8 months ago
view post
Post
3062
πŸš€πŸš€πŸš€πŸš€ Introducing AutoTrain Configs! πŸš€πŸš€πŸš€πŸš€
Now you can train models using yaml config files! πŸ’₯ These configs are easy to understand and are not at all overwhelming. So, even a person with almost zero knowledge of machine learning can train state of the art models without writing any code. Check out example configs in the config directory of autotrain-advanced github repo and feel free to share configs by creating a pull request πŸ€—
Github repo: https://github.com/huggingface/autotrain-advanced
  • 2 replies
Β·
abhishekΒ 
posted an update 8 months ago
abhishekΒ 
posted an update 8 months ago
view post
Post
2371
Trained another version of llama3-8b-instruct which beats the base model. This time without losing too many points on gsm8k benchmark. Again, using AutoTrain πŸ’₯ pip install autotrain-advanced
Trained model: abhishek/autotrain-llama3-orpo-v2
  • 1 reply
Β·
abhishekΒ 
posted an update 8 months ago
view post
Post
3476
With AutoTrain, you can already finetune the latest llama3 models without writing a single line of code. Here's an example finetune of llama3 8b model: abhishek/autotrain-llama3-no-robots
  • 2 replies
Β·
multimodalartΒ 
posted an update 10 months ago
view post
Post
The Stable Diffusion 3 research paper broken down, including some overlooked details! πŸ“

Model
πŸ“ 2 base model variants mentioned: 2B and 8B sizes

πŸ“ New architecture in all abstraction levels:
- πŸ”½ UNet; ⬆️ Multimodal Diffusion Transformer, bye cross attention πŸ‘‹
- πŸ†• Rectified flows for the diffusion process
- 🧩 Still a Latent Diffusion Model

πŸ“„ 3 text-encoders: 2 CLIPs, one T5-XXL; plug-and-play: removing the larger one maintains competitiveness

πŸ—ƒοΈ Dataset was deduplicated with SSCD which helped with memorization (no more details about the dataset tho)

Variants
πŸ” A DPO fine-tuned model showed great improvement in prompt understanding and aesthetics
✏️ An Instruct Edit 2B model was trained, and learned how to do text-replacement

Results
βœ… State of the art in automated evals for composition and prompt understanding
βœ… Best win rate in human preference evaluation for prompt understanding, aesthetics and typography (missing some details on how many participants and the design of the experiment)

Paper: https://stabilityai-public-packages.s3.us-west-2.amazonaws.com/Stable+Diffusion+3+Paper.pdf
Β·
multimodalartΒ 
posted an update 10 months ago
multimodalartΒ 
posted an update 11 months ago
view post
Post
It seems February started with a fully open source AI renaissance 🌟

Models released with fully open dataset, training code, weights βœ…

LLM - allenai/olmo-suite-65aeaae8fe5b6b2122b46778 🧠
Embedding - nomic-ai/nomic-embed-text-v1 πŸ“š (sota!)

And it's literally February 1st - can't wait to see what else the community will bring πŸ‘€