Data Is Better Together

community
Activity Feed

AI & ML interests

Building better datasets together

Recent Activity

data-is-better-together's activity

davidberenstein1957ย 
posted an update about 9 hours ago
burtenshawย 
posted an update 1 day ago
view post
Post
1554
Iโ€™m super excited to work with @mlabonne to build the first practical example in the reasoning course.

๐Ÿ”— https://huggingface.co/reasoning-course

Here's a quick walk through of the first drop of material that works toward the use case:

- a fundamental introduction to reinforcement learning. Answering questions like, โ€˜what is a reward?โ€™ and โ€˜how do we create an environment for a language model?โ€™

- Then it focuses on Deepseek R1 by walking through the paper and highlighting key aspects. This is an old school way to learn ML topics, but it always works.

- Next, it takes to you Transformers Reinforcement Learning and demonstrates potential reward functions you could use. This is cool because it uses Marimo notebooks to visualise the reward.

- Finally, Maxime walks us through a real training notebook that uses GRPO to reduce generation length. Iโ€™m really into this because it works and Maxime took the time to validate it share assets and logging from his own runs for you to compare with.

Maximeโ€™s work and notebooks have been a major part of the open source community over the last few years. I, like everyone, have learnt so much from them.
davidberenstein1957ย 
posted an update 2 days ago
view post
Post
3614
๐ŸฅŠ Epic Agent Framework Showdown! Available today!

๐Ÿ”ต In the blue corner, the versatile challenger with a proven track record of knowledge retrieval: LlamaIndex!

๐Ÿ›‘ In the red corner, the defender, weighing in with lightweight efficiency: Hugging Face smolagents!

๐Ÿ”— URL: https://huggingface.co/agents-course

We just published the LlamaIndex unit for the agents course, and it is set to offer a great contrast between the smolagents unit by looking at

- What makes llama-index stand-out
- How the LlamaHub is used for integrations
- Creating QueryEngine components
- Using agents and tools
- Agentic and multi-agent workflows

The team has been working flat-out on this for a few weeks. Supported by Logan Markewich and Laurie Voss over at LlamaIndex.

Who won? You decide!
davidberenstein1957ย 
posted an update 2 days ago
view post
Post
2849
๐Ÿซธ New release to push vector search to the Hub with vicinity and work with any serialisable objects.

๐Ÿง‘โ€๐Ÿซ KNN, HNSW, USEARCH, ANNOY, PYNNDESCENT, FAISS, and VOYAGER.

๐Ÿ”— Example Repo: minishlab/my-vicinity-repo
davanstrienย 
posted an update 6 days ago
view post
Post
2552
๐Ÿ“Š Introducing "Hugging Face Dataset Spotlight" ๐Ÿ“Š

I'm excited to share the first episode of our AI-generated podcast series focusing on nice datasets from the Hugging Face Hub!

This first episode explores mathematical reasoning datasets:

- SynthLabsAI/Big-Math-RL-Verified: Over 250,000 rigorously verified problems spanning multiple difficulty levels and mathematical domains
- open-r1/OpenR1-Math-220k: 220,000 math problems with multiple reasoning traces, verified for accuracy using Math Verify and Llama-3.3-70B models.
- facebook/natural_reasoning: 1.1 million general reasoning questions carefully deduplicated and decontaminated from existing benchmarks, showing superior scaling effects when training models like Llama3.1-8B-Instruct.

Plus a bonus segment on bespokelabs/bespoke-manim!

https://www.youtube.com/watch?v=-TgmRq45tW4
davanstrienย 
posted an update 7 days ago
view post
Post
3548
Quick POC: Turn a Hugging Face dataset card into a short podcast introducing the dataset using all open models.

I think I'm the only weirdo who would enjoy listening to something like this though ๐Ÿ˜…

Here is an example for eth-nlped/stepverify
  • 2 replies
ยท
burtenshawย 
posted an update 7 days ago
view post
Post
5201
I made a real time voice agent with FastRTC, smolagents, and hugging face inference providers. Check it out in this space:

๐Ÿ”— burtenshaw/coworking_agent
ยท
burtenshawย 
posted an update 8 days ago
view post
Post
6019
Now the Hugging Face agent course is getting real! With frameworks like smolagents, LlamaIndex, and LangChain.

๐Ÿ”— Follow the org for updates https://huggingface.co/agents-course

This week we are releasing the first framework unit in the course and itโ€™s on smolagents. This is what the unit covers:

- why should you use smolagents vs another library?
- how to build agents that use code
- build multiagents systems
- use vision language models for browser use

The team has been working flat out on this for a few weeks. Led by @sergiopaniego and supported by smolagents author @m-ric .
alvarobarttย 
posted an update 9 days ago
view post
Post
2748
๐Ÿ”ฅ Agents can do anything! @microsoft Research just announced the release of Magma 8B!

Magma is a new Visual Language Model (VLM) with 8B parameters for multi-modal agents designed to handle complex interactions across virtual and real environments; and it's MIT licensed!

Magma comes with exciting new features such as:
- Introduces the Set-of-Mark and Trace-of-Mark techniques for fine-tuning
- Leverages a large amount of unlabeled video data to learn the spatial-temporal grounding and planning
- A strong generalization and ability to be fine-tuned for other agentic tasks
- SOTA in different multi-modal benchmarks spanning across UI navigation, robotics manipulation, image / video understanding and spatial understanding and reasoning
- Generates goal-driven visual plans and actions for agentic use cases

Model: microsoft/Magma-8B
Technical Report: Magma: A Foundation Model for Multimodal AI Agents (2502.13130)
davanstrienย 
posted an update 14 days ago
view post
Post
2558
Hacked together a way to log trl GRPO training completions to a ๐Ÿค— dataset repo. This allows you to:

- Track rewards from multiple reward functions
- Treat the completion and rewards from training as a "proper" dataset and do EDA
- Share results for open science

The implementation is super hacky, but I'm curious if people would find this useful.

To push completions to the Hub, you just need two extra parameters:

log_completions=True
log_completions_hub_repo='your-username/repo-name'

Example dataset: davanstrien/test-logs
Colab: https://colab.research.google.com/drive/1wzBFPVthRYYTp-mEYlznLg_e_0Za1M3g

burtenshawย 
posted an update 15 days ago
view post
Post
7153
AGENTS + FINETUNING! This week Hugging Face learn has a whole pathway on finetuning for agentic applications. You can follow these two courses to get knowledge on levelling up your agent game beyond prompts:

1๏ธโƒฃ New Supervised Fine-tuning unit in the NLP Course https://huggingface.co/learn/nlp-course/en/chapter11/1
2๏ธโƒฃNew Finetuning for agents bonus module in the Agents Course https://huggingface.co/learn/agents-course/bonus-unit1/introduction

Fine-tuning will squeeze everything out of your model for how youโ€™re using it, more than any prompt.
  • 2 replies
ยท
sayakpaulย 
posted an update 17 days ago
view post
Post
2989
Inference-time scaling meets Flux.1-Dev (and others) ๐Ÿ”ฅ

Presenting a simple re-implementation of "Inference-time scaling diffusion models beyond denoising steps" by Ma et al.

I did the simplest random search strategy, but results can potentially be improved with better-guided search methods.

Supports Gemini 2 Flash & Qwen2.5 as verifiers for "LLMGrading" ๐Ÿค—

The steps are simple:

For each round:

1> Starting by sampling 2 starting noises with different seeds.
2> Score the generations w.r.t a metric.
3> Obtain the best generation from the current round.

If you have more compute budget, go to the next search round. Scale the noise pool (2 ** search_round) and repeat 1 - 3.

This constitutes the random search method as done in the paper by Google DeepMind.

Code, more results, and a bunch of other stuff are in the repository. Check it out here: https://github.com/sayakpaul/tt-scale-flux/ ๐Ÿค—
burtenshawย 
posted an update 17 days ago
view post
Post
3441
NEW COURSE! Weโ€™re cooking hard on Hugging Face courses, and itโ€™s not just agents. The NLP course is getting the same treatment with a new chapter on Supervised Fine-Tuning!

๐Ÿ‘‰ Follow to get more updates https://huggingface.co/nlp-course

The new SFT chapter will guide you through these topics:

1๏ธโƒฃ Chat Templates: Master the art of structuring AI conversations for consistent and helpful responses.

2๏ธโƒฃ Supervised Fine-Tuning (SFT): Learn the core techniques to adapt pre-trained models to your specific outputs.

3๏ธโƒฃ Low Rank Adaptation (LoRA): Discover efficient fine-tuning methods that save memory and resources.

4๏ธโƒฃ Evaluation: Measure your model's performance and ensure top-notch results.

This is the first update in a series, so follow along if youโ€™re upskilling in AI.
  • 2 replies
ยท
davanstrienย 
posted an update 18 days ago
davanstrienย 
posted an update 20 days ago
view post
Post
1894
How do you make 1M+ Hugging Face models & datasets more discoverable?

davanstrien/Smol-Hub-tldr!

I fine-tuned HuggingFaceTB/SmolLM2-360M to generate one-line summaries from a model or dataset README.

Its own self-description?
"A model for generating concise summaries of model & dataset cards from the Hugging Face Hub"

The goal? Make it easier to find the right models and datasets for your specific needs. It's already powering a semantic search for datasets Space.

It's still a WIP but thanks to @loubnabnl , @anton-l , @eliebak et al, for cooking such a nice base model for fine-tuning small, efficient models for specific domains and tasks. ๐Ÿ™
burtenshawย 
posted an update 20 days ago
view post
Post
3535
Hey, Iโ€™m Ben and I work at Hugging Face.

Right now, Iโ€™m focusing on educational stuff and getting loads of new people to build open AI models using free and open source tools.

Iโ€™ve made a collection of some of the tools Iโ€™m building and using for teaching. Stuff like quizzes, code challenges, and certificates.

burtenshaw/tools-for-learning-ai-6797453caae193052d3638e2
  • 1 reply
ยท
davanstrienย 
posted an update 21 days ago
davidberenstein1957ย 
posted an update 22 days ago
view post
Post
3271
๐Ÿš€ Find banger tools for your smolagents!

I created the Tools gallery, which makes tools specifically developed by/for smolagents searchable and visible. This will help with:
- inspiration
- best practices
- finding cool tools

Space: davidberenstein1957/smolagents-and-tools
  • 1 reply
ยท
burtenshawย 
posted an update 23 days ago
view post
Post
9084
The Hugging Face agents course is finally out!

๐Ÿ‘‰ https://huggingface.co/agents-course

This first unit of the course sets you up with all the fundamentals to become a pro in agents.

- What's an AI Agent?
- What are LLMs?
- Messages and Special Tokens
- Understanding AI Agents through the Thought-Action-Observation Cycle
- Thought, Internal Reasoning and the Re-Act Approach
- Actions, Enabling the Agent to Engage with Its Environment
- Observe, Integrating Feedback to Reflect and Adapt