Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
1
12
Robbie Pasquale
RobbiePasquale
Follow
DrishtiSharma's profile picture
sanggusti's profile picture
sliceofham's profile picture
4 followers
·
18 following
robbiepasquale
rpasquale
AI & ML interests
AI
Recent Activity
reacted
to
tomaarsen
's
post
with ❤️
7 days ago
‼️Sentence Transformers v4.0 is out! You can now train and finetune reranker models with multi-GPU training, bf16 support, loss logging, callbacks & much more. I also prove that finetuning on your domain helps much more than you might think. 1️⃣ Reranker Training Refactor Reranker models can now be trained using an extensive trainer with a lot of powerful features: - MultiGPU Training (Data Parallelism (DP) and Distributed Data Parallelism (DDP)) - bf16 training support; loss logging - Evaluation datasets + evaluation loss - Improved callback support + an excellent Weights & Biases integration - Gradient checkpointing, gradient accumulation - Model card generation - Resuming from a training checkpoint without performance loss - Hyperparameter Optimization and much more! Read my detailed blogpost to learn about the components that make up this new training approach: https://huggingface.co/blog/train-reranker Notably, the release is fully backwards compatible: all deprecations are soft, meaning that they still work but emit a warning informing you how to upgrade. 2️⃣ New Reranker Losses - 11 new losses: - 2 traditional losses: BinaryCrossEntropy and CrossEntropy - 2 distillation losses: MSE and MarginMSE - 2 in-batch negatives losses: MNRL (a.k.a. InfoNCE) and CMNRL - 5 learning to rank losses: Lambda, p-ListMLE, ListNet, RankNet, ListMLE 3️⃣ New Reranker Documentation - New Training Overview, Loss Overview, API Reference docs - 5 new, 1 refactored training examples docs pages - 13 new, 6 refactored training scripts - Migration guides (2.x -> 3.x, 3.x -> 4.x) 4️⃣ Blogpost Alongside the release, I've written a blogpost where I finetune ModernBERT on a generic question-answer dataset. My finetunes easily outperform all general-purpose reranker models, even models 4x as big. Finetuning on your domain is definitely worth it: https://huggingface.co/blog/train-reranker See the full release notes here: https://github.com/UKPLab/sentence-transformers/releases/v4.0.1
liked
a model
8 days ago
tomaarsen/reranker-ModernBERT-large-gooaq-bce
updated
a model
4 months ago
RobbiePasquale/gpt-moe-mcts
View all activity
Organizations
RobbiePasquale
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
Articles
liked
a model
8 days ago
tomaarsen/reranker-ModernBERT-large-gooaq-bce
Text Ranking
•
Updated
12 days ago
•
208
•
3
liked
a model
9 months ago
meta-llama/Meta-Llama-3-70B-Instruct
Text Generation
•
Updated
Dec 15, 2024
•
544k
•
•
1.47k
liked
3 datasets
9 months ago
allenai/ultrafeedback_binarized_cleaned
Viewer
•
Updated
Dec 1, 2023
•
186k
•
822
•
69
allenai/reward-bench
Viewer
•
Updated
Sep 9, 2024
•
8.11k
•
6.99k
•
91
nielsr/funsd
Updated
Jul 27, 2021
•
1.21k
•
11
liked
a dataset
about 1 year ago
Intel/orca_dpo_pairs
Viewer
•
Updated
Nov 29, 2023
•
12.9k
•
1.68k
•
299
liked
a model
about 1 year ago
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO-adapter
Updated
Feb 20, 2024
•
15
liked
2 datasets
over 1 year ago
teknium/GPT4-LLM-Cleaned
Viewer
•
Updated
May 4, 2023
•
54.6k
•
1.28k
•
157
legacy-datasets/wikipedia
Updated
Mar 11, 2024
•
33.5k
•
589
liked
a model
over 1 year ago
TinyLlama/TinyLlama-1.1B-Chat-v0.6
Text Generation
•
Updated
Nov 20, 2023
•
13.1k
•
•
98
liked
a dataset
over 1 year ago
teknium/GPTeacher-General-Instruct
Viewer
•
Updated
Apr 29, 2023
•
89.3k
•
148
•
43
liked
a model
over 1 year ago
meta-llama/Llama-2-7b
Text Generation
•
Updated
Apr 17, 2024
•
4.3k