view article Article Illustrating Reinforcement Learning from Human Feedback (RLHF) Dec 9, 2022 • 197
Pairwise RM: Perform Best-of-N Sampling with Knockout Tournament Paper • 2501.13007 • Published Jan 22 • 20
RM-Bench: Benchmarking Reward Models of Language Models with Subtlety and Style Paper • 2410.16184 • Published Oct 21, 2024 • 24