Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region Paper • 2502.13946 • Published 22 days ago • 9
Instruct Once, Chat Consistently in Multiple Rounds: An Efficient Tuning Framework for Dialogue Paper • 2402.06967 • Published Feb 10, 2024
TokenSkip: Controllable Chain-of-Thought Compression in LLMs Paper • 2502.12067 • Published 24 days ago • 1
TokenSkip: Controllable Chain-of-Thought Compression in LLMs Paper • 2502.12067 • Published 24 days ago • 1
Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region Paper • 2502.13946 • Published 22 days ago • 9
Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region Paper • 2502.13946 • Published 22 days ago • 9 • 2
Running 2.23k 2.23k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters
Self-Detoxifying Language Models via Toxification Reversal Paper • 2310.09573 • Published Oct 14, 2023
E2CL: Exploration-based Error Correction Learning for Embodied Agents Paper • 2409.03256 • Published Sep 5, 2024 • 1
Subtle Errors Matter: Preference Learning via Error-injected Self-editing Paper • 2410.06638 • Published Oct 9, 2024
Direct Preference Optimization Using Sparse Feature-Level Constraints Paper • 2411.07618 • Published Nov 12, 2024 • 16
Direct Preference Optimization Using Sparse Feature-Level Constraints Paper • 2411.07618 • Published Nov 12, 2024 • 16