-
Video Creation by Demonstration
Paper • 2412.09551 • Published • 8 -
DiffSensei: Bridging Multi-Modal LLMs and Diffusion Models for Customized Manga Generation
Paper • 2412.07589 • Published • 45 -
Unraveling the Complexity of Memory in RL Agents: an Approach for Classification and Evaluation
Paper • 2412.06531 • Published • 71 -
APOLLO: SGD-like Memory, AdamW-level Performance
Paper • 2412.05270 • Published • 38
Collections
Discover the best community collections!
Collections including paper arxiv:2411.17116
-
Beyond Examples: High-level Automated Reasoning Paradigm in In-Context Learning via MCTS
Paper • 2411.18478 • Published • 32 -
o1-Coder: an o1 Replication for Coding
Paper • 2412.00154 • Published • 41 -
A Simple and Provable Scaling Law for the Test-Time Compute of Large Language Models
Paper • 2411.19477 • Published • 5 -
Reverse Thinking Makes LLMs Stronger Reasoners
Paper • 2411.19865 • Published • 19
-
Natural Language Reinforcement Learning
Paper • 2411.14251 • Published • 26 -
The Impossible Test: A 2024 Unsolvable Dataset and A Chance for an AGI Quiz
Paper • 2411.14486 • Published • 7 -
Star Attention: Efficient LLM Inference over Long Sequences
Paper • 2411.17116 • Published • 47 -
Beyond Examples: High-level Automated Reasoning Paradigm in In-Context Learning via MCTS
Paper • 2411.18478 • Published • 32
-
Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering
Paper • 2411.11504 • Published • 19 -
Top-nσ: Not All Logits Are You Need
Paper • 2411.07641 • Published • 18 -
Adaptive Decoding via Latent Preference Optimization
Paper • 2411.09661 • Published • 10 -
When Precision Meets Position: BFloat16 Breaks Down RoPE in Long-Context Training
Paper • 2411.13476 • Published • 15
-
Selective Attention Improves Transformer
Paper • 2410.02703 • Published • 23 -
Differential Transformer
Paper • 2410.05258 • Published • 168 -
TidalDecode: Fast and Accurate LLM Decoding with Position Persistent Sparse Attention
Paper • 2410.05076 • Published • 7 -
SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs
Paper • 2410.13276 • Published • 25
-
Revisit Large-Scale Image-Caption Data in Pre-training Multimodal Foundation Models
Paper • 2410.02740 • Published • 52 -
From Code to Correctness: Closing the Last Mile of Code Generation with Hierarchical Debugging
Paper • 2410.01215 • Published • 30 -
Molmo and PixMo: Open Weights and Open Data for State-of-the-Art Multimodal Models
Paper • 2409.17146 • Published • 104 -
EuroLLM: Multilingual Language Models for Europe
Paper • 2409.16235 • Published • 25