-
DataComp-LM: In search of the next generation of training sets for language models
Paper • 2406.11794 • Published • 48 -
Training Language Models on Synthetic Edit Sequences Improves Code Synthesis
Paper • 2410.02749 • Published • 12 -
Fewer Truncations Improve Language Modeling
Paper • 2404.10830 • Published • 3 -
How to Train Long-Context Language Models (Effectively)
Paper • 2410.02660 • Published
Yi Cui
yicui
·
AI & ML interests
None yet
Organizations
None yet
Collections
7
-
Is In-Context Learning Sufficient for Instruction Following in LLMs?
Paper • 2405.19874 • Published -
Implicit In-context Learning
Paper • 2405.14660 • Published -
Why does in-context learning fail sometimes? Evaluating in-context learning on open and closed questions
Paper • 2407.02028 • Published -
Towards Better Understanding of In-Context Learning Ability from In-Context Uncertainty Quantification
Paper • 2405.15115 • Published