Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models Paper • 2412.16545 • Published 5 days ago
A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression Paper • 2412.17483 • Published 3 days ago
Large Language Models for Information Retrieval: A Survey Paper • 2308.07107 • Published Aug 14, 2023 • 3
CORAL: Benchmarking Multi-turn Conversational Retrieval-Augmentation Generation Paper • 2410.23090 • Published Oct 30 • 54