netzkontrast 's Collections Performance
updated
ShortGPT: Layers in Large Language Models are More Redundant Than You
Expect
Paper
• 2403.03853
• Published
• 66
SLEB: Streamlining LLMs through Redundancy Verification and Elimination
of Transformer Blocks
Paper
• 2402.09025
• Published
• 10
Shortened LLaMA: A Simple Depth Pruning for Large Language Models
Paper
• 2402.02834
• Published
• 17
Algorithmic progress in language models
Paper
• 2403.05812
• Published
• 19
Larimar: Large Language Models with Episodic Memory Control
Paper
• 2403.11901
• Published
• 33
Qihoo-T2X: An Efficiency-Focused Diffusion Transformer via Proxy Tokens
for Text-to-Any-Task
Paper
• 2409.04005
• Published
• 19
Self-Discover: Large Language Models Self-Compose Reasoning Structures
Paper
• 2402.03620
• Published
• 117
Chain-of-Thought Reasoning Without Prompting
Paper
• 2402.10200
• Published
• 109
CAS-ViT: Convolutional Additive Self-attention Vision Transformers for
Efficient Mobile Applications
Paper
• 2408.03703
• Published
LLaVA-Mini: Efficient Image and Video Large Multimodal Models with One
Vision Token
Paper
• 2501.03895
• Published
• 52