Youtu-LLM: Unlocking the Native Agentic Potential for Lightweight Large Language Models Paper • 2512.24618 • Published 3 days ago • 60
SSA: Sparse Sparse Attention by Aligning Full and Sparse Attention Outputs in Feature Space Paper • 2511.20102 • Published Nov 25, 2025 • 27
CODI: Compressing Chain-of-Thought into Continuous Space via Self-Distillation Paper • 2502.21074 • Published Feb 28, 2025 • 4