Glyph: Scaling Context Windows via Visual-Text Compression Paper β’ 2510.17800 β’ Published 13 days ago β’ 64
Language Models Can Learn from Verbal Feedback Without Scalar Rewards Paper β’ 2509.22638 β’ Published Sep 26 β’ 67
GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models Paper β’ 2508.06471 β’ Published Aug 8 β’ 188
GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable Reinforcement Learning Paper β’ 2507.01006 β’ Published Jul 1 β’ 237
LongWriter-Zero: Mastering Ultra-Long Text Generation via Reinforcement Learning Paper β’ 2506.18841 β’ Published Jun 23 β’ 56
SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models Paper β’ 2506.04180 β’ Published Jun 4 β’ 33
DAPO: An Open-Source LLM Reinforcement Learning System at Scale Paper β’ 2503.14476 β’ Published Mar 18 β’ 141
Light-R1: Curriculum SFT, DPO and RL for Long COT from Scratch and Beyond Paper β’ 2503.10460 β’ Published Mar 13 β’ 29
Shifting Long-Context LLMs Research from Input to Output Paper β’ 2503.04723 β’ Published Mar 6 β’ 22
^RFLAV: Rolling Flow matching for infinite Audio Video generation Paper β’ 2503.08307 β’ Published Mar 11 β’ 9
REF-VLM: Triplet-Based Referring Paradigm for Unified Visual Decoding Paper β’ 2503.07413 β’ Published Mar 10 β’ 2
What's in a Latent? Leveraging Diffusion Latent Space for Domain Generalization Paper β’ 2503.06698 β’ Published Mar 9 β’ 4
NeuGrasp: Generalizable Neural Surface Reconstruction with Background Priors for Material-Agnostic Object Grasp Detection Paper β’ 2503.03511 β’ Published Mar 5 β’ 2
Beyond Decoder-only: Large Language Models Can be Good Encoders for Machine Translation Paper β’ 2503.06594 β’ Published Mar 9 β’ 6
LongWriter-V: Enabling Ultra-Long and High-Fidelity Generation in Vision-Language Models Paper β’ 2502.14834 β’ Published Feb 20 β’ 24
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper β’ 2502.11089 β’ Published Feb 16 β’ 165
LongGenBench: Long-context Generation Benchmark Paper β’ 2410.04199 β’ Published Oct 5, 2024 β’ 22
MiniMax-01: Scaling Foundation Models with Lightning Attention Paper β’ 2501.08313 β’ Published Jan 14 β’ 298
rStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking Paper β’ 2501.04519 β’ Published Jan 8 β’ 285