Nemotron-Post-Training-v3 Collection Collection of datasets used in the post-training phase of Nemotron Nano v3. β’ 7 items β’ Updated 7 days ago β’ 52
Nemotron-Pre-Training-Datasets Collection Large scale pre-training datasets used in the Nemotron family of models. β’ 11 items β’ Updated 7 days ago β’ 84
NVIDIA Nemotron v3 Collection Open, Production-ready Enterprise Models β’ 6 items β’ Updated 7 days ago β’ 107
Bolmo: Byteifying the Next Generation of Language Models Paper β’ 2512.15586 β’ Published 13 days ago β’ 12
Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory Paper β’ 2504.19413 β’ Published Apr 28 β’ 36
SmolDocling: An ultra-compact vision-language model for end-to-end multi-modal document conversion Paper β’ 2503.11576 β’ Published Mar 14 β’ 122
TurboDiffusion: Accelerating Video Diffusion Models by 100-200 Times Paper β’ 2512.16093 β’ Published 13 days ago β’ 85
Emergent temporal abstractions in autoregressive models enable hierarchical reinforcement learning Paper β’ 2512.20605 β’ Published 7 days ago β’ 56
β Long-context post-training π§Ά β Collection Resources for post-training LLMs with long-context samples β’ 5 items β’ Updated Sep 14 β’ 6
VL-JEPA: Joint Embedding Predictive Architecture for Vision-language Paper β’ 2512.10942 β’ Published 19 days ago β’ 13
V-JEPA 2 Collection A frontier video understanding model developed by FAIR, Meta, which extends the pretraining objectives of https://ai.meta.com/blog/v-jepa-yann β’ 8 items β’ Updated Jun 13 β’ 177