- 
	
	
	
TokDrift: When LLM Speaks in Subwords but Code Speaks in Grammar
Paper • 2510.14972 • Published • 29 - 
	
	
	
LightMem: Lightweight and Efficient Memory-Augmented Generation
Paper • 2510.18866 • Published • 107 - 
	
	
	
Every Attention Matters: An Efficient Hybrid Architecture for Long-Context Reasoning
Paper • 2510.19338 • Published • 101 - 
	
	
	1.22k
The Smol Training Playbook: The Secrets to Building World-Class LLMs
📝 
Jonatan Borkowski
j14i
		AI & ML interests
None yet
		Recent Activity
						upvoted 
								a
								paper
							
						about 6 hours ago
						
					
						
						
						To Code, or Not To Code? Exploring Impact of Code in Pre-training
						
						liked
								a model
							
						1 day ago
						
					
						
						
						
						Qwen/Qwen3-Coder-480B-A35B-Instruct