Fathom-4B-AIO-GGUF
Fathom-DeepResearch is a dual-model agentic system consisting of Fathom-Search-4B and Fathom-Synthesizer-4B, optimized for long-horizon web search, evidence verification, and synthesis into citation-dense reports. These 4B-parameter models are trained to browse, extract, verify, and reason over live web content, setting state-of-the-art open-weight performance on search-heavy tasks (SimpleQA, FRAMES, WebWalkerQA, Seal0) and surpassing closed-source DeepResearch agents like Claude, Grok, Perplexity, and GPT-4o on DeepResearch-Bench. The system leverages a multi-agent self-play pipeline to generate datasets like DuetQA, enforces live web search with multi-hop QA pairs, and employs RAPO (Reward-Aware Policy Optimization) and novel steerable, step-level rewards for stable multi-turn RL with verifiable rewards. The DeepResearch plan-then-write protocol ensures rigorous decomposition, evidence mapping, and insight generation for end-to-end report synthesis. The released stack includes a web-agent search tool server built on Jina-AI, Crawl4AI, Trafilatura, and Serper.dev, supporting high-volume, asynchronous search with URL handlers for YouTube, PDFs, Reddit, and more. The models and code are licensed under MIT to promote open research and community collaboration, with a focus on democratizing complex web-enhanced LLMs for research, development, and AI practitioners worldwide.
Fathom-4B Models GGUF
| Model | Link | 
|---|---|
| Fathom-Search-4B-GGUF | https://huggingface.co/prithivMLmods/Fathom-4B-AIO-GGUF/tree/main/Fathom-Search-4B-GGUF | 
| Fathom-Synthesizer-4B-GGUF | https://huggingface.co/prithivMLmods/Fathom-4B-AIO-GGUF/tree/main/Fathom-Synthesizer-4B-GGUF | 
Model Files
Fathom-Search-4B
| File Name | Quant Type | File Size | 
|---|---|---|
| Fathom-Search-4B.BF16.gguf | BF16 | 8.05 GB | 
| Fathom-Search-4B.F16.gguf | F16 | 8.05 GB | 
| Fathom-Search-4B.F32.gguf | F32 | 16.1 GB | 
| Fathom-Search-4B.Q2_K.gguf | Q2_K | 1.67 GB | 
| Fathom-Search-4B.Q3_K_L.gguf | Q3_K_L | 2.24 GB | 
| Fathom-Search-4B.Q3_K_M.gguf | Q3_K_M | 2.08 GB | 
| Fathom-Search-4B.Q3_K_S.gguf | Q3_K_S | 1.89 GB | 
| Fathom-Search-4B.Q4_K_M.gguf | Q4_K_M | 2.5 GB | 
| Fathom-Search-4B.Q4_K_S.gguf | Q4_K_S | 2.38 GB | 
| Fathom-Search-4B.Q5_K_M.gguf | Q5_K_M | 2.89 GB | 
| Fathom-Search-4B.Q5_K_S.gguf | Q5_K_S | 2.82 GB | 
| Fathom-Search-4B.Q6_K.gguf | Q6_K | 3.31 GB | 
| Fathom-Search-4B.Q8_0.gguf | Q8_0 | 4.28 GB | 
Fathom-Synthesizer-4B
| File Name | Quant Type | File Size | 
|---|---|---|
| Fathom-Synthesizer-4B.BF16.gguf | BF16 | 8.05 GB | 
| Fathom-Synthesizer-4B.F16.gguf | F16 | 8.05 GB | 
| Fathom-Synthesizer-4B.F32.gguf | F32 | 16.1 GB | 
| Fathom-Synthesizer-4B.Q2_K.gguf | Q2_K | 1.67 GB | 
| Fathom-Synthesizer-4B.Q3_K_L.gguf | Q3_K_L | 2.24 GB | 
| Fathom-Synthesizer-4B.Q3_K_M.gguf | Q3_K_M | 2.08 GB | 
| Fathom-Synthesizer-4B.Q3_K_S.gguf | Q3_K_S | 1.89 GB | 
| Fathom-Synthesizer-4B.Q4_K_M.gguf | Q4_K_M | 2.5 GB | 
| Fathom-Synthesizer-4B.Q4_K_S.gguf | Q4_K_S | 2.38 GB | 
| Fathom-Synthesizer-4B.Q5_K_M.gguf | Q5_K_M | 2.89 GB | 
| Fathom-Synthesizer-4B.Q5_K_S.gguf | Q5_K_S | 2.82 GB | 
| Fathom-Synthesizer-4B.Q6_K.gguf | Q6_K | 3.31 GB | 
| Fathom-Synthesizer-4B.Q8_0.gguf | Q8_0 | 4.28 GB | 
Quants Usage
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better):
- Downloads last month
- 673
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
32-bit

