Update README.md
Browse files
    	
        README.md
    CHANGED
    
    | @@ -34,6 +34,10 @@ NVIDIA Nemotron-H-4B-Instruct-128K is a large language model (LLM) developed by | |
| 34 |  | 
| 35 | 
             
            The model underwent a multi-phase post-training process including multiple supervised fine-tuning stages for math, code, science, and then chat, instruction following, and tool-calling, followed by multiple preference tuning stages using Reward-aware Preference Optimization (RPO) for both chat and instruction-following.
         | 
| 36 |  | 
|  | |
|  | |
|  | |
|  | |
| 37 | 
             
            This model is for research and development only.
         | 
| 38 |  | 
| 39 | 
             
            ## License/Terms of Use
         | 
|  | |
| 34 |  | 
| 35 | 
             
            The model underwent a multi-phase post-training process including multiple supervised fine-tuning stages for math, code, science, and then chat, instruction following, and tool-calling, followed by multiple preference tuning stages using Reward-aware Preference Optimization (RPO) for both chat and instruction-following.
         | 
| 36 |  | 
| 37 | 
            +
            The model was pruned and distilled from [Nemotron-H-Base-8K](https://huggingface.co/nvidia/Nemotron-H-8B-Base-8K) using our hybrid language model compression technique. For more details, please refer to the [paper](https://arxiv.org/abs/2504.11409).
         | 
| 38 | 
            +
             | 
| 39 | 
            +
            The paper has been accepted for publication at NeurIPS 2025.
         | 
| 40 | 
            +
             | 
| 41 | 
             
            This model is for research and development only.
         | 
| 42 |  | 
| 43 | 
             
            ## License/Terms of Use
         | 
