Text Generation
Transformers
Safetensors
PyTorch
English
nvidia
conversational
suhara commited on
Commit
0bba2a3
·
verified ·
1 Parent(s): f5b038a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -34,6 +34,10 @@ NVIDIA Nemotron-H-4B-Instruct-128K is a large language model (LLM) developed by
34
 
35
  The model underwent a multi-phase post-training process including multiple supervised fine-tuning stages for math, code, science, and then chat, instruction following, and tool-calling, followed by multiple preference tuning stages using Reward-aware Preference Optimization (RPO) for both chat and instruction-following.
36
 
 
 
 
 
37
  This model is for research and development only.
38
 
39
  ## License/Terms of Use
 
34
 
35
  The model underwent a multi-phase post-training process including multiple supervised fine-tuning stages for math, code, science, and then chat, instruction following, and tool-calling, followed by multiple preference tuning stages using Reward-aware Preference Optimization (RPO) for both chat and instruction-following.
36
 
37
+ The model was pruned and distilled from [Nemotron-H-Base-8K](https://huggingface.co/nvidia/Nemotron-H-8B-Base-8K) using our hybrid language model compression technique. For more details, please refer to the [paper](https://arxiv.org/abs/2504.11409).
38
+
39
+ The paper has been accepted for publication at NeurIPS 2025.
40
+
41
  This model is for research and development only.
42
 
43
  ## License/Terms of Use