Update README.md
Browse files
README.md
CHANGED
|
@@ -18,6 +18,8 @@ Compared with the state-of-the-art opensource language models, including the pre
|
|
| 18 |
Qwen2-7B-Instruct-GPTQ-Int4 supports a context length of up to 131,072 tokens, enabling the processing of extensive inputs. Please refer to [this section](#processing-long-texts) for detailed instructions on how to deploy Qwen2 for handling long texts.
|
| 19 |
|
| 20 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2/) and [GitHub](https://github.com/QwenLM/Qwen2).
|
|
|
|
|
|
|
| 21 |
<br>
|
| 22 |
|
| 23 |
## Model Details
|
|
|
|
| 18 |
Qwen2-7B-Instruct-GPTQ-Int4 supports a context length of up to 131,072 tokens, enabling the processing of extensive inputs. Please refer to [this section](#processing-long-texts) for detailed instructions on how to deploy Qwen2 for handling long texts.
|
| 19 |
|
| 20 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwen2/) and [GitHub](https://github.com/QwenLM/Qwen2).
|
| 21 |
+
|
| 22 |
+
**Note**: If you encounter ``RuntimeError: probability tensor contains either `inf`, `nan` or element < 0`` during inference with ``transformer``, we recommand installing ``autogpq>=0.7.1`` or [deploying this model with vLLM](https://qwen.readthedocs.io/en/latest/deployment/vllm.html).
|
| 23 |
<br>
|
| 24 |
|
| 25 |
## Model Details
|