Update vLLM command
Browse files
    	
        README.md
    CHANGED
    
    | @@ -94,7 +94,12 @@ transformers chat localhost:8000 --model-name-or-path openai/gpt-oss-120b | |
| 94 | 
             
            vLLM recommends using [uv](https://docs.astral.sh/uv/) for Python dependency management. You can use vLLM to spin up an OpenAI-compatible webserver. The following command will automatically download the model and start the server.
         | 
| 95 |  | 
| 96 | 
             
            ```bash
         | 
| 97 | 
            -
            uv  | 
|  | |
|  | |
|  | |
|  | |
|  | |
| 98 | 
             
            ```
         | 
| 99 |  | 
| 100 | 
             
            [Learn more about how to use gpt-oss with vLLM.](https://cookbook.openai.com/articles/gpt-oss/run-vllm)
         | 
|  | |
| 94 | 
             
            vLLM recommends using [uv](https://docs.astral.sh/uv/) for Python dependency management. You can use vLLM to spin up an OpenAI-compatible webserver. The following command will automatically download the model and start the server.
         | 
| 95 |  | 
| 96 | 
             
            ```bash
         | 
| 97 | 
            +
            uv pip install --pre vllm==0.10.1+gptoss \
         | 
| 98 | 
            +
                --extra-index-url https://wheels.vllm.ai/gpt-oss/ \
         | 
| 99 | 
            +
                --extra-index-url https://download.pytorch.org/whl/nightly/cu128 \
         | 
| 100 | 
            +
                --index-strategy unsafe-best-match
         | 
| 101 | 
            +
             | 
| 102 | 
            +
            vllm serve openai/gpt-oss-120b
         | 
| 103 | 
             
            ```
         | 
| 104 |  | 
| 105 | 
             
            [Learn more about how to use gpt-oss with vLLM.](https://cookbook.openai.com/articles/gpt-oss/run-vllm)
         | 

