yangapku commited on
Commit
6130ef3
·
verified ·
1 Parent(s): a9c98e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -95,7 +95,7 @@ print("thinking content:", thinking_content)
95
  print("content:", content)
96
  ```
97
 
98
- For deployment, you can use `sglang>=0.4.6.post1` or `vllm>=0.8.4` or to create an OpenAI-compatible API endpoint:
99
  - SGLang:
100
  ```shell
101
  python -m sglang.launch_server --model-path Qwen/Qwen3-0.6B --reasoning-parser qwen3
@@ -105,7 +105,7 @@ For deployment, you can use `sglang>=0.4.6.post1` or `vllm>=0.8.4` or to create
105
  vllm serve Qwen/Qwen3-0.6B --enable-reasoning --reasoning-parser deepseek_r1
106
  ```
107
 
108
- For local use, applications such as llama.cpp, Ollama, LMStudio, and MLX-LM have also supported Qwen3.
109
 
110
  ## Switching Between Thinking and Non-Thinking Mode
111
 
 
95
  print("content:", content)
96
  ```
97
 
98
+ For deployment, you can use `sglang>=0.4.6.post1` or `vllm>=0.8.5` or to create an OpenAI-compatible API endpoint:
99
  - SGLang:
100
  ```shell
101
  python -m sglang.launch_server --model-path Qwen/Qwen3-0.6B --reasoning-parser qwen3
 
105
  vllm serve Qwen/Qwen3-0.6B --enable-reasoning --reasoning-parser deepseek_r1
106
  ```
107
 
108
+ For local use, applications such as Ollama, LMStudio, MLX-LM, llama.cpp, and KTransformers have also supported Qwen3.
109
 
110
  ## Switching Between Thinking and Non-Thinking Mode
111