diff --git a/README.md b/README.md index 26671653976e552d81346c1c8f7b301eb40306a6..4181bb01de9c8c140932e2f4f77ed906b9a55907 100644 --- a/README.md +++ b/README.md @@ -172,6 +172,9 @@ Qwen2.5 is supported by multiple inference frameworks. Here we demonstrate the u ### vLLM +> [!Warning] +> Not working with vllm 0.6.3. + We advise you to use the latest version of vLLM to build OpenAI-compatible API service, including tool use support. Start the server with a chat model, e.g. `Qwen2.5-7B-Instruct`: ```shell vllm serve Qwen/Qwen2.5-7B-Instruct