From 571ef2b8050ae058d7491a07ee4284b279400895 Mon Sep 17 00:00:00 2001 From: Marut Pandya Date: Fri, 9 Aug 2024 14:58:46 -0700 Subject: [PATCH] Update README.md --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 7c567c2..a6eba8e 100644 --- a/README.md +++ b/README.md @@ -18,8 +18,10 @@ Deploy OpenAI-Compatible Blazing-Fast LLM Endpoints powered by the [vLLM](https: ### 1. UI for Deploying vLLM Worker on RunPod console: ![Demo of Deploying vLLM Worker on RunPod console with new UI](media/ui_demo.gif) -### 2. Worker vLLM `v1.2.0` with vLLM `0.5.4` now available under `stable` tags -Update v1.1 is now available, use the image tag `runpod/worker-v1-vllm:stable-cuda12.1.0`. +### 2. Worker vLLM `v1.2.0` with vLLM `0.5.4` now available under `stable` tags +**[Note]**: Current stable docker image version still runs vllm v0.5.3, It will be updated soon. + +Update v1.1.0 is now available, use the image tag `runpod/worker-v1-vllm:stable-cuda12.1.0`. ### 3. OpenAI-Compatible [Embedding Worker](https://github.com/runpod-workers/worker-infinity-embedding) Released Deploy your own OpenAI-compatible Serverless Endpoint on RunPod with multiple embedding models and fast inference for RAG and more!