Skip to content

Commit 87058b0

Browse files
authored
Fixing vLLM image (#651)
1 parent 3f8194b commit 87058b0

File tree

4 files changed

+3
-4
lines changed

4 files changed

+3
-4
lines changed

pyproject.toml

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[tool.poetry]
22
name = "truss"
3-
version = "0.7.2"
3+
version = "0.7.2rc2"
44
description = "A seamless bridge from model development to model delivery"
55
license = "MIT"
66
readme = "README.md"

truss/templates/vllm/proxy.conf.jinja

+1-1
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ server {
2121
location ~ ^/v1/models/model:predict$ {
2222
proxy_redirect off;
2323

24-
rewrite ^/v1/models/model:predict$ /generate break;
24+
rewrite ^/v1/models/model:predict$ {{server_endpoint}} break;
2525

2626
proxy_pass http://127.0.0.1:8081;
2727
}

truss/templates/vllm/supervisord.conf.jinja

+1-1
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ logfile=/dev/null
44
logfile_maxbytes=0
55

66
[program:vllm-server]
7-
command=python -m vllm.entrypoints.api_server --host=0.0.0.0 --port=8081 {{extra_args}}
7+
command=python -m vllm.entrypoints.openai.api_server --host=0.0.0.0 --port=8081 {{extra_args}}
88
startsecs=0
99
autostart=true
1010
autorestart=true

truss/templates/vllm/vllm.Dockerfile.jinja

-1
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,6 @@
44

55
FROM baseten/vllm:v0.3 as vllm
66

7-
FROM baseten/vllm:v0.1 as vllm
87
EXPOSE 8080-9000
98

109
RUN apt-get update && DEBIAN_FRONTEND=noninteractive apt-get install -y --no-install-recommends \

0 commit comments

Comments
 (0)