Skip to content

Commit 2d3a9e8

Browse files
authored
Merge pull request #647 from basetenlabs/bump-version-0.7.2
Release 0.7.2
2 parents 05c6e39 + f5a2a98 commit 2d3a9e8

File tree

6 files changed

+37
-8
lines changed

6 files changed

+37
-8
lines changed

.github/workflows/commit_new_release_to_main.yml

+4-1
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,8 @@ jobs:
1313
steps:
1414
- name: Check out code
1515
uses: actions/checkout@v3
16+
with:
17+
token: ${{ secrets.BASETENBOT_GITHUB_TOKEN }}
1618

1719
- name: Configure Git user as basetenbot
1820
run: |
@@ -22,12 +24,13 @@ jobs:
2224
- name: Fetch all branches
2325
run: |
2426
git fetch --all --unshallow
27+
git checkout release
2528
git pull origin release
2629
2730
- name: Merge release into main with priority on main changes
2831
run: |
2932
git checkout main
30-
git merge --strategy-option=ours release -m "Merge release into main prioritizing main changes"
33+
git merge --strategy-option=ours release
3134
git push origin main
3235
env:
3336
GH_TOKEN: ${{ secrets.BASETENBOT_GITHUB_TOKEN }}

docker/base_images/vllm.Dockerfile

+26
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,26 @@
1+
# Use an official CUDA runtime as a parent image
2+
FROM nvidia/cuda:11.8.0-devel-ubuntu22.04
3+
4+
ENV DEBIAN_FRONTEND=noninteractive
5+
ENV SHELL=/bin/bash
6+
7+
# Update, upgrade, install packages and clean up
8+
RUN apt-get update --yes && \
9+
apt-get upgrade --yes && \
10+
apt install --yes --no-install-recommends git wget curl bash software-properties-common nginx && \
11+
apt install python3.10-dev python3.10-venv -y --no-install-recommends && \
12+
apt-get autoremove -y && \
13+
apt-get clean && \
14+
rm -rf /var/lib/apt/lists/* && \
15+
echo "en_US.UTF-8 UTF-8" > /etc/locale.gen
16+
17+
18+
# Set up Python and pip
19+
RUN ln -s /usr/bin/python3.10 /usr/bin/python && \
20+
rm /usr/bin/python3 && \
21+
ln -s /usr/bin/python3.10 /usr/bin/python3 && \
22+
curl https://bootstrap.pypa.io/get-pip.py -o get-pip.py && \
23+
python get-pip.py
24+
25+
# Install vllm and pandas (dependency of Ray)
26+
RUN pip install --upgrade --no-cache-dir vllm pandas>=1.3 && rm -rf /root/.cache/pip

pyproject.toml

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[tool.poetry]
22
name = "truss"
3-
version = "0.7.1"
3+
version = "0.7.2"
44
description = "A seamless bridge from model development to model delivery"
55
license = "MIT"
66
readme = "README.md"

truss/templates/server/model_wrapper.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -366,7 +366,7 @@ def _handle_exception():
366366
# Note that logger.exception logs the stacktrace, such that the user can
367367
# debug this error from the logs.
368368
logging.exception("Internal Server Error")
369-
raise HTTPException(status_code=500, detail={"message": "Internal Server Error"})
369+
raise HTTPException(status_code=500, detail="Internal Server Error")
370370

371371

372372
def _intercept_exceptions_sync(func):

truss/templates/vllm/vllm.Dockerfile.jinja

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
FROM baseten/vllm:v0.1
1+
FROM baseten/vllm:v0.3
22

33
EXPOSE 8080-9000
44

truss/tests/test_model_inference.py

+4-4
Original file line numberDiff line numberDiff line change
@@ -424,7 +424,7 @@ def predict(self, request):
424424

425425
assert_logs_contain_error(container.logs(), "ValueError: error")
426426

427-
assert "Internal Server Error" in response.json()["error"]["message"]
427+
assert "Internal Server Error" in response.json()["error"]
428428

429429
model_preprocess_error = """
430430
class Model:
@@ -452,7 +452,7 @@ def predict(self, request):
452452
assert "error" in response.json()
453453

454454
assert_logs_contain_error(container.logs(), "ValueError: error")
455-
assert "Internal Server Error" in response.json()["error"]["message"]
455+
assert "Internal Server Error" in response.json()["error"]
456456

457457
model_postprocess_error = """
458458
class Model:
@@ -479,7 +479,7 @@ def postprocess(self, response):
479479
assert response.status_code == 500
480480
assert "error" in response.json()
481481
assert_logs_contain_error(container.logs(), "ValueError: error")
482-
assert "Internal Server Error" in response.json()["error"]["message"]
482+
assert "Internal Server Error" in response.json()["error"]
483483

484484
model_async = """
485485
class Model:
@@ -505,7 +505,7 @@ async def predict(self, request):
505505

506506
assert_logs_contain_error(container.logs(), "ValueError: error")
507507

508-
assert "Internal Server Error" in response.json()["error"]["message"]
508+
assert "Internal Server Error" in response.json()["error"]
509509

510510

511511
@pytest.mark.integration

0 commit comments

Comments
 (0)