Skip to content

Commit 8aaf43d

Browse files
committed
cleanup: upgrade vllm version to v0.7.3
1 parent 9eba500 commit 8aaf43d

File tree

3 files changed

+5
-4
lines changed

3 files changed

+5
-4
lines changed

chart/templates/backends/vllm.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ spec:
1313
- -m
1414
- vllm.entrypoints.openai.api_server
1515
image: vllm/vllm-openai
16-
version: v0.6.0
16+
version: v0.7.3
1717
# Do not edit the preset argument name unless you know what you're doing.
1818
# Free to add more arguments with your requirements.
1919
recommendedConfigs:

test/integration/webhook/backendruntime_test.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ var _ = ginkgo.Describe("BackendRuntime default and validation", func() {
8888
ginkgo.Entry("BackendRuntime creation with no resources", &testValidatingCase{
8989
creationFunc: func() *inferenceapi.BackendRuntime {
9090
return wrapper.MakeBackendRuntime("vllm").
91-
Image("vllm/vllm-openai").Version("v0.6.0").
91+
Image("vllm/vllm-openai").Version(util.VllmImageVersion).
9292
Command([]string{"python3", "-m", "vllm.entrypoints.openai.api_server"}).
9393
Arg("default", []string{"--model", "{{.ModelPath}}", "--served-model-name", "{{.ModelName}}", "--host", "0.0.0.0", "--port", "8080"}).
9494
Obj()

test/util/mock.go

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,8 @@ import (
2828
)
2929

3030
const (
31-
sampleModelName = "llama3-8b"
31+
sampleModelName = "llama3-8b"
32+
VllmImageVersion = "v0.7.3"
3233
)
3334

3435
func MockASampleModel() *coreapi.OpenModel {
@@ -54,7 +55,7 @@ func MockASampleService(ns string) *inferenceapi.Service {
5455

5556
func MockASampleBackendRuntime() *wrapper.BackendRuntimeWrapper {
5657
return wrapper.MakeBackendRuntime("vllm").
57-
Image("vllm/vllm-openai").Version("v0.6.0").
58+
Image("vllm/vllm-openai").Version(VllmImageVersion).
5859
Command([]string{"python3", "-m", "vllm.entrypoints.openai.api_server"}).
5960
Arg("default", []string{"--model", "{{.ModelPath}}", "--served-model-name", "{{.ModelName}}", "--host", "0.0.0.0", "--port", "8080"}).
6061
Request("default", "cpu", "4").Limit("default", "cpu", "4")

0 commit comments

Comments
 (0)