Skip to content

Commit 917ba59

Browse files
Merge pull request #289 from nayihz/update_vllm_version
cleanup: upgrade vllm version to v0.7.3
2 parents 9eba500 + cb8c2ba commit 917ba59

File tree

4 files changed

+5
-3
lines changed

4 files changed

+5
-3
lines changed

chart/templates/backends/vllm.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ spec:
1313
- -m
1414
- vllm.entrypoints.openai.api_server
1515
image: vllm/vllm-openai
16-
version: v0.6.0
16+
version: v0.7.3
1717
# Do not edit the preset argument name unless you know what you're doing.
1818
# Free to add more arguments with your requirements.
1919
recommendedConfigs:

test/integration/webhook/backendruntime_test.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ var _ = ginkgo.Describe("BackendRuntime default and validation", func() {
8888
ginkgo.Entry("BackendRuntime creation with no resources", &testValidatingCase{
8989
creationFunc: func() *inferenceapi.BackendRuntime {
9090
return wrapper.MakeBackendRuntime("vllm").
91-
Image("vllm/vllm-openai").Version("v0.6.0").
91+
Image("vllm/vllm-openai").Version(util.VllmImageVersion).
9292
Command([]string{"python3", "-m", "vllm.entrypoints.openai.api_server"}).
9393
Arg("default", []string{"--model", "{{.ModelPath}}", "--served-model-name", "{{.ModelName}}", "--host", "0.0.0.0", "--port", "8080"}).
9494
Obj()

test/util/consts.go

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,4 +23,6 @@ const (
2323
Interval = time.Millisecond * 250
2424
E2ETimeout = 5 * time.Minute
2525
E2EInterval = 1 * time.Second
26+
27+
VllmImageVersion = "v0.7.3"
2628
)

test/util/mock.go

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ func MockASampleService(ns string) *inferenceapi.Service {
5454

5555
func MockASampleBackendRuntime() *wrapper.BackendRuntimeWrapper {
5656
return wrapper.MakeBackendRuntime("vllm").
57-
Image("vllm/vllm-openai").Version("v0.6.0").
57+
Image("vllm/vllm-openai").Version(VllmImageVersion).
5858
Command([]string{"python3", "-m", "vllm.entrypoints.openai.api_server"}).
5959
Arg("default", []string{"--model", "{{.ModelPath}}", "--served-model-name", "{{.ModelName}}", "--host", "0.0.0.0", "--port", "8080"}).
6060
Request("default", "cpu", "4").Limit("default", "cpu", "4")

0 commit comments

Comments
 (0)