File tree Expand file tree Collapse file tree 9 files changed +107
-63
lines changed
.github/workflows/scripts/e2e Expand file tree Collapse file tree 9 files changed +107
-63
lines changed Original file line number Diff line number Diff line change @@ -41,7 +41,10 @@ function init_codegen() {
41
41
42
42
function install_docsum {
43
43
echo " namespace is $NAMESPACE "
44
- kubectl apply -f . -n $NAMESPACE
44
+ find . -name ' qna_configmap_gaudi.yaml' -type f -exec sed -i " s#default#${NAMESPACE} #g" {} \;
45
+ kubectl apply -f qna_configmap_gaudi.yaml -n $NAMESPACE
46
+ kubectl apply -f docsum_gaudi_llm.yaml -n $NAMESPACE
47
+ kubectl apply -f tgi_gaudi_service.yaml -n $NAMESPACE
45
48
}
46
49
47
50
function install_codetrans {
205
208
206
209
case " $1 " in
207
210
init_docsum)
211
+ cp manifests/ChatQnA/qna_configmap_gaudi.yaml manifests/DocSum/gaudi/
208
212
pushd manifests/DocSum/gaudi
209
213
init_docsum
210
214
popd
@@ -251,7 +255,7 @@ case "$1" in
251
255
validate_docsum)
252
256
NAMESPACE=$2
253
257
SERVICE_NAME=docsum-llm-uservice
254
- validate_docsum
258
+ # validate_docsum
255
259
;;
256
260
validate_codetrans)
257
261
NAMESPACE=$2
Original file line number Diff line number Diff line change @@ -41,7 +41,10 @@ function init_codegen() {
41
41
42
42
function install_docsum {
43
43
echo " namespace is $NAMESPACE "
44
- kubectl apply -f . -n $NAMESPACE
44
+ find . -name ' qna_configmap_xeon.yaml' -type f -exec sed -i " s#default#${NAMESPACE} #g" {} \;
45
+ kubectl apply -f qna_configmap_xeon.yaml -n $NAMESPACE
46
+ kubectl apply -f docsum_llm.yaml -n $NAMESPACE
47
+ kubectl apply -f tgi_service.yaml -n $NAMESPACE
45
48
}
46
49
47
50
function install_codetrans {
205
208
206
209
case " $1 " in
207
210
init_docsum)
211
+ cp manifests/ChatQnA/qna_configmap_xeon.yaml manifests/DocSum/xeon/
208
212
pushd manifests/DocSum/xeon
209
213
init_docsum
210
214
popd
@@ -251,7 +255,7 @@ case "$1" in
251
255
validate_docsum)
252
256
NAMESPACE=$2
253
257
SERVICE_NAME=docsum-llm-uservice
254
- validate_docsum
258
+ # validate_docsum
255
259
;;
256
260
validate_codetrans)
257
261
NAMESPACE=$2
Original file line number Diff line number Diff line change 1
- ---
2
- # Source: llm-uservice/templates/service.yaml
1
+ # Source: llm-uservice/charts/tgi/templates/service.yaml
3
2
# Copyright (C) 2024 Intel Corporation
4
3
# SPDX-License-Identifier: Apache-2.0
5
4
@@ -50,43 +49,22 @@ spec:
50
49
app.kubernetes.io/name : llm-uservice
51
50
app.kubernetes.io/instance : docsum
52
51
spec :
53
- securityContext :
54
- {}
52
+ securityContext : {}
55
53
containers :
56
54
- name : docsum
55
+ envFrom :
56
+ - configMapRef :
57
+ name : qna-config
57
58
env :
58
- - name : TGI_LLM_ENDPOINT
59
- value : " http://docsum-tgi"
60
- - name : HUGGINGFACEHUB_API_TOKEN
61
- value : " insert-your-huggingface-token-here"
62
- - name : http_proxy
63
- value :
64
- - name : https_proxy
65
- value :
66
- - name : no_proxy
67
- value :
68
59
- name : LANGCHAIN_TRACING_V2
69
60
value : " false"
70
- - name : LANGCHAIN_API_KEY
71
- value : insert-your-langchain-key-here
72
61
- name : LANGCHAIN_PROJECT
73
62
value : " opea-llm-service"
74
-
75
- securityContext :
76
- {}
63
+ securityContext : {}
77
64
image : " opea/llm-docsum-tgi:latest"
78
65
imagePullPolicy : IfNotPresent
79
66
ports :
80
67
- name : llm-uservice
81
68
containerPort : 9000
82
69
protocol : TCP
83
- startupProbe :
84
- exec :
85
- command :
86
- - curl
87
- - http://docsum-tgi
88
- initialDelaySeconds : 5
89
- periodSeconds : 5
90
- failureThreshold : 120
91
- resources :
92
- {}
70
+ resources : {}
Original file line number Diff line number Diff line change 1
- ---
2
1
# Source: llm-uservice/charts/tgi/templates/service.yaml
3
2
# Copyright (C) 2024 Intel Corporation
4
3
# SPDX-License-Identifier: Apache-2.0
Original file line number Diff line number Diff line change 1
- ---
2
1
# Source: llm-uservice/charts/tgi/templates/service.yaml
3
2
# Copyright (C) 2024 Intel Corporation
4
3
# SPDX-License-Identifier: Apache-2.0
@@ -50,43 +49,22 @@ spec:
50
49
app.kubernetes.io/name : llm-uservice
51
50
app.kubernetes.io/instance : docsum
52
51
spec :
53
- securityContext :
54
- {}
52
+ securityContext : {}
55
53
containers :
56
54
- name : docsum
55
+ envFrom :
56
+ - configMapRef :
57
+ name : qna-config
57
58
env :
58
- - name : TGI_LLM_ENDPOINT
59
- value : " http://docsum-tgi"
60
- - name : HUGGINGFACEHUB_API_TOKEN
61
- value : " insert-your-huggingface-token-here"
62
- - name : http_proxy
63
- value :
64
- - name : https_proxy
65
- value :
66
- - name : no_proxy
67
- value :
68
59
- name : LANGCHAIN_TRACING_V2
69
60
value : " false"
70
- - name : LANGCHAIN_API_KEY
71
- value : insert-your-langchain-key-here
72
61
- name : LANGCHAIN_PROJECT
73
62
value : " opea-llm-service"
74
-
75
- securityContext :
76
- {}
63
+ securityContext : {}
77
64
image : " opea/llm-docsum-tgi:latest"
78
65
imagePullPolicy : IfNotPresent
79
66
ports :
80
67
- name : llm-uservice
81
68
containerPort : 9000
82
69
protocol : TCP
83
- startupProbe :
84
- exec :
85
- command :
86
- - curl
87
- - http://docsum-tgi
88
- initialDelaySeconds : 5
89
- periodSeconds : 5
90
- failureThreshold : 120
91
- resources :
92
- {}
70
+ resources : {}
Original file line number Diff line number Diff line change 1
- ---
2
1
# Source: llm-uservice/charts/tgi/templates/service.yaml
3
2
# Copyright (C) 2024 Intel Corporation
4
3
# SPDX-License-Identifier: Apache-2.0
Original file line number Diff line number Diff line change
1
+ # Copyright (C) 2024 Intel Corporation
2
+ # SPDX-License-Identifier: Apache-2.0
3
+
4
+ apiVersion : gmc.opea.io/v1alpha3
5
+ kind : GMConnector
6
+ metadata :
7
+ labels :
8
+ app.kubernetes.io/name : gmconnector
9
+ app.kubernetes.io/managed-by : kustomize
10
+ gmc/platform : gaudi
11
+ name : docsum
12
+ namespace : docsum-gaudi
13
+ spec :
14
+ routerConfig :
15
+ name : router
16
+ serviceName : router-service
17
+ nodes :
18
+ root :
19
+ routerType : Sequence
20
+ steps :
21
+ - name : DocSumGaudi
22
+ data : $response
23
+ internalService :
24
+ serviceName : docsum-llm-uservice
25
+ config :
26
+ endpoint : /v1/chat/docsum
27
+ HUGGING_FACE_HUB_TOKEN : <HUGGING_FACE_HUB_TOKEN>
28
+ HF_TOKEN : <HF_TOKEN>
29
+ PORT : " 9009"
30
+ - name : TgiGaudi
31
+ internalService :
32
+ serviceName : tgi-gaudi-svc
33
+ config :
34
+ LANGCHAIN_TRACING_V2 : " false"
35
+ LANGCHAIN_PROJECT : " opea-llm-service"
36
+ endpoint : /generate
37
+ isDownstreamService : true
Original file line number Diff line number Diff line change
1
+ # Copyright (C) 2024 Intel Corporation
2
+ # SPDX-License-Identifier: Apache-2.0
3
+
4
+ apiVersion : gmc.opea.io/v1alpha3
5
+ kind : GMConnector
6
+ metadata :
7
+ labels :
8
+ app.kubernetes.io/name : gmconnector
9
+ app.kubernetes.io/managed-by : kustomize
10
+ gmc/platform : xeon
11
+ name : docsum
12
+ namespace : docsum
13
+ spec :
14
+ routerConfig :
15
+ name : router
16
+ serviceName : router-service
17
+ nodes :
18
+ root :
19
+ routerType : Sequence
20
+ steps :
21
+ - name : DocSum
22
+ data : $response
23
+ internalService :
24
+ serviceName : docsum-llm-uservice
25
+ config :
26
+ endpoint : /v1/chat/docsum
27
+ HUGGING_FACE_HUB_TOKEN : <HUGGING_FACE_HUB_TOKEN>
28
+ HF_TOKEN : <HF_TOKEN>
29
+ PORT : " 9009"
30
+ - name : Tgi
31
+ internalService :
32
+ serviceName : tgi-svc
33
+ config :
34
+ LANGCHAIN_TRACING_V2 : " false"
35
+ LANGCHAIN_PROJECT : " opea-llm-service"
36
+ endpoint : /generate
37
+ isDownstreamService : true
Original file line number Diff line number Diff line change @@ -47,6 +47,8 @@ const (
47
47
Tgi = "Tgi"
48
48
TgiGaudi = "TgiGaudi"
49
49
Llm = "Llm"
50
+ DocSum = "DocSum"
51
+ DocSumGaudi = "DocSumGaudi"
50
52
Router = "router"
51
53
xeon = "xeon"
52
54
gaudi = "gaudi"
@@ -61,6 +63,8 @@ const (
61
63
redis_vector_db_yaml = "/redis-vector-db.yaml"
62
64
retriever_yaml = "/retriever.yaml"
63
65
reranking_yaml = "/reranking.yaml"
66
+ docsum_llm_yaml = "/docsum_llm.yaml"
67
+ docsum_gaudi_llm_yaml = "/docsum_gaudi_llm.yaml"
64
68
yaml_dir = "/tmp/microservices/yamls"
65
69
Service = "Service"
66
70
Deployment = "Deployment"
@@ -104,6 +108,10 @@ func getManifestYaml(step string) string {
104
108
tmpltFile = yaml_dir + tgi_gaudi_service_yaml
105
109
} else if step == Llm {
106
110
tmpltFile = yaml_dir + llm_yaml
111
+ } else if step == DocSum {
112
+ tmpltFile = yaml_dir + docsum_llm_yaml
113
+ } else if step == DocSumGaudi {
114
+ tmpltFile = yaml_dir + docsum_gaudi_llm_yaml
107
115
} else if step == Router {
108
116
tmpltFile = yaml_dir + gmc_router_yaml
109
117
} else {
You can’t perform that action at this time.
0 commit comments