Skip to content

Commit d994618

Browse files
authored
doc: fix minor issue in GMC doc (#383)
Signed-off-by: Ruoyu Ying <[email protected]>
1 parent 11a56e0 commit d994618

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

ChatQnA/kubernetes/manifests/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ kubectl create deployment client-test -n chatqa --image=python:3.8.13 -- sleep i
6464
5. Access the application using the above URL from the client pod
6565

6666
```sh
67-
export CLIENT_POD=$(kubectl get pod -l app=client-test -o jsonpath={.items..metadata.name})
67+
export CLIENT_POD=$(kubectl get pod -n chatqa -l app=client-test -o jsonpath={.items..metadata.name})
6868
export accessUrl=$(kubectl get gmc -n chatqa -o jsonpath="{.items[?(@.metadata.name=='chatqa')].status.accessUrl}")
6969
kubectl exec "$CLIENT_POD" -n chatqa -- curl $accessUrl -X POST -d '{"text":"What is the revenue of Nike in 2023?","parameters":{"max_new_tokens":17, "do_sample": true}}' -H 'Content-Type: application/json'
7070
```
@@ -79,7 +79,7 @@ For example, to use Llama-2-7b-chat-hf make the following edit:
7979
```yaml
8080
- name: Tgi
8181
internalService:
82-
serviceName: tgi-svc
82+
serviceName: tgi-service-m
8383
config:
8484
LLM_MODEL_ID: Llama-2-7b-chat-hf
8585
```
@@ -92,7 +92,7 @@ kubectl apply -f $(pwd)/chatQnA_xeon.yaml
9292
8. Check that the tgi-svc-deployment has been changed to use the new LLM Model
9393

9494
```sh
95-
kubectl get deployment tgi-svc-deployment -n chatqa -o jsonpath="{.spec.template.spec.containers[*].env[?(@.name=='LLM_MODEL_ID')].value}"
95+
kubectl get deployment tgi-service-m-deployment -n chatqa -o jsonpath="{.spec.template.spec.containers[*].env[?(@.name=='LLM_MODEL_ID')].value}"
9696
```
9797

9898
9. Access the updated pipeline using the same URL from above using the client pod

DocSum/kubernetes/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ kubectl create deployment client-test -n ${ns} --image=python:3.8.13 -- sleep in
5858
6. Access the pipeline using the above URL from the client pod and execute a request
5959

6060
```bash
61-
export CLIENT_POD=$(kubectl get pod -l app=client-test -o jsonpath={.items..metadata.name})
61+
export CLIENT_POD=$(kubectl get pod -n ${ns} -l app=client-test -o jsonpath={.items..metadata.name})
6262
export accessUrl=$(kubectl get gmc -n $ns -o jsonpath="{.items[?(@.metadata.name=='docsum')].status.accessUrl}")
6363
kubectl exec "$CLIENT_POD" -n $ns -- curl $accessUrl -X POST -d '{"query":"Text Embeddings Inference (TEI) is a toolkit for deploying and serving open source text embeddings and sequence classification models. TEI enables high-performance extraction for the most popular models, including FlagEmbedding, Ember, GTE and E5."}' -H 'Content-Type: application/json'
6464
```

0 commit comments

Comments
 (0)