Skip to content

Commit 3505bd2

Browse files
chensuyueSpycsh
andauthored
update tgi-gaudi version (#389)
Signed-off-by: chensuyue <suyue.chen@intel.com> Co-authored-by: Spycsh <sihan.chen@intel.com>
1 parent 4978959 commit 3505bd2

File tree

22 files changed

+63
-41
lines changed

22 files changed

+63
-41
lines changed

AudioQnA/docker/gaudi/docker_compose.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ services:
5454
environment:
5555
TTS_ENDPOINT: ${TTS_ENDPOINT}
5656
tgi-service:
57-
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
57+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
5858
container_name: tgi-gaudi-server
5959
ports:
6060
- "3006:80"
@@ -73,7 +73,7 @@ services:
7373
cap_add:
7474
- SYS_NICE
7575
ipc: host
76-
command: --model-id ${LLM_MODEL_ID}
76+
command: --model-id ${LLM_MODEL_ID} --max-input-length 1024 --max-total-tokens 2048
7777
llm:
7878
image: opea/llm-tgi:latest
7979
container_name: llm-tgi-gaudi-server

AudioQnA/tests/test_audioqna_on_gaudi.sh

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ function build_docker_images() {
2020
docker build -t opea/speecht5:latest -f comps/tts/speecht5/Dockerfile_hpu .
2121
docker build -t opea/tts:latest -f comps/tts/Dockerfile .
2222

23-
docker pull ghcr.io/huggingface/tgi-gaudi:1.2.1
23+
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1
2424

2525
cd ..
2626

@@ -76,9 +76,17 @@ function start_services() {
7676

7777
function validate_megaservice() {
7878
result=$(http_proxy="" curl http://${ip_address}:3008/v1/audioqna -XPOST -d '{"audio": "UklGRigAAABXQVZFZm10IBIAAAABAAEARKwAAIhYAQACABAAAABkYXRhAgAAAAEA", "max_tokens":64}' -H 'Content-Type: application/json')
79+
echo "result is === $result"
7980
if [[ $result == *"AAA"* ]]; then
8081
echo "Result correct."
8182
else
83+
docker logs whisper-service > $LOG_PATH/whisper-service.log
84+
docker logs asr-service > $LOG_PATH/asr-service.log
85+
docker logs speecht5-service > $LOG_PATH/tts-service.log
86+
docker logs tts-service > $LOG_PATH/tts-service.log
87+
docker logs tgi-gaudi-server > $LOG_PATH/tgi-gaudi-server.log
88+
docker logs llm-tgi-gaudi-server > $LOG_PATH/llm-tgi-gaudi-server.log
89+
8290
echo "Result wrong."
8391
exit 1
8492
fi

AudioQnA/tests/test_audioqna_on_xeon.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ function build_docker_images() {
1919
docker build -t opea/speecht5:latest -f comps/tts/speecht5/Dockerfile .
2020
docker build -t opea/tts:latest -f comps/tts/Dockerfile .
2121

22-
docker pull ghcr.io/huggingface/tgi-gaudi:1.2.1
22+
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1
2323

2424
cd $WORKPATH/docker
2525
docker build --no-cache -t opea/audioqna:latest -f Dockerfile .

ChatQnA/chatqna.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ opea_micro_services:
4848
tgi_service:
4949
host: ${TGI_SERVICE_IP}
5050
ports: ${TGI_SERVICE_PORT}
51-
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
51+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
5252
volumes:
5353
- "./data:/data"
5454
runtime: habana

ChatQnA/docker/gaudi/docker_compose.yaml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,6 +43,7 @@ services:
4343
https_proxy: ${https_proxy}
4444
HABANA_VISIBLE_DEVICES: all
4545
OMPI_MCA_btl_vader_single_copy_mechanism: none
46+
MAX_WARMUP_SEQUENCE_LENGTH: 512
4647
command: --model-id ${EMBEDDING_MODEL_ID}
4748
embedding:
4849
image: opea/embedding-tei:latest
@@ -116,7 +117,7 @@ services:
116117
LANGCHAIN_PROJECT: "opea-reranking-service"
117118
restart: unless-stopped
118119
tgi-service:
119-
image: ghcr.io/huggingface/tgi-gaudi:2.0.0
120+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
120121
container_name: tgi-gaudi-server
121122
ports:
122123
- "8008:80"

ChatQnA/tests/test_chatqna_on_gaudi.sh

Lines changed: 14 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -20,12 +20,12 @@ function build_docker_images() {
2020
docker build -t opea/llm-tgi:latest -f comps/llms/text-generation/tgi/Dockerfile .
2121
docker build -t opea/dataprep-redis:latest -f comps/dataprep/redis/langchain/docker/Dockerfile .
2222

23-
cd ..
24-
git clone https://github.com/huggingface/tei-gaudi
25-
cd tei-gaudi/
26-
docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest .
23+
# cd ..
24+
# git clone https://github.com/huggingface/tei-gaudi
25+
# cd tei-gaudi/
26+
# docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest .
2727

28-
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.0
28+
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1
2929
docker pull ghcr.io/huggingface/text-embeddings-inference:cpu-1.2
3030

3131
cd $WORKPATH/docker
@@ -38,8 +38,13 @@ function build_docker_images() {
3838
}
3939

4040
function start_services() {
41-
cd $WORKPATH/docker/gaudi
41+
# build tei-gaudi for each test instead of pull from local registry
42+
cd $WORKPATH
43+
git clone https://github.com/huggingface/tei-gaudi
44+
cd tei-gaudi/
45+
docker build --no-cache -f Dockerfile-hpu -t opea/tei-gaudi:latest .
4246

47+
cd $WORKPATH/docker/gaudi
4348
export EMBEDDING_MODEL_ID="BAAI/bge-base-en-v1.5"
4449
export RERANK_MODEL_ID="BAAI/bge-reranker-base"
4550
export LLM_MODEL_ID="Intel/neural-chat-7b-v3-3"
@@ -66,6 +71,9 @@ function start_services() {
6671
sed -i "s#image: opea/chatqna-ui:latest#image: opea/chatqna-ui:${IMAGE_TAG}#g" docker_compose.yaml
6772
sed -i "s#image: opea/chatqna-conversation-ui:latest#image: opea/chatqna-conversation-ui:${IMAGE_TAG}#g" docker_compose.yaml
6873
sed -i "s#image: opea/*#image: ${IMAGE_REPO}opea/#g" docker_compose.yaml
74+
sed -i "s#image: ${IMAGE_REPO}opea/tei-gaudi:latest#image: opea/tei-gaudi:latest#g" docker_compose.yaml
75+
echo "cat docker_compose.yaml"
76+
cat docker_compose.yaml
6977
fi
7078

7179
# Start Docker Containers

CodeGen/codegen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ opea_micro_services:
66
tgi_service:
77
host: ${TGI_SERVICE_IP}
88
ports: ${TGI_SERVICE_PORT}
9-
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
9+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
1010
volumes:
1111
- "./data:/data"
1212
runtime: habana

CodeGen/docker/gaudi/docker_compose.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ version: "3.8"
66

77
services:
88
tgi_service:
9-
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
9+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
1010
container_name: tgi-gaudi-server
1111
ports:
1212
- "8028:80"
@@ -23,7 +23,7 @@ services:
2323
cap_add:
2424
- SYS_NICE
2525
ipc: host
26-
command: --model-id ${LLM_MODEL_ID}
26+
command: --model-id ${LLM_MODEL_ID} --max-input-length 1024 --max-total-tokens 2048
2727
llm:
2828
image: opea/llm-tgi:latest
2929
container_name: llm-tgi-gaudi-server

CodeGen/tests/test_codegen_on_gaudi.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ function build_docker_images() {
1515

1616
docker build -t opea/llm-tgi:latest -f comps/llms/text-generation/tgi/Dockerfile .
1717

18-
docker pull ghcr.io/huggingface/tgi-gaudi:1.2.1
18+
docker pull ghcr.io/huggingface/tgi-gaudi:2.0.1
1919

2020
cd $WORKPATH/docker
2121
docker build --no-cache -t opea/codegen:latest -f Dockerfile .

CodeTrans/codetrans.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ opea_micro_services:
66
tgi_service:
77
host: ${TGI_SERVICE_IP}
88
ports: ${TGI_SERVICE_PORT}
9-
image: ghcr.io/huggingface/tgi-gaudi:1.2.1
9+
image: ghcr.io/huggingface/tgi-gaudi:2.0.1
1010
volumes:
1111
- "./data:/data"
1212
runtime: habana

0 commit comments

Comments
 (0)