Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
44 commits
Select commit Hold shift + click to select a range
0c2cb4c
Compose file to use Milvus as vdb
Feb 13, 2025
1f0a26e
Make a fix in the main README.md of the ChatQnA. (#1551)
kendall1997 Feb 14, 2025
d3db280
Expand CI test scope for common test scripts (#1554)
chensuyue Feb 14, 2025
1c6e267
Refactor AudioQnA README (#1508)
Spycsh Feb 15, 2025
43b0ad1
update vLLM CPU to the latest stable version (#1546)
xiguiw Feb 17, 2025
fa9f5d9
Fix minor typo in README (#1559)
jotpalch Feb 17, 2025
79ef340
Remove perf test code from test scripts. (#1510)
ZePan110 Feb 18, 2025
6af942b
Bump gradio from 5.5.0 to 5.11.0 in /MultimodalQnA/ui/gradio (#1391)
dependabot[bot] Feb 19, 2025
18f6342
Simplify ChatQnA AIPC user setting (#1573)
xiguiw Feb 19, 2025
f10e493
Fix mismatched environment variable (#1575)
xiguiw Feb 19, 2025
d0d5910
Fix trivy issue (#1569)
ZePan110 Feb 20, 2025
231d9b8
Update AgentQnA and DocIndexRetriever (#1564)
minmin-intel Feb 22, 2025
77041d6
Update README.md of AIPC quick start (#1578)
yinghu5 Feb 23, 2025
71c9bba
Fix "OpenAI" & "response" spelling (#1561)
eero-t Feb 25, 2025
3a92022
Bump gradio from 5.5.0 to 5.11.0 in /DocSum/ui/gradio (#1576)
dependabot[bot] Feb 25, 2025
287c7ca
Align mongo related image names with comps (#1543)
Spycsh Feb 27, 2025
fff6840
Add test script for Milvus on Xeon
ezelanza Feb 27, 2025
23f992e
Fix ChatQnA ROCm compose Readme file and absolute path for ROCM CI te…
artem-astafev Feb 27, 2025
d758818
Update compose_milvus.yaml
ezelanza Feb 27, 2025
e8c33c7
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Feb 27, 2025
192dff8
Fix async in chatqna bug (#1589)
XinyaoWa Feb 27, 2025
87e5716
Update compose_milvus.yaml
ezelanza Feb 27, 2025
a33c504
Create milvus.yaml
ezelanza Feb 27, 2025
4ab6916
Update compose_milvus.yaml
ezelanza Feb 28, 2025
e4a2794
Docker compose updated
ezelanza Feb 28, 2025
17499b0
Docker compose updated
ezelanza Feb 28, 2025
5e33b3e
Docker compose updated
ezelanza Feb 28, 2025
d25969f
Docker compose updated
ezelanza Feb 28, 2025
de0b324
Docker compose updated
ezelanza Feb 28, 2025
9895448
Docker compose updated
ezelanza Feb 28, 2025
4e61ef2
Docker compose updated
ezelanza Feb 28, 2025
16456e7
Docker compose updated
ezelanza Feb 28, 2025
bdf8014
Docker compose updated
ezelanza Feb 28, 2025
b7663cd
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Feb 28, 2025
b7431f0
Fix benchmark scripts (#1517)
chensuyue Feb 28, 2025
ea0cfcd
Top level README: add link to github.io documentation (#1584)
alexsin368 Feb 28, 2025
ec679a9
Update test_compose_milvus_on_xeon.sh
ezelanza Feb 28, 2025
5755e0c
fix click example button issue (#1586)
WenjiaoYue Feb 28, 2025
3d09215
Milvus compose
ezelanza Feb 28, 2025
80d9b32
Docker compose updated
ezelanza Feb 28, 2025
6ebe408
Docker compose updated
ezelanza Feb 28, 2025
5db12a5
Docker compose updated
ezelanza Feb 28, 2025
ceaa7bc
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot] Feb 28, 2025
cf1c114
Merge branch 'main' into milvus_compose
ezelanza Feb 28, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
227 changes: 227 additions & 0 deletions ChatQnA/docker_compose/intel/cpu/xeon/compose_milvus.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,227 @@
# Copyright (C) 2024 Intel Corporation
# SPDX-License-Identifier: Apache-2.0
services:
etcd:
container_name: milvus-etcd
image: quay.io/coreos/etcd:v3.5.5
environment:
- ETCD_AUTO_COMPACTION_MODE=revision
- ETCD_AUTO_COMPACTION_RETENTION=1000
- ETCD_QUOTA_BACKEND_BYTES=4294967296
- ETCD_SNAPSHOT_COUNT=50000
volumes:
- ${DOCKER_VOLUME_DIRECTORY:-.}/volumes/etcd:/etcd
command: etcd -advertise-client-urls=http://127.0.0.1:2379 -listen-client-urls http://0.0.0.0:2379 --data-dir /etcd
healthcheck:
test: ["CMD", "etcdctl", "endpoint", "health"]
interval: 30s
timeout: 20s
retries: 3

minio:
container_name: milvus-minio
image: minio/minio:RELEASE.2023-03-20T20-16-18Z
environment:
MINIO_ACCESS_KEY: minioadmin
MINIO_SECRET_KEY: minioadmin
ports:
- "${MINIO_PORT1:-5044}:9001"
- "${MINIO_PORT2:-5043}:9000"
volumes:
- ${DOCKER_VOLUME_DIRECTORY:-.}/volumes/minio:/minio_data
command: minio server /minio_data --console-address ":9001"
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9000/minio/health/live"]
interval: 30s
timeout: 20s
retries: 3

milvus-standalone:
container_name: milvus-standalone
image: milvusdb/milvus:v2.4.6
command: ["milvus", "run", "standalone"]
security_opt:
- seccomp:unconfined
environment:
ETCD_ENDPOINTS: etcd:2379
MINIO_ADDRESS: minio:9000
volumes:
- ${DOCKER_VOLUME_DIRECTORY:-.}/milvus.yaml:/milvus/configs/milvus.yaml
- ${DOCKER_VOLUME_DIRECTORY:-.}/volumes/milvus:/var/lib/milvus
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:9091/healthz"]
interval: 30s
start_period: 90s
timeout: 20s
retries: 3
ports:
- "19530:19530"
- "${MILVUS_STANDALONE_PORT:-9091}:9091"
depends_on:
- "etcd"
- "minio"

dataprep-milvus-service:
image: ${REGISTRY:-opea}/dataprep:${TAG:-latest}
container_name: dataprep-milvus-server
ports:
- "${DATAPREP_PORT:-11101}:5000"
ipc: host
environment:
no_proxy: ${no_proxy}
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
DATAPREP_COMPONENT_NAME: "OPEA_DATAPREP_MILVUS"
MILVUS_HOST: ${host_ip}
MILVUS_PORT: 19530
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
EMBEDDING_MODEL_ID: ${EMBEDDING_MODEL_ID}
LOGFLAG: ${LOGFLAG}
restart: unless-stopped
depends_on:
milvus-standalone:
condition: service_healthy
etcd:
condition: service_healthy
minio:
condition: service_healthy

retriever:
image: ${REGISTRY:-opea}/retriever:${TAG:-latest}
container_name: retriever-milvus-server
depends_on:
- milvus-standalone
ports:
- "7000:7000"
ipc: host
environment:
no_proxy: ${no_proxy}
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
MILVUS_HOST: ${host_ip}
MILVUS_PORT: 19530
TEI_EMBEDDING_ENDPOINT: http://tei-embedding-service:80
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
LOGFLAG: ${LOGFLAG}
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_MILVUS"
restart: unless-stopped

tei-embedding-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
container_name: tei-embedding-server
ports:
- "6006:80"
volumes:
- "./data:/data"
shm_size: 1g
environment:
no_proxy: ${no_proxy}
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
command: --model-id ${EMBEDDING_MODEL_ID} --auto-truncate

tei-reranking-service:
image: ghcr.io/huggingface/text-embeddings-inference:cpu-1.5
container_name: tei-reranking-server
ports:
- "8808:80"
volumes:
- "./data:/data"
shm_size: 1g
environment:
no_proxy: ${no_proxy}
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
HUGGINGFACEHUB_API_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
HF_HUB_DISABLE_PROGRESS_BARS: 1
HF_HUB_ENABLE_HF_TRANSFER: 0
command: --model-id ${RERANK_MODEL_ID} --auto-truncate

vllm-service:
image: ${REGISTRY:-opea}/vllm:${TAG:-latest}
container_name: vllm-service
ports:
- "9009:80"
volumes:
- "./data:/data"
shm_size: 128g
environment:
no_proxy: ${no_proxy}
http_proxy: ${http_proxy}
https_proxy: ${https_proxy}
HF_TOKEN: ${HUGGINGFACEHUB_API_TOKEN}
LLM_MODEL_ID: ${LLM_MODEL_ID}
VLLM_TORCH_PROFILER_DIR: "/mnt"
healthcheck:
test: ["CMD-SHELL", "curl -f http://$host_ip:9009/health || exit 1"]
interval: 10s
timeout: 10s
retries: 100
command: --model $LLM_MODEL_ID --host 0.0.0.0 --port 80

chatqna-xeon-backend-server:
image: ${REGISTRY:-opea}/chatqna:${TAG:-latest}
container_name: chatqna-xeon-backend-server
depends_on:
- milvus-standalone
- tei-embedding-service
- dataprep-milvus-service
- retriever
- tei-reranking-service
- vllm-service
ports:
- "8888:8888"
environment:
- no_proxy=${no_proxy}
- https_proxy=${https_proxy}
- http_proxy=${http_proxy}
- MEGA_SERVICE_HOST_IP=chatqna-xeon-backend-server
- EMBEDDING_SERVER_HOST_IP=tei-embedding-service
- RETRIEVER_SERVICE_HOST_IP=retriever
- EMBEDDING_SERVER_PORT=${EMBEDDING_SERVER_PORT:-80}
- RERANK_SERVER_HOST_IP=tei-reranking-service
- RERANK_SERVER_PORT=${RERANK_SERVER_PORT:-80}
- LLM_SERVER_HOST_IP=vllm-service
- LLM_SERVER_PORT=${LLM_SERVER_PORT:-80}
- LLM_MODEL=${LLM_MODEL_ID}
- LOGFLAG=${LOGFLAG}
ipc: host
restart: always

chatqna-xeon-ui-server:
image: ${REGISTRY:-opea}/chatqna-ui:${TAG:-latest}
container_name: chatqna-xeon-ui-server
depends_on:
- chatqna-xeon-backend-server
ports:
- "5173:5173"
ipc: host
restart: always

chatqna-xeon-nginx-server:
image: ${REGISTRY:-opea}/nginx:${TAG:-latest}
container_name: chatqna-xeon-nginx-server
depends_on:
- chatqna-xeon-backend-server
- chatqna-xeon-ui-server
ports:
- "${NGINX_PORT:-80}:80"
environment:
- no_proxy=${no_proxy}
- https_proxy=${https_proxy}
- http_proxy=${http_proxy}
- FRONTEND_SERVICE_IP=chatqna-xeon-ui-server
- FRONTEND_SERVICE_PORT=5173
- BACKEND_SERVICE_NAME=chatqna
- BACKEND_SERVICE_IP=chatqna-xeon-backend-server
- BACKEND_SERVICE_PORT=8888
- DATAPREP_SERVICE_IP=dataprep-milvus-service
- DATAPREP_SERVICE_PORT=5000
ipc: host
restart: always


networks:
default:
driver: bridge
Loading