RAGflow/docker/docker-compose_gpu.yml

127 lines
3.1 KiB
YAML

include:
- ./docker-compose-base.yml
services:
ragflow:
depends_on:
mysql:
condition: service_healthy
image: ${RAGFLOW_IMAGE}
container_name: ragflowplus-server
ports:
- ${SVR_HTTP_PORT}:9380
- 80:80
- 443:443
volumes:
- ./ragflow-logs:/ragflow/logs
- ./nginx/ragflow.conf:/etc/nginx/conf.d/ragflow.conf
- ./nginx/proxy.conf:/etc/nginx/proxy.conf
- ./nginx/nginx.conf:/etc/nginx/nginx.conf
env_file: .env
environment:
- TZ=${TIMEZONE}
- HF_ENDPOINT=${HF_ENDPOINT}
- MACOS=${MACOS}
networks:
- ragflow
restart: on-failure
extra_hosts:
- "host.docker.internal:host-gateway"
# 后台信息管理系统前端
management-frontend:
container_name: ragflowplus-management-frontend
image: ${RAGFLOWPLUS_MANAGEMENT_WEB_IMAGE}
volumes:
- ./nginx/management_nginx.conf:/etc/nginx/conf.d/default.conf
ports:
- "8888:80"
depends_on:
- management-backend
environment:
- API_BASE_URL=/api
networks:
- ragflow
restart: on-failure
extra_hosts:
- "host.docker.internal:host-gateway"
# 后台信息管理系统后端
management-backend:
container_name: ragflowplus-management-backend
image: ${RAGFLOWPLUS_MANAGEMENT_SERVER_IMAGE}
ports:
- "5000:5000"
volumes:
- ./magic-pdf.json:/root/magic-pdf.json
depends_on:
mysql:
condition: service_healthy
es01:
condition: service_healthy
environment:
- FLASK_ENV=development
- CORS_ALLOWED_ORIGINS=http://management-frontend
- MANAGEMENT_ADMIN_USERNAME=${MANAGEMENT_ADMIN_USERNAME:-admin}
- MANAGEMENT_ADMIN_PASSWORD=${MANAGEMENT_ADMIN_PASSWORD:-12345678}
- MANAGEMENT_JWT_SECRET=${MANAGEMENT_JWT_SECRET:-12345678}
networks:
- ragflow
restart: on-failure
extra_hosts:
- "host.docker.internal:host-gateway"
deploy:
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
# vllm 嵌入模型
vllm-bge:
container_name: vllm-bge
image: vllm/vllm-openai:latest
ipc: host
volumes:
- ./models/bge-m3:/models
command: [
"--model", "/models",
"--served-model-name", "bge-m3",
"--dtype", "float16",
"--gpu-memory-utilization", "0.9",
]
ports:
- "8000:8000"
deploy:
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
networks:
- ragflow
# vllm 语言模型
vllm-deepseek:
container_name: vllm-deepseek
image: vllm/vllm-openai:latest
ipc: host
volumes:
- ./models/DeepSeek-R1-1.5B:/models
command: [
"--model", "/models",
"--served-model-name", "deepseek-r1",
"--dtype", "float16",
"--tensor-parallel-size", "1",
"--max-model-len", "4096"
]
ports:
- "8001:8000"
deploy:
resources:
reservations:
devices:
- driver: nvidia
capabilities: [gpu]
networks:
- ragflow