From c24898bf027abe95f807c70fd6fe8d8d204b4754 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?St=C3=A9phan?= Date: Sun, 15 Jun 2025 11:56:24 +0200 Subject: [PATCH] =?UTF-8?q?fix(pgpt)=20:=20suite=20=C3=A0=20mont=C3=A9e=20?= =?UTF-8?q?de=20version=20de=20docker?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit régénération du docker-compose.yaml pour pouvoir relancer le service --- pgpt/docker-compose.yaml | 30 +++++++----------------------- 1 file changed, 7 insertions(+), 23 deletions(-) diff --git a/pgpt/docker-compose.yaml b/pgpt/docker-compose.yaml index 1575757..6354075 100644 --- a/pgpt/docker-compose.yaml +++ b/pgpt/docker-compose.yaml @@ -3,10 +3,8 @@ services: #---- Private-GPT services --------- #----------------------------------- - # Private-GPT service for the Ollama CPU and GPU modes - # This service builds from an external Dockerfile and runs the Ollama mode. private-gpt-ollama: - image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-ollama # x-release-please-version + image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-ollama user: root build: context: . @@ -29,12 +27,10 @@ services: - ollama-api depends_on: ollama: - condition: service_healthy + condition: service_started - # Private-GPT service for the local mode - # This service builds from a local Dockerfile and runs the application in local mode. private-gpt-llamacpp-cpu: - image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-llamacpp-cpu # x-release-please-version + image: ${PGPT_IMAGE:-zylonai/private-gpt}:${PGPT_TAG:-0.6.2}-llamacpp-cpu user: root build: context: . @@ -56,22 +52,8 @@ services: #---- Ollama services -------------- #----------------------------------- - # Traefik reverse proxy for the Ollama service - # This will route requests to the Ollama service based on the profile. ollama: image: traefik:v2.10 - healthcheck: - test: - [ - "CMD", - "sh", - "-c", - "wget -q --spider http://ollama:11434 || exit 1", - ] - interval: 10s - retries: 3 - start_period: 5s - timeout: 5s ports: - "127.0.0.1:8080:8080" command: @@ -86,24 +68,26 @@ services: - ./.docker/router.yml:/etc/router.yml:ro extra_hosts: - "host.docker.internal:host-gateway" + security_opt: + - label:disable profiles: - "" - ollama-cpu - ollama-cuda - ollama-api - # Ollama service for the CPU mode ollama-cpu: image: ollama/ollama:latest ports: - "127.0.0.1:11434:11434" volumes: - ./models:/root/.ollama:Z + healthcheck: + disable: true profiles: - "" - ollama-cpu - # Ollama service for the CUDA mode ollama-cuda: image: ollama/ollama:latest ports: