From 54cbad69a3bce346f70f3fe7fc20683100ec7594 Mon Sep 17 00:00:00 2001 From: "renovate[bot]" <29139614+renovate[bot]@users.noreply.github.com> Date: Thu, 28 Nov 2024 01:25:46 +0000 Subject: [PATCH 1/2] chore(deps): update keglin/pinchflat docker tag to v2024.11.27 (#5738) --- apps/pinchflat/config.json | 6 +++--- apps/pinchflat/docker-compose.yml | 2 +- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/apps/pinchflat/config.json b/apps/pinchflat/config.json index 2e3091b942..3b272fc84e 100644 --- a/apps/pinchflat/config.json +++ b/apps/pinchflat/config.json @@ -5,8 +5,8 @@ "available": true, "exposable": true, "id": "pinchflat", - "tipi_version": 14, - "version": "2024.10.30", + "tipi_version": 15, + "version": "2024.11.27", "categories": ["media"], "description": "Your next YouTube media manager", "short_desc": "Your next YouTube media manager", @@ -34,5 +34,5 @@ ], "supported_architectures": ["arm64", "amd64"], "created_at": 1691943801422, - "updated_at": 1730916717000 + "updated_at": 1732757066000 } diff --git a/apps/pinchflat/docker-compose.yml b/apps/pinchflat/docker-compose.yml index 4751e840d9..73a1e6122b 100644 --- a/apps/pinchflat/docker-compose.yml +++ b/apps/pinchflat/docker-compose.yml @@ -1,6 +1,6 @@ services: pinchflat: - image: keglin/pinchflat:v2024.10.30 + image: keglin/pinchflat:v2024.11.27 container_name: pinchflat environment: - BASIC_AUTH_USERNAME=${PINCHFLAT_BASIC_AUTH_USERNAME} From a60a808ecd321d3078abaa5a23c4e9a6c0406525 Mon Sep 17 00:00:00 2001 From: "renovate[bot]" <29139614+renovate[bot]@users.noreply.github.com> Date: Thu, 28 Nov 2024 01:27:47 +0000 Subject: [PATCH 2/2] chore(deps): update ollama/ollama docker tag to v0.4.6 (#5739) --- apps/ollama-amd/config.json | 6 +++--- apps/ollama-amd/docker-compose.yml | 2 +- apps/ollama-cpu/config.json | 6 +++--- apps/ollama-cpu/docker-compose.yml | 2 +- apps/ollama-nvidia/config.json | 6 +++--- apps/ollama-nvidia/docker-compose.yml | 2 +- 6 files changed, 12 insertions(+), 12 deletions(-) diff --git a/apps/ollama-amd/config.json b/apps/ollama-amd/config.json index 22cae62575..9f4625963c 100644 --- a/apps/ollama-amd/config.json +++ b/apps/ollama-amd/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11434, "id": "ollama-amd", - "tipi_version": 42, - "version": "0.4.5-rocm", + "tipi_version": 43, + "version": "0.4.6-rocm", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", @@ -16,5 +16,5 @@ "form_fields": [], "supported_architectures": ["arm64", "amd64"], "created_at": 1691943801422, - "updated_at": 1732597295000 + "updated_at": 1732757185000 } diff --git a/apps/ollama-amd/docker-compose.yml b/apps/ollama-amd/docker-compose.yml index 16aafa7f38..0c6ea47a40 100755 --- a/apps/ollama-amd/docker-compose.yml +++ b/apps/ollama-amd/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-amd: - image: ollama/ollama:0.4.5-rocm + image: ollama/ollama:0.4.6-rocm restart: unless-stopped container_name: ollama-amd environment: diff --git a/apps/ollama-cpu/config.json b/apps/ollama-cpu/config.json index fd86f30703..0d971a5d9a 100644 --- a/apps/ollama-cpu/config.json +++ b/apps/ollama-cpu/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11436, "id": "ollama-cpu", - "tipi_version": 42, - "version": "0.4.5", + "tipi_version": 43, + "version": "0.4.6", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", @@ -16,5 +16,5 @@ "form_fields": [], "supported_architectures": ["arm64", "amd64"], "created_at": 1691943801422, - "updated_at": 1732597314000 + "updated_at": 1732757189000 } diff --git a/apps/ollama-cpu/docker-compose.yml b/apps/ollama-cpu/docker-compose.yml index 1244e58c30..cf7516bd34 100755 --- a/apps/ollama-cpu/docker-compose.yml +++ b/apps/ollama-cpu/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-cpu: - image: ollama/ollama:0.4.5 + image: ollama/ollama:0.4.6 restart: unless-stopped container_name: ollama-cpu ports: diff --git a/apps/ollama-nvidia/config.json b/apps/ollama-nvidia/config.json index a09e2b1e9b..c159781337 100644 --- a/apps/ollama-nvidia/config.json +++ b/apps/ollama-nvidia/config.json @@ -5,8 +5,8 @@ "exposable": true, "port": 11435, "id": "ollama-nvidia", - "tipi_version": 42, - "version": "0.4.5", + "tipi_version": 43, + "version": "0.4.6", "categories": ["ai"], "description": "Get up and running with Llama 3, Mistral, Gemma, and other large language models.", "short_desc": "LLMs inference server with OpenAI compatible API", @@ -16,5 +16,5 @@ "form_fields": [], "supported_architectures": ["arm64", "amd64"], "created_at": 1691943801422, - "updated_at": 1732597317000 + "updated_at": 1732757191000 } diff --git a/apps/ollama-nvidia/docker-compose.yml b/apps/ollama-nvidia/docker-compose.yml index 2c18e189c3..5175ba5545 100755 --- a/apps/ollama-nvidia/docker-compose.yml +++ b/apps/ollama-nvidia/docker-compose.yml @@ -2,7 +2,7 @@ version: '3.7' services: ollama-nvidia: - image: ollama/ollama:0.4.5 + image: ollama/ollama:0.4.6 restart: unless-stopped container_name: ollama-nvidia ports: