From b58701bcd0b08e2018e6024a3ca1db2502b67e6d Mon Sep 17 00:00:00 2001 From: Dex Burgess Date: Wed, 28 May 2025 03:01:18 +0000 Subject: [PATCH] Update qnap_mediaui.yaml --- qnap_mediaui.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/qnap_mediaui.yaml b/qnap_mediaui.yaml index dcafe27..1e683d2 100644 --- a/qnap_mediaui.yaml +++ b/qnap_mediaui.yaml @@ -3,7 +3,7 @@ services: # Photos immich-server: container_name: immich_server - image: ghcr.io/immich-app/immich-server:{$IMMICH_VERSION:-release} + image: ghcr.io/immich-app/immich-server:$IMMICH_VERSION # extends: # file: hwaccel.transcoding.yml # service: cpu # set to one of [nvenc, quicksync, rkmpp, vaapi, vaapi-wsl] for accelerated transcoding @@ -27,7 +27,7 @@ services: container_name: immich_machine_learning # For hardware acceleration, add one of -[armnn, cuda, rocm, openvino, rknn] to the image tag. # Example tag: ${IMMICH_VERSION:-release}-cuda - image: ghcr.io/immich-app/immich-machine-learning:{$IMMICH_VERSION:-release} + image: ghcr.io/immich-app/immich-machine-learning:$IMMICH_VERSION # extends: # uncomment this section for hardware acceleration - see https://immich.app/docs/features/ml-hardware-acceleration # file: hwaccel.ml.yml # service: cpu # set to one of [armnn, cuda, rocm, openvino, openvino-wsl, rknn] for accelerated inference - use the `-wsl` version for WSL2 where applicable