From 8cf825d1076bfb9f4d0d3aec762c940b2aa87bff Mon Sep 17 00:00:00 2001 From: Lianhao Lu Date: Fri, 24 Jan 2025 09:52:08 +0800 Subject: [PATCH] lvm-uservice: adapt to config changes (#755) Adapt to config changes of lvm source code. Signed-off-by: Lianhao Lu (cherry picked from commit bdb50403270fe77cfd6c8780cd193260b9452bc7) --- helm-charts/common/lvm-uservice/templates/configmap.yaml | 4 +++- helm-charts/common/lvm-uservice/values.yaml | 2 ++ 2 files changed, 5 insertions(+), 1 deletion(-) diff --git a/helm-charts/common/lvm-uservice/templates/configmap.yaml b/helm-charts/common/lvm-uservice/templates/configmap.yaml index f65d5f494..616efe956 100644 --- a/helm-charts/common/lvm-uservice/templates/configmap.yaml +++ b/helm-charts/common/lvm-uservice/templates/configmap.yaml @@ -10,6 +10,7 @@ metadata: data: {{- if eq "TGI" .Values.LVM_BACKEND }} LVM_COMPONENT_NAME: "OPEA_TGI_LLAVA_LVM" + MAX_IMAGES: {{ .Values.MAX_IMAGES | default 1 | quote }} {{- if not .Values.LVM_ENDPOINT }} LVM_ENDPOINT: "http://{{ .Release.Name }}-tgi" {{- end }} @@ -19,10 +20,11 @@ data: {{- end }} {{- if eq "LLaVA" .Values.LVM_BACKEND }} LVM_COMPONENT_NAME: "OPEA_LLAVA_LVM" + MAX_IMAGES: {{ .Values.MAX_IMAGES | default 1 | quote }} {{- else if eq "VideoLlama" .Values.LVM_BACKEND }} LVM_COMPONENT_NAME: "OPEA_VIDEO_LLAMA_LVM" {{- else if eq "LlamaVision" .Values.LVM_BACKEND }} - LVM_COMPONENT_NAME: "OPEA_LLAVA_VISION_LVM" + LVM_COMPONENT_NAME: "OPEA_LLAMA_VISION_LVM" {{- else if eq "PredictionGuard" .Values.LVM_BACKEND }} LVM_COMPONENT_NAME: "OPEA_PREDICTION_GUARD_LVM" {{- else }} diff --git a/helm-charts/common/lvm-uservice/values.yaml b/helm-charts/common/lvm-uservice/values.yaml index ff1c16f72..988f19d9e 100644 --- a/helm-charts/common/lvm-uservice/values.yaml +++ b/helm-charts/common/lvm-uservice/values.yaml @@ -9,6 +9,8 @@ LOGFLAG: "" # backend inference engine to use, i.e. TGI, LLaVA, VideoLlama, LlamaVision, PredictionGuard LVM_BACKEND: "TGI" +# maximum image number sent to backend, only valid for TGI, LLaVa backend +MAX_IMAGES: 1 # inference engine service URL, e.g. http://tgi:80 LVM_ENDPOINT: ""