Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions helm-charts/common/llm-uservice/templates/configmap.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -70,6 +70,9 @@ data:
{{- if .Values.LLM_MODEL_ID }}
LLM_MODEL_ID: {{ .Values.LLM_MODEL_ID | quote }}
{{- end }}
{{- if .Values.OPENAI_API_KEY }}
OPENAI_API_KEY: {{ .Values.OPENAI_API_KEY | quote }}
{{- end }}
HF_HOME: "/tmp/.cache/huggingface"
{{- if not .Values.global.offline }}
HF_TOKEN: {{ .Values.global.HUGGINGFACEHUB_API_TOKEN | default .Values.global.HF_TOKEN | quote }}
Expand Down
3 changes: 3 additions & 0 deletions helm-charts/docsum/Chart.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,16 @@ dependencies:
- name: llm-uservice
version: 0-latest
repository: "file://../common/llm-uservice"
condition: llm-uservice.enabled
- name: whisper
version: 0-latest
repository: "file://../common/whisper"
condition: whisper.enabled
- name: ui
version: 0-latest
repository: "file://../common/ui"
alias: docsum-ui
condition: docsum-ui.enabled
- name: nginx
version: 0-latest
repository: "file://../common/nginx"
Expand Down
7 changes: 6 additions & 1 deletion helm-charts/docsum/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,12 @@ helm install docsum docsum --set global.HF_TOKEN=${HFTOKEN} --set global.modelUs
# helm install docsum docsum --set global.HF_TOKEN=${HFTOKEN} --values docsum/rocm-values.yaml
# To use AMD ROCm device with TGI
# helm install docsum docsum --set global.HF_TOKEN=${HFTOKEN} --values docsum/rocm-tgi-values.yaml

# To use with external OpenAI-compatible LLM endpoints (OpenAI, vLLM, TGI, etc.)
# This configures the llm-uservice to connect to external LLM providers while maintaining DocSum compatibility
# For OpenAI:
# helm install docsum docsum --values docsum/variant_external-llm-values.yaml --set llm-uservice.OPENAI_API_KEY="your-api-key" --set llm-uservice.LLM_ENDPOINT="https://api.openai.com" --set llm-uservice.LLM_MODEL_ID="gpt-4-turbo"
# For vLLM/TGI or other OpenAI-compatible endpoints:
# helm install docsum docsum --values docsum/variant_external-llm-values.yaml --set llm-uservice.LLM_ENDPOINT="http://your-server-url" --set llm-uservice.LLM_MODEL_ID="your-model"
Comment on lines +29 to +34
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you do separate PR adding similar stuff for FaqGen in ChatQnA Helm chart?

(Separate variant file, README instructions etc.)

```

## Verify
Expand Down
2 changes: 2 additions & 0 deletions helm-charts/docsum/templates/deployment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -38,10 +38,12 @@ spec:
value: {{ include "llm-uservice.fullname" (index .Subcharts "llm-uservice") }}
- name: LLM_SERVICE_PORT
value: {{ index .Values "llm-uservice" "service" "port" | quote }}
{{- if .Values.whisper.enabled }}
- name: ASR_SERVICE_HOST_IP
value: {{ include "whisper.fullname" (index .Subcharts "whisper") }}
- name: ASR_SERVICE_PORT
value: {{ index .Values "whisper" "service" "port" | quote }}
{{- end }}
securityContext:
{{- toYaml .Values.securityContext | nindent 12 }}
image: "{{ .Values.image.repository }}:{{ .Values.image.tag | default .Chart.AppVersion }}"
Expand Down
13 changes: 13 additions & 0 deletions helm-charts/docsum/values.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -57,7 +57,14 @@ tolerations: []
affinity: {}

# To override values in subchart llm-uservice
# For external LLM configuration, you can override these values here or use variant_external-llm-values.yaml
# Example for external OpenAI-compatible endpoints:
# DOCSUM_BACKEND: "vLLM" # Use vLLM backend for OpenAI-compatible APIs
# LLM_ENDPOINT: "https://api.openai.com" # External endpoint (omit /v1 suffix)
# LLM_MODEL_ID: "gpt-4-turbo" # Model to use
# OPENAI_API_KEY: "your-api-key" # API key for authentication
llm-uservice:
enabled: true
image:
repository: opea/llm-docsum
DOCSUM_BACKEND: "vLLM"
Expand All @@ -79,6 +86,8 @@ vllm:
nginx:
enabled: false
docsum-ui:
# if false, set also nginx.enabled=false
enabled: true
image:
repository: opea/docsum-gradio-ui
tag: "latest"
Expand All @@ -101,8 +110,12 @@ docsum-ui:
# type: ClusterIP

dashboard:
enabled: true
prefix: "OPEA DocSum"

whisper:
enabled: true

global:
http_proxy: ""
https_proxy: ""
Expand Down
17 changes: 17 additions & 0 deletions helm-charts/docsum/variant_external-llm-values.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,17 @@
# Copyright (C) 2024 Intel Corporation
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
# Copyright (C) 2024 Intel Corporation
# Copyright (C) 2025 Intel Corporation

# SPDX-License-Identifier: Apache-2.0

# External LLM configuration - configures llm-uservice to use external LLM providers
# This keeps the llm-uservice wrapper (required for /v1/docsum endpoint) but connects it to external LLMs
llm-uservice:
enabled: true # Keep the wrapper service for DocSum compatibility
DOCSUM_BACKEND: "vLLM" # Use vLLM backend for OpenAI-compatible APIs
LLM_ENDPOINT: "https://api.openai.com" # External LLM API endpoint (omit /v1 suffix)
OPENAI_API_KEY: "${OPENAI_API_KEY}" # API key for authentication
LLM_MODEL_ID: "gpt-4-turbo" # Model to use

# Disable local inference services since we're using external LLMs
vllm:
enabled: false
tgi:
enabled: false
Loading