-
Notifications
You must be signed in to change notification settings - Fork 98
Add support for external OpenAI-compatible LLM endpoints across Helm charts (chatqna, codegen) #993
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 19 commits
587db38
b53e498
3d9a276
ca216f8
82d8ae4
42e5e08
0c78cd9
b41d6cc
0362d65
f361956
3813ff5
f16d6fb
057fac1
8b1b207
3e57f52
5894ba0
509815c
1a00ed2
273fa44
6d20fca
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -38,6 +38,7 @@ dependencies: | |
| - name: tei | ||
| version: 0-latest | ||
| repository: "file://../common/tei" | ||
| condition: tei.enabled | ||
| - name: embedding-usvc | ||
| version: 0-latest | ||
| repository: "file://../common/embedding-usvc" | ||
|
|
@@ -65,15 +66,19 @@ dependencies: | |
| - name: retriever-usvc | ||
| version: 0-latest | ||
| repository: "file://../common/retriever-usvc" | ||
| condition: retriever-usvc.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| - name: data-prep | ||
| version: 0-latest | ||
| repository: "file://../common/data-prep" | ||
| condition: data-prep.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| - name: ui | ||
| alias: chatqna-ui | ||
| version: 0-latest | ||
| repository: "file://../common/ui" | ||
| condition: chatqna-ui.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
Comment on lines
+73
to
+78
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Disabling either (Because their names have dashes, Helm does not understand if one adds |
||
| - name: nginx | ||
| version: 0-latest | ||
| repository: "file://../common/nginx" | ||
| condition: nginx.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| version: 0-latest | ||
| appVersion: "v1.0" | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -45,6 +45,9 @@ helm install chatqna chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} -- | |
| # To use AMD ROCm device with TGI | ||
| #helm install chatqna chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} --set global.modelUseHostPath=${MODELDIR} --set tgi.LLM_MODEL_ID=${MODELNAME} -f chatqna/rocm-tgi-values.yaml | ||
|
|
||
| # To use with external OpenAI compatible LLM endpoint | ||
| #helm install chatqna chatqna -f chatqna/external-llm-values.yaml --set externalLLM.LLM_SERVER_HOST_IP="http://your-llm-server" --set externalLLM.LLM_MODEL="your-model" --set externalLLM.OPENAI_API_KEY="your-api-key" | ||
|
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Please change the file name
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. |
||
|
|
||
| # To deploy FaqGen | ||
| #helm install faqgen chatqna --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} --set global.modelUseHostPath=${MODELDIR} -f chatqna/faqgen-cpu-values.yaml | ||
|
|
||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,20 @@ | ||
| # Copyright (C) 2025 Intel Corporation | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
|
|
||
| # External LLM configuration override | ||
| externalLLM: | ||
| enabled: true # Enable external LLM service | ||
| LLM_SERVER_HOST_IP: "http://your-llm-server" # External LLM service host | ||
| LLM_MODEL: "your-model" # LLM model to use | ||
| OPENAI_API_KEY: "your-api-key" # OpenAI API key for authentication | ||
| LLM_SERVER_PORT: "80" # Port for the external LLM service | ||
|
|
||
| # Disable internal LLM services when using external LLM | ||
| llm-uservice: | ||
| enabled: false | ||
|
|
||
| vllm: | ||
| enabled: false | ||
|
|
||
| tgi: | ||
| enabled: false | ||
|
Comment on lines
+12
to
+20
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. This is missing: |
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -17,25 +17,32 @@ dependencies: | |
| - name: llm-uservice | ||
| version: 0-latest | ||
| repository: "file://../common/llm-uservice" | ||
| condition: llm-uservice.enabled | ||
| - name: tei | ||
| version: 0-latest | ||
| repository: "file://../common/tei" | ||
| condition: tei.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| - name: embedding-usvc | ||
| version: 0-latest | ||
| repository: "file://../common/embedding-usvc" | ||
| condition: embedding-usvc.enabled | ||
| - name: redis-vector-db | ||
| version: 0-latest | ||
| repository: "file://../common/redis-vector-db" | ||
| condition: redis-vector-db.enabled | ||
lianhao marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| - name: retriever-usvc | ||
| version: 0-latest | ||
| repository: "file://../common/retriever-usvc" | ||
| condition: retriever-usvc.enabled | ||
| - name: data-prep | ||
| version: 0-latest | ||
| repository: "file://../common/data-prep" | ||
| condition: data-prep.enabled | ||
| - name: ui | ||
| version: 0-latest | ||
| repository: "file://../common/ui" | ||
| alias: codegen-ui | ||
| condition: codegen-ui.enabled | ||
|
Comment on lines
+40
to
+45
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Disabling either |
||
| - name: nginx | ||
| version: 0-latest | ||
| repository: "file://../common/nginx" | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -34,7 +34,8 @@ helm install codegen codegen --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} -- | |
| # helm install codegen codegen --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} --set global.modelUseHostPath=${MODELDIR} --set llm-uservcie.LLM_MODEL_ID=${MODELNAME} --set vllm.LLM_MODEL_ID=${MODELNAME} -f codegen/rocm-values.yaml | ||
| # To use AMD ROCm device with TGI | ||
| # helm install codegen codegen --set global.HUGGINGFACEHUB_API_TOKEN=${HFTOKEN} --set global.modelUseHostPath=${MODELDIR} --set llm-uservcie.LLM_MODEL_ID=${MODELNAME} --set tgi.LLM_MODEL_ID=${MODELNAME} -f codegen/rocm-tgi-values.yaml | ||
|
|
||
| # To use with external OpenAI compatible LLM endpoint | ||
| # helm install codegen codegen -f codegen/external-llm-values.yaml --set externalLLM.LLM_SERVER_HOST_IP="http://your-llm-server" --set externalLLM.LLM_MODEL="your-model" --set externalLLM.OPENAI_API_KEY="your-api-key" | ||
|
Collaborator
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Please change the file name |
||
| ``` | ||
|
|
||
| ### IMPORTANT NOTE | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,18 @@ | ||
| # Copyright (C) 2024 Intel Corporation | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
|
|
||
| # External LLM configuration | ||
| externalLLM: | ||
| enabled: true # Enable external LLM service | ||
| LLM_SERVICE_HOST_IP: "http://your-llm-server" # External LLM service host | ||
| LLM_MODEL_ID: "your-model" # LLM model to use | ||
| OPENAI_API_KEY: "your-api-key" # OpenAI API key for authentication | ||
| LLM_SERVER_PORT: "80" # Port for the external LLM service | ||
|
|
||
| # Disable internal LLM services when using external LLM | ||
| tgi: | ||
| enabled: false | ||
| vllm: | ||
| enabled: false | ||
| llm-uservice: | ||
| enabled: false |
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,21 @@ | ||
| # Copyright (C) 2024 Intel Corporation | ||
| # SPDX-License-Identifier: Apache-2.0 | ||
|
|
||
| image: | ||
| repository: opea/productivity-suite-react-ui-server | ||
| tag: "latest" | ||
|
|
||
| containerPort: 5173 | ||
| APP_BACKEND_SERVICE_ENDPOINT_CHATQNA: "/v1/chatqna" | ||
| APP_DATAPREP_DELETE_FILE_ENDPOINT: "/v1/dataprep/delete" | ||
| APP_BACKEND_SERVICE_ENDPOINT_CODEGEN: "/v1/codegen" | ||
| APP_BACKEND_SERVICE_ENDPOINT_DOCSUM: "/v1/docsum" | ||
| APP_DATAPREP_SERVICE_ENDPOINT: "/v1/dataprep/ingest" | ||
| APP_DATAPREP_GET_FILE_ENDPOINT: "/v1/dataprep/get" | ||
| APP_CHAT_HISTORY_CREATE_ENDPOINT: "/v1/chathistory/create" | ||
| APP_CHAT_HISTORY_DELETE_ENDPOINT: "/v1/chathistory/delete" | ||
| APP_CHAT_HISTORY_GET_ENDPOINT: "/v1/chathistory/get" | ||
| APP_PROMPT_SERVICE_GET_ENDPOINT: "/v1/prompt/get" | ||
| APP_PROMPT_SERVICE_CREATE_ENDPOINT: "/v1/prompt/create" | ||
| APP_PROMPT_SERVICE_DELETE_ENDPOINT: "/v1/prompt/delete" | ||
| APP_KEYCLOAK_SERVICE_ENDPOINT: "/auth" |
Uh oh!
There was an error while loading. Please reload this page.