Skip to content

Conversation

@lianhao
Copy link
Collaborator

@lianhao lianhao commented May 30, 2025

Description

  • Support air gapped environment
  • Temporary workaround for GenAIComps issue #1719

Issues

n/a.

Type of change

List the type of change like below. Please delete options that are not relevant.

  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds new functionality)
  • Breaking change (fix or feature that would break existing design and interface)

Dependencies

List the newly introduced 3rd party dependency if exists.

Tests

Describe the tests that you ran to verify your changes.

@lianhao lianhao requested a review from yongfengdu as a code owner May 30, 2025 03:32
@lianhao
Copy link
Collaborator Author

lianhao commented May 30, 2025

Let's wait for opea-project/GenAIComps#1758, opea-project/GenAIComps#1743 to land in first

Copy link
Collaborator

@eero-t eero-t left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Few typo fixes + text suggestions

@lianhao lianhao force-pushed the airgap-llm-uservice branch from bfa4919 to 93b6aea Compare June 4, 2025 01:27
@yongfengdu
Copy link
Collaborator

Not sure if this flag help with the air gapped mode? HF_HUB_OFFLINE=1
https://huggingface.co/docs/transformers/installation#offline-mode

@eero-t
Copy link
Collaborator

eero-t commented Jun 5, 2025

Not sure if this flag help with the air gapped mode? HF_HUB_OFFLINE=1 https://huggingface.co/docs/transformers/installation#offline-mode

If I understood correctly, it would help by blocking engines using HF download facilities, from trying to access online resources. Does that mean it having impact only on TEI / TGI, but not on vLLM, Ollama etc?

@lianhao
Copy link
Collaborator Author

lianhao commented Jun 6, 2025

Not sure if this flag help with the air gapped mode? HF_HUB_OFFLINE=1 https://huggingface.co/docs/transformers/installation#offline-mode

HF_HUB_OFFLINE only prevent the huggingface modules to download the model data from online resources. If the data is not existing in local cache dir, it will still cause problems.

The only way to use it is when the data is already exiting in local dir, but sometimes the underlying huggingface modules will still try to access online resources which cause the failure. Just like what we did in #1072. However, for llm-uservice, we don't need it.

lianhao and others added 3 commits June 9, 2025 09:28
Following GenAIExample to change the default model of vllm as a
temporary workaround for GenAIComps issue #1719.

Signed-off-by: Lianhao Lu <[email protected]>
@lianhao lianhao force-pushed the airgap-llm-uservice branch from 64246ee to 2251d0c Compare June 9, 2025 01:28
@yongfengdu yongfengdu merged commit b43a06e into opea-project:main Jun 9, 2025
14 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

[Feature] llm-docsum/llm-faqgen: support air-gapped environment

3 participants