Skip to content

Conversation

@sgurunat
Copy link
Contributor

Description

Added langchain vLLM support for Faq-Generation. Added new folder called vllm under faq-generation and have maintained the same file structure similar to tgi. Dockerfile, compose yaml and README contains details about how to build and use FaQ with vLLM.

This also contains changes related to authentication support for remote vLLM endpoints in llm.py file.

Since model is mandatory parameter for vLLM updated the FaqGenGateway to include the new parameter called "model" in Mega service gateway.py file

Issues

N/A

Type of change

  • New feature (non-breaking change which adds new functionality)

Dependencies

N/A

@codecov
Copy link

codecov bot commented Nov 11, 2024

Codecov Report

All modified and coverable lines are covered by tests ✅

Files with missing lines Coverage Δ
comps/cores/mega/gateway.py 29.01% <ø> (ø)

@lvliang-intel
Copy link
Collaborator

@sgurunat,

Please update comps/llms/faq-generation/vllm/langchain/Dockerfile to .github/workflows/docker/compose/llms-compose-cd.yaml. The yaml is used for release images build.

@sgurunat
Copy link
Contributor Author

sgurunat commented Nov 12, 2024

@lvliang-intel - I don't see the llm-compose-cd.yaml file. I did Update Branch and looks like the latest changes in main doesn't have this file.

@chensuyue chensuyue added this to the v1.1 milestone Nov 13, 2024
@sgurunat
Copy link
Contributor Author

@chensuyue @lvliang-intel Updated the git workflows to include vllm faqgen build in llm-compose.yaml file

@ftian1
Copy link
Collaborator

ftian1 commented Nov 13, 2024

at next release, we need consider how to merge such instances with others to avoid code duplication.

@ftian1 ftian1 merged commit f5c60f1 into opea-project:main Nov 13, 2024
madison-evans pushed a commit to SAPD-Intel/GenAIComps that referenced this pull request May 12, 2025
* Add model parameter for FaqGenGateway in gateway.py file

Signed-off-by: sgurunat <[email protected]>

* Add langchain vllm support for FaqGen along with authentication support for vllm endpoints

Signed-off-by: sgurunat <[email protected]>

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Updated docker_compose_llm.yaml and README file with vLLM information

Signed-off-by: sgurunat <[email protected]>

* [pre-commit.ci] auto fixes from pre-commit.com hooks

for more information, see https://pre-commit.ci

* Updated faq-vllm Dockerfile into llm-compose-cd.yaml under github workflows

Signed-off-by: sgurunat <[email protected]>

* Updated llm-compose.yaml file to include vllm faqgen build

Signed-off-by: sgurunat <[email protected]>

---------

Signed-off-by: sgurunat <[email protected]>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants