-
Notifications
You must be signed in to change notification settings - Fork 218
vLLM support for FAQGen #884
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Signed-off-by: sgurunat <[email protected]>
…rt for vllm endpoints Signed-off-by: sgurunat <[email protected]>
for more information, see https://pre-commit.ci
Codecov ReportAll modified and coverable lines are covered by tests ✅
|
comps/llms/faq-generation/vllm/langchain/docker_compose_llm.yaml
Outdated
Show resolved
Hide resolved
Signed-off-by: sgurunat <[email protected]>
for more information, see https://pre-commit.ci
|
Please update comps/llms/faq-generation/vllm/langchain/Dockerfile to .github/workflows/docker/compose/llms-compose-cd.yaml. The yaml is used for release images build. |
…kflows Signed-off-by: sgurunat <[email protected]>
|
@lvliang-intel - I don't see the llm-compose-cd.yaml file. I did Update Branch and looks like the latest changes in main doesn't have this file. |
Signed-off-by: sgurunat <[email protected]>
Signed-off-by: sgurunat <[email protected]>
|
@chensuyue @lvliang-intel Updated the git workflows to include vllm faqgen build in llm-compose.yaml file |
|
at next release, we need consider how to merge such instances with others to avoid code duplication. |
* Add model parameter for FaqGenGateway in gateway.py file Signed-off-by: sgurunat <[email protected]> * Add langchain vllm support for FaqGen along with authentication support for vllm endpoints Signed-off-by: sgurunat <[email protected]> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Updated docker_compose_llm.yaml and README file with vLLM information Signed-off-by: sgurunat <[email protected]> * [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci * Updated faq-vllm Dockerfile into llm-compose-cd.yaml under github workflows Signed-off-by: sgurunat <[email protected]> * Updated llm-compose.yaml file to include vllm faqgen build Signed-off-by: sgurunat <[email protected]> --------- Signed-off-by: sgurunat <[email protected]> Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Description
Added langchain vLLM support for Faq-Generation. Added new folder called vllm under faq-generation and have maintained the same file structure similar to tgi. Dockerfile, compose yaml and README contains details about how to build and use FaQ with vLLM.
This also contains changes related to authentication support for remote vLLM endpoints in llm.py file.
Since model is mandatory parameter for vLLM updated the FaqGenGateway to include the new parameter called "model" in Mega service gateway.py file
Issues
N/A
Type of change
Dependencies
N/A