diff --git a/comps/third_parties/llama-vision/src/requirements-cpu.txt b/comps/third_parties/llama-vision/src/requirements-cpu.txt index 70cb2d74e9..336a66fccd 100644 --- a/comps/third_parties/llama-vision/src/requirements-cpu.txt +++ b/comps/third_parties/llama-vision/src/requirements-cpu.txt @@ -1,6 +1,6 @@ # This file was autogenerated by uv via the following command: # uv pip compile --index-strategy unsafe-best-match ./comps/third_parties/llama-vision/src/requirements.in --universal -o ./comps/third_parties/llama-vision/src/requirements-cpu.txt -accelerate==1.7.0 +accelerate==1.8.1 # via # -r ./comps/third_parties/llama-vision/src/requirements.in # optimum-habana @@ -41,14 +41,10 @@ colorama==0.4.6 ; sys_platform == 'win32' # click # colorlog # tqdm -coloredlogs==15.0.1 - # via optimum colorlog==6.9.0 # via trimesh datasets==3.6.0 - # via - # -r ./comps/third_parties/llama-vision/src/requirements.in - # optimum + # via -r ./comps/third_parties/llama-vision/src/requirements.in diffusers==0.33.1 # via optimum-habana dill==0.3.8 @@ -87,7 +83,7 @@ h11==0.16.0 # via # httpcore # uvicorn -hf-xet==1.1.4 ; platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64' +hf-xet==1.1.5 ; platform_machine == 'aarch64' or platform_machine == 'amd64' or platform_machine == 'arm64' or platform_machine == 'x86_64' # via huggingface-hub httpcore==1.0.9 # via httpx @@ -99,10 +95,10 @@ huggingface-hub==0.33.0 # datasets # diffusers # optimum + # optimum-habana + # sentence-transformers # tokenizers # transformers -humanfriendly==10.0 - # via coloredlogs idna==3.10 # via # anyio @@ -117,6 +113,8 @@ jax==0.5.3 # via docarray jaxlib==0.5.3 # via jax +joblib==1.5.1 + # via scikit-learn jsonschema==4.24.0 # via trimesh jsonschema-specifications==2025.4.1 @@ -171,6 +169,7 @@ numpy==1.26.4 # optimum # pandas # pycollada + # scikit-learn # scipy # shapely # transformers @@ -207,11 +206,11 @@ opentelemetry-semantic-conventions==0.55b1 # via opentelemetry-sdk opt-einsum==3.4.0 # via jax -optimum==1.13.2 +optimum==1.26.1 # via # -r ./comps/third_parties/llama-vision/src/requirements.in # optimum-habana -optimum-habana==1.8.0 +optimum-habana==1.18.0 # via optimum orjson==3.10.18 # via docarray @@ -230,6 +229,7 @@ pillow==11.2.1 # via # diffusers # docarray + # sentence-transformers # trimesh prometheus-client==0.22.1 # via prometheus-fastapi-instrumentator @@ -244,28 +244,25 @@ protobuf==5.29.5 # docarray # googleapis-common-protos # opentelemetry-proto - # transformers psutil==7.0.0 # via accelerate pyarrow==20.0.0 # via datasets pycollada==0.9 # via trimesh -pydantic==2.9.2 +pydantic==2.11.7 # via # -r ./comps/third_parties/llama-vision/src/requirements.in # docarray # fastapi -pydantic-core==2.23.4 +pydantic-core==2.33.2 # via pydantic pydub==0.25.1 # via # -r ./comps/third_parties/llama-vision/src/requirements.in # docarray -pygments==2.19.1 +pygments==2.19.2 # via rich -pyreadline3==3.5.4 ; sys_platform == 'win32' - # via humanfriendly python-dateutil==2.9.0.post0 # via # pandas @@ -306,13 +303,24 @@ safetensors==0.5.3 # accelerate # diffusers # transformers -scipy==1.15.3 +scikit-learn==1.7.0 + # via sentence-transformers +scipy==1.15.3 ; python_full_version < '3.11' # via # jax # jaxlib + # scikit-learn + # sentence-transformers # trimesh -sentencepiece==0.2.0 - # via transformers +scipy==1.16.0 ; python_full_version >= '3.11' + # via + # jax + # jaxlib + # scikit-learn + # sentence-transformers + # trimesh +sentence-transformers==3.3.1 + # via optimum-habana shapely==2.1.1 # via trimesh shortuuid==1.0.13 @@ -327,21 +335,22 @@ starlette==0.46.2 # prometheus-fastapi-instrumentator svg-path==6.3 # via trimesh -sympy==1.14.0 - # via - # optimum +threadpoolctl==3.6.0 + # via scikit-learn tokenizers==0.21.1 # via transformers tqdm==4.67.1 # via # datasets # huggingface-hub + # sentence-transformers # transformers -transformers==4.52.4 +transformers==4.49.0 # via # -r ./comps/third_parties/llama-vision/src/requirements.in # optimum # optimum-habana + # sentence-transformers trimesh==4.6.12 # via docarray types-pillow==10.2.0.20240822 @@ -365,9 +374,12 @@ typing-extensions==4.14.0 # referencing # rich # typing-inspect + # typing-inspection # uvicorn typing-inspect==0.9.0 # via docarray +typing-inspection==0.4.1 + # via pydantic tzdata==2025.2 # via pandas urllib3==2.5.0 diff --git a/comps/third_parties/llama-vision/src/requirements.in b/comps/third_parties/llama-vision/src/requirements.in index 2f250f1569..9ce435eb74 100644 --- a/comps/third_parties/llama-vision/src/requirements.in +++ b/comps/third_parties/llama-vision/src/requirements.in @@ -8,8 +8,8 @@ opentelemetry-exporter-otlp opentelemetry-sdk optimum[habana] prometheus-fastapi-instrumentator -pydantic==2.9.2 +pydantic pydub shortuuid -transformers>=4.50.0 +transformers uvicorn diff --git a/tests/language_detection/test_language_detection.sh b/tests/language_detection/test_language_detection.sh index a2a3651508..6ea0119765 100644 --- a/tests/language_detection/test_language_detection.sh +++ b/tests/language_detection/test_language_detection.sh @@ -7,7 +7,6 @@ set -x WORKPATH=$(dirname "$PWD") ip_address=$(hostname -I | awk '{print $1}') - export TAG=comps export PORT=8069 export service_name="language-detection" @@ -28,7 +27,7 @@ function start_service() { unset http_proxy cd $WORKPATH/comps/language_detection/deployment/docker_compose docker compose -f compose.yaml up ${service_name} -d - sleep 3s + sleep 10s } function validate_microservice() {