diff --git a/demos/python_demos/requirements.txt b/demos/python_demos/requirements.txt index 1595fc03c5..beda6bd0b4 100644 --- a/demos/python_demos/requirements.txt +++ b/demos/python_demos/requirements.txt @@ -1,13 +1,13 @@ --extra-index-url "https://download.pytorch.org/whl/cpu" -optimum-intel@git+https://github.com/huggingface/optimum-intel.git -pillow==12.1.1 +optimum-intel@git+https://github.com/huggingface/optimum-intel.git@2ca7f8bb09d536c0b6136830dfe31c7fefe14701 +pillow==12.2.0 tritonclient[grpc]==2.57.0 # Required to use batch string serialization/deserialization (4byte length prepend) numpy<2.0 huggingface_hub==0.32.0 nncf>=2.11.0 -sentence_transformers +sentence_transformers==5.3.0 sentencepiece==0.2.1 -transformers<=4.53 +transformers<4.58 einops torchvision timm==1.0.15 diff --git a/prepare_llm_models.sh b/prepare_llm_models.sh index 86395bbe31..67517d2d49 100755 --- a/prepare_llm_models.sh +++ b/prepare_llm_models.sh @@ -45,7 +45,7 @@ echo "Downloading LLM testing models to directory $1" export PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cpu https://storage.openvinotoolkit.org/simple/wheels/nightly" if [ "$2" = "docker" ]; then export PATH=$PATH:/opt/intel/openvino/python/bin - python3 -m pip install "optimum-intel"@git+https://github.com/huggingface/optimum-intel.git nncf sentence_transformers einops timm==1.0.22 sentencepiece + python3 -m pip install "optimum-intel"@git+https://github.com/huggingface/optimum-intel.git@2ca7f8bb09d536c0b6136830dfe31c7fefe14701 nncf sentence_transformers==5.3.0 einops timm==1.0.22 sentencepiece else python3 -m venv .venv . .venv/bin/activate