From dfc774632c2712d0c747fad314c2740fc2aa211b Mon Sep 17 00:00:00 2001 From: Ekaterina Aidova Date: Thu, 12 Sep 2024 20:36:00 +0400 Subject: [PATCH] revise peft installation in notebooks (#2387) --- .../distil-whisper-asr.ipynb | 2 +- notebooks/instant-id/instant-id.ipynb | 4 +--- ...-consistency-models-image-generation.ipynb | 21 +------------------ .../lcm-lora-controlnet.ipynb | 2 +- .../paint-by-example/paint-by-example.ipynb | 2 +- notebooks/photo-maker/photo-maker.ipynb | 2 +- .../pix2struct-docvqa/pix2struct-docvqa.ipynb | 2 +- notebooks/qrcode-monster/qrcode-monster.ipynb | 2 +- notebooks/sdxl-turbo/sdxl-turbo.ipynb | 2 +- .../sketch-to-image-pix2pix-turbo.ipynb | 2 +- .../sound-generation-audioldm2.ipynb | 17 ++++++++++++++- .../speculative-sampling.ipynb | 2 +- .../stable-diffusion-ip-adapter.ipynb | 2 +- .../stable-diffusion-xl/segmind-vegart.ipynb | 2 +- .../stable-diffusion-xl.ipynb | 2 +- .../stable-video-diffusion.ipynb | 2 +- .../whisper-subtitles-generation.ipynb | 2 +- .../wuerstchen-image-generation.ipynb | 2 +- 18 files changed, 33 insertions(+), 39 deletions(-) diff --git a/notebooks/distil-whisper-asr/distil-whisper-asr.ipynb b/notebooks/distil-whisper-asr/distil-whisper-asr.ipynb index b453d0d90d8..363e85d378a 100644 --- a/notebooks/distil-whisper-asr/distil-whisper-asr.ipynb +++ b/notebooks/distil-whisper-asr/distil-whisper-asr.ipynb @@ -70,7 +70,7 @@ }, "outputs": [], "source": [ - "%pip install -q \"transformers>=4.35\" \"torch>=2.1,<2.4.0\" \"torchvision<0.19.0\" \"onnx<1.16.2\" \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"transformers>=4.35\" \"torch>=2.1,<2.4.0\" \"torchvision<0.19.0\" \"onnx<1.16.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"git+https://github.com/huggingface/optimum-intel.git\"\n", "%pip install -q \"openvino>=2023.2.0\" datasets \"gradio>=4.0\" \"librosa\" \"soundfile\"\n", "%pip install -q \"nncf>=2.6.0\" \"jiwer\"\n", diff --git a/notebooks/instant-id/instant-id.ipynb b/notebooks/instant-id/instant-id.ipynb index dac69096736..d96ec508a06 100644 --- a/notebooks/instant-id/instant-id.ipynb +++ b/notebooks/instant-id/instant-id.ipynb @@ -102,7 +102,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"openvino>=2023.3.0\" opencv-python transformers diffusers accelerate gdown \"scikit-image>=0.19.2\" \"gradio>=4.19\" \"nncf>=2.9.0\" \"datasets>=2.14.6\" \"peft==0.6.2\"" + "%pip install -q \"openvino>=2023.3.0\" opencv-python transformers \"diffusers>=0.24.0\" accelerate gdown \"scikit-image>=0.19.2\" \"gradio>=4.19\" \"nncf>=2.9.0\" \"datasets>=2.14.6\" \"peft>=0.6.2\"" ] }, { @@ -246,8 +246,6 @@ " outputs = self.ov_model.outputs\n", " self.input_mean = 127.5\n", " self.input_std = 128.0\n", - " # print(self.output_names)\n", - " # assert len(outputs)==10 or len(outputs)==15\n", " self.use_kps = False\n", " self._anchor_ratio = 1.0\n", " self._num_anchors = 1\n", diff --git a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb index c572108d2cf..154fd153de4 100644 --- a/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb +++ b/notebooks/latent-consistency-models-image-generation/latent-consistency-models-image-generation.ipynb @@ -1,7 +1,6 @@ { "cells": [ { - "attachments": {}, "cell_type": "markdown", "id": "37b660e4-3cba-412b-ab1a-b4acb88ff329", "metadata": {}, @@ -47,7 +46,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "ee66539d-99cc-45a3-80e4-4fbd6b520650", "metadata": {}, @@ -64,11 +62,10 @@ "outputs": [], "source": [ "%pip install -q \"torch>=2.1\" --index-url https://download.pytorch.org/whl/cpu\n", - "%pip install -q \"openvino>=2023.1.0\" transformers \"diffusers>=0.23.1\" pillow \"gradio>=4.19\" \"nncf>=2.7.0\" \"datasets>=2.14.6\" \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu" + "%pip install -q \"openvino>=2024.3.0\" transformers \"diffusers>=0.30.1\" pillow \"gradio>=4.19\" \"nncf>=2.12.0\" \"datasets>=2.14.6\" --extra-index-url https://download.pytorch.org/whl/cpu" ] }, { - "attachments": {}, "cell_type": "markdown", "id": "9523be62-2c38-480d-ae53-13910ae6d049", "metadata": {}, @@ -221,7 +218,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "8972c4f5-4dbb-4650-a97a-8debbfab1d58", "metadata": {}, @@ -310,7 +306,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "ea926988-97d0-4722-9c5d-72a16635c86e", "metadata": {}, @@ -371,7 +366,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "6ecdac70-4194-4345-b816-26454e5941d4", "metadata": {}, @@ -435,7 +429,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "d38976ba-29bf-49f2-8fc8-02d66f342d07", "metadata": {}, @@ -672,7 +665,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "8de4233d-f6df-44a5-a319-af24fd2d6ee1", "metadata": {}, @@ -738,7 +730,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "8ca4d4ac-66e0-4636-bb26-d3e66eb90639", "metadata": {}, @@ -765,7 +756,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "21bc3f1b-c324-4f8d-bcf9-19504f5826f5", "metadata": {}, @@ -837,7 +827,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "7600eaf2-6a6a-4f07-8fdc-fa7155acc842", "metadata": {}, @@ -846,7 +835,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "919bf570", "metadata": {}, @@ -899,7 +887,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "7383b987", "metadata": {}, @@ -927,7 +914,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "2931a018", "metadata": {}, @@ -1040,7 +1026,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "0421a05a", "metadata": {}, @@ -1231,7 +1216,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "d90608e5", "metadata": {}, @@ -1294,7 +1278,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "3ea2cfd0", "metadata": {}, @@ -1372,7 +1355,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "1cfcd78e", "metadata": {}, @@ -1409,7 +1391,6 @@ ] }, { - "attachments": {}, "cell_type": "markdown", "id": "035dffa6-0ce9-4d22-8638-a2217d06e828", "metadata": {}, diff --git a/notebooks/latent-consistency-models-image-generation/lcm-lora-controlnet.ipynb b/notebooks/latent-consistency-models-image-generation/lcm-lora-controlnet.ipynb index fbb1ac762eb..fb20a2a6f7e 100644 --- a/notebooks/latent-consistency-models-image-generation/lcm-lora-controlnet.ipynb +++ b/notebooks/latent-consistency-models-image-generation/lcm-lora-controlnet.ipynb @@ -127,7 +127,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"torch\" transformers \"diffusers>=0.22.0\" \"controlnet-aux>=0.0.6\" \"peft==0.6.2\" accelerate --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"torch\" transformers \"diffusers>=0.24.0\" \"controlnet-aux>=0.0.6\" \"peft>=0.6.2\" accelerate --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"openvino>=2023.2.0\" pillow \"gradio>=4.19\" \"datasets>=2.14.6\" \"nncf>=2.7.0\"" ] }, diff --git a/notebooks/paint-by-example/paint-by-example.ipynb b/notebooks/paint-by-example/paint-by-example.ipynb index 9a6b2f584f4..9c81095af5b 100644 --- a/notebooks/paint-by-example/paint-by-example.ipynb +++ b/notebooks/paint-by-example/paint-by-example.ipynb @@ -58,7 +58,7 @@ "outputs": [], "source": [ "%pip install -q \"torch>=2.1\" torchvision --extra-index-url \"https://download.pytorch.org/whl/cpu\"\n", - "%pip install -q \"diffusers>=0.25.0\" \"peft==0.6.2\" \"openvino>=2023.2.0\" \"transformers>=4.25.1\" ipywidgets opencv-python pillow \"nncf>=2.7.0\" \"gradio==3.44.1\" tqdm" + "%pip install -q \"diffusers>=0.25.0\" \"peft>=0.6.2\" \"openvino>=2023.2.0\" \"transformers>=4.25.1\" ipywidgets opencv-python pillow \"nncf>=2.7.0\" \"gradio==3.44.1\" tqdm" ] }, { diff --git a/notebooks/photo-maker/photo-maker.ipynb b/notebooks/photo-maker/photo-maker.ipynb index 5e1fc2d8e2a..773e5062df2 100644 --- a/notebooks/photo-maker/photo-maker.ipynb +++ b/notebooks/photo-maker/photo-maker.ipynb @@ -107,7 +107,7 @@ "outputs": [], "source": [ "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \\\n", - "transformers \"torch>=2.1\" \"diffusers>=0.26,<0.30\" \"gradio>=4.19\" \"openvino>=2024.0.0\" \"einops\" torchvision \"peft==0.6.2\" \"nncf>=2.9.0\" \"protobuf==3.20.3\" \"insightface\" \"onnxruntime\"" + "transformers \"torch>=2.1\" \"diffusers>=0.26,<0.30\" \"gradio>=4.19\" \"openvino>=2024.0.0\" \"einops\" torchvision \"peft>=0.6.2\" \"nncf>=2.9.0\" \"protobuf==3.20.3\" \"insightface\" \"onnxruntime\"" ] }, { diff --git a/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb b/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb index cba6296b4a4..a5c624a0ed8 100644 --- a/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb +++ b/notebooks/pix2struct-docvqa/pix2struct-docvqa.ipynb @@ -75,7 +75,7 @@ "outputs": [], "source": [ "%pip install -q \"torch>=2.1\" torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu\n", - "%pip install -q \"git+https://github.com/huggingface/optimum-intel.git\" \"openvino>=2023.1.0\" \"transformers>=4.33.0\" \"peft==0.6.2\" \"onnx<1.16.2\" \"gradio>=4.19\" --extra-index-url https://download.pytorch.org/whl/cpu" + "%pip install -q \"git+https://github.com/huggingface/optimum-intel.git\" \"openvino>=2023.1.0\" \"onnx<1.16.2\" \"gradio>=4.19\" --extra-index-url https://download.pytorch.org/whl/cpu" ] }, { diff --git a/notebooks/qrcode-monster/qrcode-monster.ipynb b/notebooks/qrcode-monster/qrcode-monster.ipynb index a4c5f2513af..bcd90d71b29 100644 --- a/notebooks/qrcode-monster/qrcode-monster.ipynb +++ b/notebooks/qrcode-monster/qrcode-monster.ipynb @@ -63,7 +63,7 @@ }, "outputs": [], "source": [ - "%pip install -q accelerate diffusers transformers \"torch>=2.1\" \"gradio>=4.19\" qrcode opencv-python \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q accelerate \"diffusers>=0.24.0\" transformers \"torch>=2.1\" \"gradio>=4.19\" qrcode opencv-python \"peft>=0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"openvino>=2023.1.0\" \"nncf>=2.7.0\"" ] }, diff --git a/notebooks/sdxl-turbo/sdxl-turbo.ipynb b/notebooks/sdxl-turbo/sdxl-turbo.ipynb index 02cdc910bea..16a58724f18 100644 --- a/notebooks/sdxl-turbo/sdxl-turbo.ipynb +++ b/notebooks/sdxl-turbo/sdxl-turbo.ipynb @@ -58,7 +58,7 @@ "outputs": [], "source": [ "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \\\n", - "\"torch>=2.1\" transformers diffusers \"git+https://github.com/huggingface/optimum-intel.git\" \"gradio>=4.19\" \"peft==0.6.2\" \"openvino>=2023.3.0\"" + "\"torch>=2.1\" transformers \"diffusers>=0.24.0\" \"git+https://github.com/huggingface/optimum-intel.git\" \"gradio>=4.19\" \"peft>=0.6.2\" \"openvino>=2023.3.0\"" ] }, { diff --git a/notebooks/sketch-to-image-pix2pix-turbo/sketch-to-image-pix2pix-turbo.ipynb b/notebooks/sketch-to-image-pix2pix-turbo/sketch-to-image-pix2pix-turbo.ipynb index 5c693fcf28b..41ec7d7f657 100644 --- a/notebooks/sketch-to-image-pix2pix-turbo/sketch-to-image-pix2pix-turbo.ipynb +++ b/notebooks/sketch-to-image-pix2pix-turbo/sketch-to-image-pix2pix-turbo.ipynb @@ -55,7 +55,7 @@ }, "outputs": [], "source": [ - "%pip install -q \"openvino>=2024.1.0\" \"torch>=2.1\" torchvision \"diffusers==0.25.1\" \"peft==0.6.2\" transformers tqdm pillow opencv-python \"gradio==3.43.1\" --extra-index-url https://download.pytorch.org/whl/cpu" + "%pip install -q \"openvino>=2024.1.0\" \"torch>=2.1\" torchvision \"diffusers==0.25.1\" \"peft>=0.6.2\" transformers tqdm pillow opencv-python \"gradio==3.43.1\" --extra-index-url https://download.pytorch.org/whl/cpu" ] }, { diff --git a/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb b/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb index 8dcc107d70a..cc7d02e31ef 100644 --- a/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb +++ b/notebooks/sound-generation-audioldm2/sound-generation-audioldm2.ipynb @@ -1,6 +1,7 @@ { "cells": [ { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -45,6 +46,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -60,11 +62,12 @@ }, "outputs": [], "source": [ - "%pip install -qU accelerate \"diffusers>=0.30.0\" \"transformers>=4.43\" \"torch>=2.1\" \"gradio>=4.19\" \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -qU accelerate \"diffusers>=0.30.0\" \"transformers>=4.43\" \"torch>=2.1\" \"gradio>=4.19\" \"peft>=0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"openvino>=2024.0.0\"" ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -237,6 +240,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -275,6 +279,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -349,6 +354,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -396,6 +402,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -445,6 +452,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -515,6 +523,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -557,6 +566,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -633,6 +643,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -689,6 +700,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -736,6 +748,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -854,6 +867,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ @@ -945,6 +959,7 @@ ] }, { + "attachments": {}, "cell_type": "markdown", "metadata": {}, "source": [ diff --git a/notebooks/speculative-sampling/speculative-sampling.ipynb b/notebooks/speculative-sampling/speculative-sampling.ipynb index 24c3f65f124..139484b5230 100644 --- a/notebooks/speculative-sampling/speculative-sampling.ipynb +++ b/notebooks/speculative-sampling/speculative-sampling.ipynb @@ -73,7 +73,7 @@ "%pip install -Uq pip\n", "%pip uninstall -q -y optimum optimum-intel\n", "%pip install --pre -Uq \"openvino>=2024.2.0\" openvino-tokenizers[transformers] --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly\n", - "%pip install -q --upgrade transformers \"torch>=2.1\" \"torchvision\" \"gradio>=4.19\" accelerate \"onnx<1.16.2\" ipywidgets \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q --upgrade transformers \"torch>=2.1\" \"torchvision\" \"gradio>=4.19\" accelerate \"onnx<1.16.2\" ipywidgets --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"git+https://github.com/huggingface/optimum-intel.git\"" ] }, diff --git a/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb b/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb index ead66506f58..aaefdd771c4 100644 --- a/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb +++ b/notebooks/stable-diffusion-ip-adapter/stable-diffusion-ip-adapter.ipynb @@ -69,7 +69,7 @@ "source": [ "import platform\n", "\n", - "%pip install -q \"torch>=2.1\" transformers accelerate diffusers \"openvino>=2023.3.0\" \"gradio>=4.19\" opencv-python \"peft==0.6.2\" \"protobuf>=3.20\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"torch>=2.1\" transformers accelerate \"diffusers>=0.24.0\" \"openvino>=2023.3.0\" \"gradio>=4.19\" opencv-python \"peft>=0.6.2\" \"protobuf>=3.20\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "\n", "if platform.system() != \"Windows\":\n", " %pip install -q \"matplotlib>=3.4\"\n", diff --git a/notebooks/stable-diffusion-xl/segmind-vegart.ipynb b/notebooks/stable-diffusion-xl/segmind-vegart.ipynb index 69530ae706e..e4b2dc21974 100644 --- a/notebooks/stable-diffusion-xl/segmind-vegart.ipynb +++ b/notebooks/stable-diffusion-xl/segmind-vegart.ipynb @@ -64,7 +64,7 @@ "outputs": [], "source": [ "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu\\\n", - "\"torch>=2.1\" transformers diffusers \"git+https://github.com/huggingface/optimum-intel.git\" \"gradio>=4.19\" \"openvino>=2023.3.0\" \"peft==0.6.2\"" + "\"torch>=2.1\" transformers \"diffusers>=0.24.0\" \"git+https://github.com/huggingface/optimum-intel.git\" \"gradio>=4.19\" \"openvino>=2023.3.0\" \"peft>=0.6.2\"" ] }, { diff --git a/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb b/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb index 6e0d0ed03ca..f9fb73ae9d6 100644 --- a/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb +++ b/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb @@ -85,7 +85,7 @@ }, "outputs": [], "source": [ - "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \"torch>=2.1,<2.4\" \"torchvision<0.19.0\" \"diffusers>=0.18.0\" \"invisible-watermark>=0.2.0\" \"transformers>=4.33.0\" \"accelerate\" \"onnx<1.16.2\" \"peft==0.6.2\"\n", + "%pip install -q --extra-index-url https://download.pytorch.org/whl/cpu \"torch>=2.1\" \"torchvision\" \"diffusers>=0.24.0\" \"invisible-watermark>=0.2.0\" \"transformers>=4.33.0\" \"accelerate\" \"onnx!=1.16.2\" \"peft>=0.6.2\"\n", "%pip install -q \"git+https://github.com/huggingface/optimum-intel.git\"\n", "%pip install -q \"openvino>=2023.1.0\" \"gradio>=4.19\" \"nncf>=2.9.0\"" ] diff --git a/notebooks/stable-video-diffusion/stable-video-diffusion.ipynb b/notebooks/stable-video-diffusion/stable-video-diffusion.ipynb index 95edab32494..91ce767b3ef 100644 --- a/notebooks/stable-video-diffusion/stable-video-diffusion.ipynb +++ b/notebooks/stable-video-diffusion/stable-video-diffusion.ipynb @@ -58,7 +58,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -q \"torch>=2.1\" \"torchvision<0.19.0\" \"diffusers>=0.25\" \"peft==0.6.2\" \"transformers\" \"openvino>=2024.1.0\" Pillow opencv-python tqdm \"gradio>=4.19\" safetensors --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"torch>=2.1\" \"diffusers>=0.25\" \"peft>=0.6.2\" \"transformers\" \"openvino>=2024.1.0\" Pillow opencv-python tqdm \"gradio>=4.19\" safetensors --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q datasets \"nncf>=2.10.0\"" ] }, diff --git a/notebooks/whisper-subtitles-generation/whisper-subtitles-generation.ipynb b/notebooks/whisper-subtitles-generation/whisper-subtitles-generation.ipynb index f0cee5aa9dd..b1f5737cd89 100644 --- a/notebooks/whisper-subtitles-generation/whisper-subtitles-generation.ipynb +++ b/notebooks/whisper-subtitles-generation/whisper-subtitles-generation.ipynb @@ -87,7 +87,7 @@ "outputs": [], "source": [ "%pip install -q \"openvino>=2024.1.0\" \"nncf>=2.10.0\"\n", - "%pip install -q \"python-ffmpeg<=1.0.16\" moviepy transformers onnx \"git+https://github.com/huggingface/optimum-intel.git\" \"peft==0.6.2\" \"torch>=2.1,<2.4\" \"torchvision<0.19.0\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"python-ffmpeg<=1.0.16\" moviepy \"onnx!=1.16.2\" \"git+https://github.com/huggingface/optimum-intel.git\" \"torch>=2.1\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q \"yt_dlp>=2024.8.6\" soundfile librosa jiwer\n", "%pip install -q \"gradio>=4.19\"" ] diff --git a/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb b/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb index 81abdab5e05..ca3832b659c 100644 --- a/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb +++ b/notebooks/wuerstchen-image-generation/wuerstchen-image-generation.ipynb @@ -71,7 +71,7 @@ "else:\n", " %pip install -q \"matplotlib>=3.4,<3.7\"\n", "\n", - "%pip install -q \"diffusers>=0.21.0\" \"torch>=2.1,<2.4\" \"torchvision<0.19.0\" transformers accelerate \"gradio>=4.19\" \"openvino>=2023.2.0\" \"peft==0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", + "%pip install -q \"diffusers>=0.24.0\" \"torch>=2.1\" \"torchvision\" transformers accelerate \"gradio>=4.19\" \"openvino>=2023.2.0\" \"peft>=0.6.2\" --extra-index-url https://download.pytorch.org/whl/cpu\n", "%pip install -q datasets \"nncf>=2.7.0\"" ] },