|
91 | 91 | " \"--extra-index-url\",\n",
|
92 | 92 | " \"https://download.pytorch.org/whl/cpu\",\n",
|
93 | 93 | " \"llama-index\",\n",
|
94 |
| - " \"llama-index-llms-huggingface==0.3.3\", # pin to keep compatibility due to https://github.com/run-llama/llama_index/commit/f037de8d0471b37f9c4069ebef5dfb329633d2c6\n", |
95 | 94 | " \"llama-index-readers-file\",\n",
|
96 | 95 | " \"llama-index-core\",\n",
|
97 |
| - " \"llama-index-llms-huggingface\",\n", |
98 |
| - " \"llama-index-embeddings-huggingface\",\n", |
99 | 96 | " \"transformers>=4.43.1\",\n",
|
100 |
| - " \"llama-index-llms-huggingface>=0.3.0,<0.3.4\",\n", |
101 |
| - " \"llama-index-embeddings-huggingface>=0.3.0\",\n", |
| 97 | + " \"llama-index-llms-huggingface>=0.4.0,<0.5.0\",\n", |
| 98 | + " \"llama-index-embeddings-huggingface>=0.4.0,<0.5.0\",\n", |
| 99 | + " \"huggingface-hub>=0.26.5\",\n", |
102 | 100 | ")\n",
|
103 | 101 | "pip_install(\"-q\", \"git+https://github.com/huggingface/optimum-intel.git\", \"git+https://github.com/openvinotoolkit/nncf.git\", \"datasets\", \"accelerate\")\n",
|
104 | 102 | "pip_install(\"--pre\", \"-Uq\", \"openvino>=2024.2.0\", \"--extra-index-url\", \"https://storage.openvinotoolkit.org/simple/wheels/nightly\")\n",
|
|
0 commit comments