|
70 | 70 | "Uninstalling openvino-2023.3.0:\n", |
71 | 71 | " Successfully uninstalled openvino-2023.3.0\n", |
72 | 72 | "\u001B[33mWARNING: Skipping openvino-nightly as it is not installed.\u001B[0m\u001B[33m\n", |
73 | | - "\u001B[0mFound existing installation: openvino-dev 2023.3.0\n", |
74 | | - "Uninstalling openvino-dev-2023.3.0:\n", |
75 | | - " Successfully uninstalled openvino-dev-2023.3.0\n", |
76 | | - "Note: you may need to restart the kernel to use updated packages.\n", |
| 73 | + "\u001B[0m\u001B[33mWARNING: Skipping openvino-dev as it is not installed.\u001B[0m\u001B[33m\n", |
| 74 | + "\u001B[0mNote: you may need to restart the kernel to use updated packages.\n", |
77 | 75 | "Note: you may need to restart the kernel to use updated packages.\n" |
78 | 76 | ] |
79 | 77 | } |
|
271 | 269 | "name": "stderr", |
272 | 270 | "output_type": "stream", |
273 | 271 | "text": [ |
274 | | - "2024-02-16 17:13:59.306808: I tensorflow/core/util/port.cc:110] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.\n", |
275 | | - "2024-02-16 17:13:59.308436: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.\n", |
276 | | - "2024-02-16 17:13:59.341678: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.\n", |
277 | | - "2024-02-16 17:13:59.342886: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\n", |
| 272 | + "2024-02-20 14:05:25.856016: I tensorflow/core/util/port.cc:110] oneDNN custom operations are on. You may see slightly different numerical results due to floating-point round-off errors from different computation orders. To turn them off, set the environment variable `TF_ENABLE_ONEDNN_OPTS=0`.\n", |
| 273 | + "2024-02-20 14:05:25.857714: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.\n", |
| 274 | + "2024-02-20 14:05:25.892124: I tensorflow/tsl/cuda/cudart_stub.cc:28] Could not find cuda drivers on your machine, GPU will not be used.\n", |
| 275 | + "2024-02-20 14:05:25.893158: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\n", |
278 | 276 | "To enable the following instructions: AVX2 AVX512F AVX512_VNNI FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n", |
279 | | - "2024-02-16 17:14:00.000099: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n" |
| 277 | + "2024-02-20 14:05:26.599940: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n" |
280 | 278 | ] |
281 | 279 | }, |
282 | 280 | { |
|
319 | 317 | "\n", |
320 | 318 | "if not model_dir.exists():\n", |
321 | 319 | " # converting the original model\n", |
322 | | - " # %pip install -U \"git+https://github.com/huggingface/optimum-intel.git\" \"nncf>=2.8.0\"a\n", |
| 320 | + " # %pip install -U \"git+https://github.com/huggingface/optimum-intel.git\" \"nncf>=2.8.0\" onnx\n", |
323 | 321 | " # %optimum-cli export openvino -m $model_id --task text-generation-with-past $model_dir\n", |
324 | 322 | " \n", |
325 | 323 | " # load already converted model\n", |
|
372 | 370 | { |
373 | 371 | "data": { |
374 | 372 | "application/vnd.jupyter.widget-view+json": { |
375 | | - "model_id": "c20e21457ddd4de4bbcd95f8bc5c7619", |
| 373 | + "model_id": "61e8811e77f34e7aaa277847730dcd50", |
376 | 374 | "version_major": 2, |
377 | 375 | "version_minor": 0 |
378 | 376 | }, |
|
467 | 465 | "model_dir = Path(Path(model_id).name)\n", |
468 | 466 | "\n", |
469 | 467 | "if not model_dir.exists():\n", |
470 | | - " %pip install -qU git+https://github.com/huggingface/optimum-intel.git\n", |
| 468 | + " %pip install -qU git+https://github.com/huggingface/optimum-intel.git onnx\n", |
471 | 469 | " !optimum-cli export openvino --model $model_id --task text-classification $model_dir\n", |
472 | 470 | " !convert_tokenizer $model_id -o $model_dir" |
473 | 471 | ] |
|
529 | 527 | "\n", |
530 | 528 | "- [Installation instructions for different environments](https://github.com/openvinotoolkit/openvino_tokenizers?tab=readme-ov-file#installation)\n", |
531 | 529 | "- [Supported Tokenizer Types](https://github.com/openvinotoolkit/openvino_tokenizers?tab=readme-ov-file#supported-tokenizer-types)\n", |
532 | | - "- [OpenVINO.genAI repository with the C++ example of OpenVINO Tokenizers usage]((https://github.com/openvinotoolkit/openvino.genai/tree/master/text_generation/causal_lm/cpp))\n", |
| 530 | + "- [OpenVINO.GenAI repository with the C++ example of OpenVINO Tokenizers usage](https://github.com/openvinotoolkit/openvino.genai/tree/master/text_generation/causal_lm/cpp)\n", |
533 | 531 | "- [HuggingFace Tokenizers Comparison Table](https://github.com/openvinotoolkit/openvino_tokenizers?tab=readme-ov-file#output-match-by-model)" |
534 | 532 | ] |
535 | 533 | } |
|
0 commit comments