-
Notifications
You must be signed in to change notification settings - Fork 332
Testing #3087
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Testing #3087
Changes from 14 commits
786b2e5
cc20fca
0d67b9e
bf603eb
c16b50b
aab75bd
948819a
e75034a
6848bed
5e49ddf
cbdaf47
bde1ede
fd1d938
138d9d5
5ee79e7
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
@@ -472,7 +472,7 @@ jobs: | |||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).tokenizers.test }} | ||||||||||||
| timeout: 60 | ||||||||||||
| - name: 'API tests' | ||||||||||||
| cmd: 'python -m pytest -v ./tests/python_tests/test_continuous_batching.py ./tests/python_tests/test_generation_config.py ./tests/python_tests/test_sampling.py ./tests/python_tests/test_text_streamer.py' | ||||||||||||
| cmd: 'python -m pytest -v ./tests/python_tests/test_continuous_batching.py -k "not eagle3" ./tests/python_tests/test_generation_config.py ./tests/python_tests/test_sampling.py ./tests/python_tests/test_text_streamer.py' | ||||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).continuous_batching.test || fromJSON(needs.smart_ci.outputs.affected_components).sampling.test || fromJSON(needs.smart_ci.outputs.affected_components).text_streamer.test }} | ||||||||||||
| timeout: 60 | ||||||||||||
| - name: 'Rag tests' | ||||||||||||
|
|
@@ -489,6 +489,12 @@ jobs: | |||||||||||
| python -m pytest -v ./tools/who_what_benchmark/tests -m nanollava | ||||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).WWB.test }} | ||||||||||||
| timeout: 90 | ||||||||||||
| - name: 'EAGLE3 speculative decoding tests' | ||||||||||||
| cmd: | | ||||||||||||
|
||||||||||||
| cmd: | | |
| cmd: | | |
| # FIXME: Using a personal fork of optimum-intel for EAGLE3 speculative decoding tests. | |
| # Reason: Required changes are not yet merged upstream. See https://github.com/huggingface/optimum-intel/pull/<PR_NUMBER> (replace with actual PR/issue link). | |
| # Remove this and use official optimum-intel release once changes are merged. |
| Original file line number | Diff line number | Diff line change | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
@@ -623,7 +623,7 @@ jobs: | |||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).tokenizers.test }} | ||||||||||||
| timeout: 60 | ||||||||||||
| - name: 'API tests' | ||||||||||||
| cmd: 'python -m pytest -s -v tests/python_tests/test_continuous_batching.py tests/python_tests/test_generation_config.py tests/python_tests/test_sampling.py tests/python_tests/test_text_streamer.py' | ||||||||||||
| cmd: 'python -m pytest -s -v tests/python_tests/test_continuous_batching.py -k "not eagle3" tests/python_tests/test_generation_config.py tests/python_tests/test_sampling.py tests/python_tests/test_text_streamer.py' | ||||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).continuous_batching.test || fromJSON(needs.smart_ci.outputs.affected_components).sampling.test || fromJSON(needs.smart_ci.outputs.affected_components).text_streamer.test }} | ||||||||||||
| timeout: 60 | ||||||||||||
| - name: 'Rag tests' | ||||||||||||
|
|
@@ -640,6 +640,12 @@ jobs: | |||||||||||
| python -m pytest -v ./tools/who_what_benchmark/tests -m nanollava | ||||||||||||
| run_condition: ${{ fromJSON(needs.smart_ci.outputs.affected_components).WWB.test }} | ||||||||||||
| timeout: 90 | ||||||||||||
| - name: 'EAGLE3 speculative decoding tests' | ||||||||||||
| cmd: | | ||||||||||||
|
||||||||||||
| cmd: | | |
| cmd: | | |
| # TODO: Using a personal fork of optimum-intel for EAGLE3 speculative decoding tests. | |
| # Reason: [Add explanation here, e.g., "Required for feature X not yet merged upstream. See PR #123."] | |
| # Remove this and use official optimum-intel release when upstream PR is merged. |
| Original file line number | Diff line number | Diff line change | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
|
|
@@ -35,6 +35,7 @@ def read_images(path: str) -> list[Tensor]: | |||||||||||
| def main(): | ||||||||||||
| parser = argparse.ArgumentParser(description="Help command") | ||||||||||||
| parser.add_argument("-m", "--model", type=str, help="Path to model and tokenizers base directory") | ||||||||||||
| parser.add_argument("-dm", "--draft_model", type=str, help="Path to draft model and tokenizers base directory") | ||||||||||||
| parser.add_argument("-p", "--prompt", type=str, default=None, help="Prompt") | ||||||||||||
| parser.add_argument("-pf", "--prompt_file", type=str, help="Read prompt from file") | ||||||||||||
| parser.add_argument("-i", "--image", type=str, default="image.jpg", help="Image") | ||||||||||||
|
|
@@ -61,6 +62,7 @@ def main(): | |||||||||||
| # Perf metrics is stored in VLMDecodedResults. | ||||||||||||
| # In order to get VLMDecodedResults instead of a string input should be a list. | ||||||||||||
| models_path = args.model | ||||||||||||
| draft_model_path = args.draft_model | ||||||||||||
| images = read_images(args.image) | ||||||||||||
| device = args.device | ||||||||||||
| num_warmup = args.num_warmup | ||||||||||||
|
|
@@ -76,7 +78,13 @@ def main(): | |||||||||||
| scheduler_config = ov_genai.SchedulerConfig() | ||||||||||||
| scheduler_config.enable_prefix_caching = False | ||||||||||||
| scheduler_config.max_num_batched_tokens = sys.maxsize | ||||||||||||
| pipe = ov_genai.VLMPipeline(models_path, device, scheduler_config=scheduler_config) | ||||||||||||
|
|
||||||||||||
| print("draft_model_path=", draft_model_path) | ||||||||||||
| print("device=", device) | ||||||||||||
| draft_model = ov_genai.draft_model(str(draft_model_path), device) | ||||||||||||
| #pipe = ov_genai.VLMPipeline(models_path, device, scheduler_config=scheduler_config) | ||||||||||||
|
Comment on lines
+82
to
+85
|
||||||||||||
| print("draft_model_path=", draft_model_path) | |
| print("device=", device) | |
| draft_model = ov_genai.draft_model(str(draft_model_path), device) | |
| #pipe = ov_genai.VLMPipeline(models_path, device, scheduler_config=scheduler_config) | |
| draft_model = ov_genai.draft_model(str(draft_model_path), device) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Installing from a personal GitHub fork using a specific commit hash is fragile and not maintainable. Consider either: 1) merging these changes into the official optimum-intel repository and using a tagged release, or 2) documenting why this fork is necessary and when it can be removed.