feat(tasks): add public model benchmark tags#1230
Merged
Luodian merged 11 commits intoEvolvingLMMs-Lab:mainfrom Mar 7, 2026
Merged
feat(tasks): add public model benchmark tags#1230Luodian merged 11 commits intoEvolvingLMMs-Lab:mainfrom
Luodian merged 11 commits intoEvolvingLMMs-Lab:mainfrom
Conversation
Sync TP ranks under external_launcher and keep max_new_tokens at least as large as the model-side setting so reasoning outputs are not task-capped.
Inject a padding request when a rank receives zero docs and align request/filter synchronization across ranks so TP+DP jobs with limit<=world_size no longer crash or hang.
fix(vllm): support TP+DP dispatch and model-side max_new_tokens precedence
…5-printer-compat fix: pin wandb 0.25 and support printer API rename
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
This PR adds curated
tagmetadata that maps runnablelmms_evalimage tasks to model families whose official public evaluation materials highlight those benchmarks.The goal is to make it possible to select a practical, vendor-aligned benchmark slice directly from YAML using existing
TaskManagertag expansion, without introducing any new evaluator logic.What changed
This PR adds four model-family tags:
public_eval_qwen3_5_familypublic_eval_seed2_familypublic_eval_gemini3_familypublic_eval_gpt5_familyIt applies those tags to the corresponding task YAMLs and adds documentation for the naming convention, scope, and maintenance rules in
docs/guides/model_benchmark_tags.md.Current tag expansion
public_eval_qwen3_5_familyai2d->lmms-lab/ai2d(test)charxiv_val_reasoning->princeton-nlp/CharXiv(validation)hallusion_bench_image->lmms-lab/HallusionBench(image)mathvista_testmini_cot->AI4Math/MathVista(testmini)mathvista_testmini_format->AI4Math/MathVista(testmini)mathvista_testmini_solution->AI4Math/MathVista(testmini)mmbench_en_dev->lmms-lab/MMBenchconfigen(dev)mmlongbench_doc->yubo2333/MMLongBench-Doc(train)mmmu_pro_standard->MMMU/MMMU_Proconfigstandard (10 options)(test)mmmu_pro_vision->MMMU/MMMU_Proconfigvision(test)mmmu_val->lmms-lab/MMMU(validation)mmstar->Lin-Chen/MMStar(val)ocrbench->echo840/OCRBench(test)omnidocbench->ouyanglinke/OmniDocBench_tsv(train)realworldqa->lmms-lab/RealWorldQA(test)public_eval_seed2_familycharxiv_val_descriptive->princeton-nlp/CharXiv(validation)charxiv_val_reasoning->princeton-nlp/CharXiv(validation)mathvision_testmini->MathLLMs/MathVision(testmini)mmlongbench_doc->yubo2333/MMLongBench-Doc(train)ocrbench_v2->ling99/OCRBench_v2(test)omnidocbench->ouyanglinke/OmniDocBench_tsv(train)public_eval_gemini3_familycharxiv_val_reasoning->princeton-nlp/CharXiv(validation)mmmu_pro_standard->MMMU/MMMU_Proconfigstandard (10 options)(test)mmmu_pro_vision->MMMU/MMMU_Proconfigvision(test)omnidocbench->ouyanglinke/OmniDocBench_tsv(train)public_eval_gpt5_familymmmu_val->lmms-lab/MMMU(validation)Why this helps
Today, the repo already supports
tagas a first-class task selector, but there is no built-in way to ask for "the public benchmark slice most associated with model family X".With these tags in place, users can run commands like:
This keeps the implementation simple, avoids adding hard-coded model-specific logic in Python, and makes future curation incremental at the YAML layer.
Validation
pre-commit run --all-filesTaskManagerindexes and expands all four new tags successfully