Skip to content

Commit

Permalink
update
Browse files Browse the repository at this point in the history
  • Loading branch information
mkrze committed Mar 4, 2025
1 parent 9ce7415 commit 41361e1
Showing 1 changed file with 10 additions and 9 deletions.
19 changes: 10 additions & 9 deletions .jenkins/lm-eval-harness/test_lm_eval_correctness.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@

LORA_ADAPTER_PATH = os.environ.get("LORA_ADAPTER_PATH", None)

INFERENCE_SERVING = os.environ.get("INFERENCE_SERVING", False)
INFERENCE_SERVING = os.environ.get("INFERENCE_SERVING", "false") in ['1', 'true']

NUM_CONCURRENT = os.environ.get("NUM_CONCURRENT", 128)

Expand Down Expand Up @@ -91,14 +91,6 @@ def launch_lm_eval(eval_config):
kwargs['fewshot_as_multiturn'] = eval_config['fewshot_as_multiturn']
if 'apply_chat_template' in eval_config:
kwargs['apply_chat_template'] = eval_config['apply_chat_template']
results = lm_eval.simple_evaluate(
model="vllm",
model_args=model_args,
tasks=[task["name"] for task in eval_config["tasks"]],
num_fewshot=eval_config["num_fewshot"],
limit=eval_config["limit"],
batch_size="auto",
**kwargs)

if INFERENCE_SERVING:
model_args = \
Expand All @@ -112,6 +104,15 @@ def launch_lm_eval(eval_config):
model_args=model_args,
tasks=[task["name"] for task in eval_config["tasks"]]
)
else:
results = lm_eval.simple_evaluate(
model="vllm",
model_args=model_args,
tasks=[task["name"] for task in eval_config["tasks"]],
num_fewshot=eval_config["num_fewshot"],
limit=eval_config["limit"],
batch_size="auto",
**kwargs)

return results

Expand Down

0 comments on commit 41361e1

Please sign in to comment.