Thanks to visit codestin.com
Credit goes to github.com

Skip to content

have an issue when running run_eval.sh #2

@s2014628

Description

@s2014628

I have faced a problem when I run bash run_eval.sh.

The error has shown as follow:

Traceback (most recent call last):
File "/home/ft/anaconda3/envs/thinkless/bin/lm_eval", line 8, in
sys.exit(cli_evaluate())
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/main.py", line 389, in cli_evaluate
results = evaluator.simple_evaluate(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/utils.py", line 422, in _wrapper
return fn(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/evaluator.py", line 308, in simple_evaluate
results = evaluate(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/utils.py", line 422, in _wrapper
return fn(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/evaluator.py", line 528, in evaluate
resps = getattr(lm, reqtype)(cloned_reqs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/models/vllm_causallms.py", line 418, in generate_until
cont = self._model_generate(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/models/vllm_causallms.py", line 271, in _model_generate
results = ray.get(object_refs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/ray/_private/auto_init_hook.py", line 21, in auto_init_wrapper
return fn(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/ray/_private/client_mode_hook.py", line 103, in wrapper
return func(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/ray/_private/worker.py", line 2822, in get
values, debugger_breakpoint = worker.get_objects(object_refs, timeout=timeout)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/ray/_private/worker.py", line 930, in get_objects
raise value.as_instanceof_cause()
ray.exceptions.RayTaskError(ValueError): ray::run_inference_one_model() (pid=4185470, ip=10.72.74.13)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/lm_eval/models/vllm_causallms.py", line 257, in run_inference_one_model
return llm.generate(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/utils.py", line 1063, in inner
return fn(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/entrypoints/llm.py", line 353, in generate
outputs = self._run_engine(use_tqdm=use_tqdm)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/entrypoints/llm.py", line 879, in _run_engine
step_outputs = self.llm_engine.step()
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 1386, in step
outputs = self.model_executor.execute_model(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/executor/ray_gpu_executor.py", line 332, in execute_model
return super().execute_model(execute_model_req)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/executor/distributed_gpu_executor.py", line 82, in execute_model
driver_outputs = self._driver_execute_model(execute_model_req)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/executor/ray_gpu_executor.py", line 325, in _driver_execute_model
return self.driver_worker.execute_method("execute_model",
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/worker_base.py", line 465, in execute_method
raise e
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/worker_base.py", line 456, in execute_method
return executor(*args, **kwargs)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/worker_base.py", line 303, in execute_model
inputs = self.prepare_input(execute_model_req)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/worker_base.py", line 291, in prepare_input
return self._get_driver_input_and_broadcast(execute_model_req)
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/worker_base.py", line 253, in _get_driver_input_and_broadcast
self.model_runner.prepare_model_input(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/model_runner.py", line 1593, in prepare_model_input
model_input = self._prepare_model_input_tensors(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/model_runner.py", line 1200, in _prepare_model_input_tensors
return builder.build() # type: ignore
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/worker/model_runner.py", line 871, in build
attn_metadata = self.attn_metadata_builder.build(
File "/home/ft/anaconda3/envs/thinkless/lib/python3.10/site-packages/vllm/attention/backends/utils.py", line 223, in build
input_block_tables[i, :len(block_table)] = block_table
ValueError: could not broadcast input array from shape (513,) into shape (512,)
Processed prompts: 98%|█████████▊| 746/759 [08:37<01:57, 9.01s/it, est. speed input: 111.14 toks/s, output: 1294.80 toks/s] [repeated 3x across cluster]

It will cause the invalid results :

AIME 2024
Full Results: []
Traceback (most recent call last):
File "/home/ft/Thinkless/scripts/eval/eval_aime.py", line 137, in
print(f"Thinking Mode: {perc_long_list} ({100*sum(perc_long_list)/len(perc_long_list):.2f}%)")
ZeroDivisionError: division by zero

Minerva
Full Results: []
Traceback (most recent call last):
File "/home/ft/Thinkless/scripts/eval/eval_minerva.py", line 179, in
print(f"Thinking Mode: {perc_long_list} ({100*sum(perc_long_list)/len(perc_long_list):.2f}%)")
ZeroDivisionError: division by zero

Math-500
Full Results: []
Traceback (most recent call last):
File "/home/ft/Thinkless/scripts/eval/eval_math500.py", line 179, in
print(f"Thinking Mode: {perc_long_list} ({100*sum(perc_long_list)/len(perc_long_list):.2f}%)")
ZeroDivisionError: division by zero

GSM8k
Full Results: []
Traceback (most recent call last):
File "/home/ft/Thinkless/scripts/eval/eval_gsm8k.py", line 179, in
print(f"Thinking Mode: {perc_long_list} ({100*sum(perc_long_list)/len(perc_long_list):.2f}%)")
ZeroDivisionError: division by zero

Could you help with this? I appreciate it

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions