You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[rank0]: Traceback (most recent call last):
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/runpy.py", line 196, in _run_module_as_main
[rank0]: return _run_code(code, main_globals, None,
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/runpy.py", line 86, in _run_code
[rank0]: exec(code, run_globals)
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/site-packages/vllm/entrypoints/openai/run_batch.py", line 146, in
[rank0]: asyncio.run(main(args))
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/asyncio/runners.py", line 44, in run
[rank0]: return loop.run_until_complete(main)
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/asyncio/base_events.py", line 649, in run_until_complete
[rank0]: return future.result()
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/site-packages/vllm/entrypoints/openai/run_batch.py", line 130, in main
[rank0]: responses = await asyncio.gather(*response_futures)
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/site-packages/vllm/entrypoints/openai/run_batch.py", line 93, in run_request
[rank0]: response=BatchResponseData(
[rank0]: File "/home/arsal/anaconda3/envs/vllm/lib/python3.10/site-packages/pydantic/main.py", line 193, in init
[rank0]: self.pydantic_validator.validate_python(data, self_instance=self)
[rank0]: pydantic_core._pydantic_core.ValidationError: 1 validation error for BatchResponseData
[rank0]: body
[rank0]: Field required [type=missing, input_value={'status_code': 400, 'req...5c4a7799ad445a114ebff4'}, input_type=dict]
[rank0]: For further information visit https://errors.pydantic.dev/2.8/v/missing
The text was updated successfully, but these errors were encountered:
Your current environment
🐛 Describe the bug
Running this command to do batch inference through API, returns the following error. The input.jsonl is as per required format.
python -m vllm.entrypoints.openai.run_batch -i input.jsonl -o results.jsonl --model Granther/Gemma-2-9B-Instruct-4Bit-GPTQ --max_model_len 3000
Error Traceback:
The text was updated successfully, but these errors were encountered: