You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm wondering if the default value of length_penalty in the beam search algorithm should be set to 0.
My understanding, based on discussions like this issue in the hf transformers repository, is that a length_penalty greater than 0 favors shorter sequences, while a penalty less than 0 favors longer ones and 0 means no length_penalty
In the vLLM project (which has code adapted from Hugging Face), particularly in the file vllm/sequence.py, the default value is indeed 0:
The default length_penalty in get_beam_search_score was changed to 1.0 last week in #2667. So that it matches the one in SamplingParams.
This is the default because it's the default in transformers and the argument for this being the default in transformers is in huggingface/transformers#19006 (comment).
I'm wondering if the default value of
length_penalty
in the beam search algorithm should be set to 0.My understanding, based on discussions like this issue in the hf transformers repository, is that a
length_penalty
greater than 0 favors shorter sequences, while a penalty less than 0 favors longer ones and 0 means nolength_penalty
In the vLLM project (which has code adapted from Hugging Face), particularly in the file vllm/sequence.py, the default value is indeed 0:
Should we consider setting the default
length_penalty
to 0?The text was updated successfully, but these errors were encountered: