-
-
Notifications
You must be signed in to change notification settings - Fork 5.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bugfix] Fix dtype mismatch in PaliGemma #6367
Conversation
@@ -111,7 +111,7 @@ def input_processor_for_paligemma(ctx: InputContext, llm_inputs: LLMInputs): | |||
orig_prompt = llm_inputs.get("prompt") | |||
orig_prompt_ids = llm_inputs.get("prompt_token_ids") | |||
|
|||
if image_token_str in orig_prompt: | |||
if orig_prompt is not None and image_token_str in orig_prompt: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My IDE flagged this so I fixed it along the way.
Hmm that's weird - I thought the vision encoder and language model itself is always https://huggingface.co/google/paligemma-3b-mix-224/blob/main/config.json |
Doesn't vLLM update the |
I wonder why we never needed to do this for Llava models - let me test the original issue and this PR and see if this fixes it. I do think we should test this model on |
Signed-off-by: Alvant <[email protected]>
FIX #6366