-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[misc] improve cloudpickle registration and tests #10202
[misc] improve cloudpickle registration and tests #10202
Conversation
Signed-off-by: youkaichao <[email protected]>
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
Signed-off-by: youkaichao <[email protected]>
Signed-off-by: youkaichao <[email protected]>
Signed-off-by: youkaichao <[email protected]>
Signed-off-by: youkaichao <[email protected]>
@@ -161,9 +171,8 @@ def iter_params(self, model_name: str): | |||
"facebook/opt-iml-max-1.3b": PPTestSettings.fast(), | |||
"OrionStarAI/Orion-14B-Chat": PPTestSettings.fast(trust_remote_code=True), | |||
"microsoft/phi-2": PPTestSettings.fast(), | |||
"microsoft/Phi-3-mini-4k-instruct": PPTestSettings.detailed(trust_remote_code=True, multi_node_only=True), # noqa: E501 | |||
"microsoft/Phi-3.5-MoE-instruct": PPTestSettings.detailed(trust_remote_code=True, multi_node_only=True, load_format="dummy", hf_overrides='{"num_hidden_layers": 4, "hidden_size": 512, "intermediate_size": 800, "num_attention_heads": 4, "num_key_value_heads": 1}'), # noqa: E501 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
test a model with a point in its name, previously this model is not tested in the ci.
@@ -161,9 +171,8 @@ def iter_params(self, model_name: str): | |||
"facebook/opt-iml-max-1.3b": PPTestSettings.fast(), | |||
"OrionStarAI/Orion-14B-Chat": PPTestSettings.fast(trust_remote_code=True), | |||
"microsoft/phi-2": PPTestSettings.fast(), | |||
"microsoft/Phi-3-mini-4k-instruct": PPTestSettings.detailed(trust_remote_code=True, multi_node_only=True), # noqa: E501 | |||
"microsoft/Phi-3.5-MoE-instruct": PPTestSettings.detailed(trust_remote_code=True, multi_node_only=True, load_format="dummy", hf_overrides='{"num_hidden_layers": 4, "hidden_size": 512, "intermediate_size": 800, "num_attention_heads": 4, "num_key_value_heads": 1}'), # noqa: E501 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks to @DarkLight1337 adding the hf_overrides
recently.
Thanks @youkaichao |
Signed-off-by: youkaichao <[email protected]> Signed-off-by: Jee Jee Li <[email protected]>
Signed-off-by: youkaichao <[email protected]>
Signed-off-by: youkaichao <[email protected]> Signed-off-by: OmerD <[email protected]>
Signed-off-by: youkaichao <[email protected]> Signed-off-by: Sumit Dubey <[email protected]>
improves over #6751
config is the only custom class we pass across processes, and we can move the function call into
get_config
, where the first import oftransformers_modules
happen.I use
cloudpickle
to serialize the wholeVllmConfig
rather than the singleModelConfig
, so that even if people movehf_config
outside ofModelConfig
, the code can still work.Also added some example code for when
transformers_modules
will be available.