-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[CI/Build] Avoid downloading all HF files in RemoteOpenAIServer
#7836
[CI/Build] Avoid downloading all HF files in RemoteOpenAIServer
#7836
Conversation
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, please make sure to run full CI as it is required to merge (or just use auto-merge). To run full CI, you can do one of these:
🚀 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice fix
tests/utils.py
Outdated
self.host = str(args.host or 'localhost') | ||
self.port = int(args.port) | ||
|
||
# download the model before starting the server to avoid timeout | ||
engine_args = AsyncEngineArgs.from_cli_args(args) | ||
engine_config = engine_args.create_engine_config() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
can we directly create a load config object? it should be simple in my opinion. just load format auto.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's fix the tests first before simplifying this.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm not that familiar with the model loading code so may need some help regarding fixing the tests.
tests/utils.py
Outdated
@@ -60,36 +61,40 @@ class RemoteOpenAIServer: | |||
|
|||
def __init__(self, | |||
model: str, | |||
cli_args: List[str], | |||
serve_args: List[str], |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why rename the arg name btw?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I meant to associate them with vllm serve
specifically, not just any CLI. Perhaps it could be clearer.
Edit: Updated the name to vllm_serve_args
|
||
parser = FlexibleArgumentParser( | ||
description="vLLM's remote OpenAI server.") | ||
parser = make_arg_parser(parser) | ||
args = parser.parse_args(cli_args) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think you just need args = parser.parse_args(["--model", model, *cli_args])
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks for the fix!
RemoteOpenAIServer
…lm-project#7836) Signed-off-by: Alvant <[email protected]>
Fixes #7834 (comment)