-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: CI - Split up "Models Test" and "Vision Language Models Test" #7439
Comments
root for this! |
This sounds great! However, we should still avoid regressions. I suggest testing a representative subset per PR and perform full testing once in a while. Any ideas on which models we should always test? |
A more concrete plan: Let's add a pytest marker Separately, we can organize the models tests as follows:
The first three are currently included in Models test while the fourth one is currently in VLM test. They are split apart further so that model PRs can easily run only the tests that are relevant to their model. (By the way, the models directory in main vLLM is pretty cluttered... perhaps we can also organize the model files as above?) In the normal CI, we can add the pytest flag Regarding which model tests to mark as core_model:
|
Hmm, it sounds unreliable to ask users to run specified model tests... in pytorch for instance it's mandatory to run quite comprehensive test suite in every PR I think that changing this behaviour will potentially introduce uncaught bugs.
I think that first step of splitting up like this is a good idea. |
After some offline discussion, we have decided to cut down on the number of models to regularly test in CI. This is due to the following reasons:
For encoder-decoder and embedding models, we will preserve all tests since there are only a few tests involved. For multimodal models, @ywang96 and I have decided to only test the following models regularly in CI:
Meanwhile, @simon-mo will help narrow down the list of decoder-only language models to test. Note that this does not remove the need to test new models. New model PRs should still implement tests so we can verify their implementation, as well as re-run those tests whenever a future PR updates related code. Until @khluu figures out how to create a separate test item for each model without clogging up the web UI, these tests will remain excluded from CI; instead, we can run these tests locally as necessary. |
🚀 The feature, motivation and pitch
Takes 1 hour+ on CI compared to others, which take <~30 min. Thus, ends up being a bottleneck
So, should be split up similar to kernels
CC: @khluu
The text was updated successfully, but these errors were encountered: