-
Notifications
You must be signed in to change notification settings - Fork 26.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix Cohere CI #31263
Fix Cohere CI #31263
Conversation
@unittest.skip("foo") | ||
def test_initialization(self): | ||
super().test_initialization() | ||
|
||
@unittest.skip("foo") | ||
def test_fast_init_context_manager(self): | ||
super().test_fast_init_context_manager() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I will open issue to keep this track, but let's follow what we have discussed earlier that we sometimes also need to skip some tests considering the amount of time we have.
"Hi there, here we are again with another great collection of free fonts for your next project. This time we have gathered 10 free fonts that you can download and use in your designs. These fonts are perfect for any kind", | ||
] | ||
|
||
model = CohereForCausalLM.from_pretrained(model_id) | ||
model = CohereForCausalLM.from_pretrained(model_id, device_map="auto") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This was previously GPU OOM. After changing to auto
, I have to update the expected output value too.
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
761b75a
to
06c81eb
Compare
# This gives CPU OOM on a single-gpu runner (~60G RAM). On multi-gpu runner, it has ~180G RAM which is enough. | ||
@require_torch_multi_gpu | ||
def test_torch_encode_plus_sent_to_model(self): | ||
super().test_torch_encode_plus_sent_to_model() |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
CPU OOM
@unittest.skip("Failing.") | ||
def test_initialization(self): | ||
super().test_initialization() | ||
|
||
@unittest.skip("Failing.") | ||
def test_fast_init_context_manager(self): | ||
super().test_fast_init_context_manager() | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I will open issue to track this, but as we discussed earlier, let's not spend too much time on some tests and focus on other priorities.
run: | | ||
export CUDA_VISIBLE_DEVICES="$(python3 utils/set_cuda_devices_for_ci.py --test_folder ${{ matrix.folders }})" | ||
echo $CUDA_VISIBLE_DEVICES | ||
python3 -m pytest -v -rsfE --make-reports=${{ matrix.machine_type }}_run_models_gpu_${{ matrix.folders }}_test_reports tests/${{ matrix.folders }} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
So far on workflow, we set CUDA_VISIBLE_DEVICES=0,1
as otherwise overall there are some more strange issues when running 4 GPUs.
However, for cohere test_eager_matches_sdpa_generate
, it requires more GPU memory. So I just tried to allow CUDA_VISIBLE_DEVICES=0,1,2,3
for this job.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
It requires a new script below (see at the end).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok, thanks Yih-Dar!
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
* [run-slow] cohere * [run-slow] cohere * [run-slow] cohere --------- Co-authored-by: ydshieh <[email protected]>
What does this PR do?
Currently some tests are failing, see
https://github.com/huggingface/transformers/actions/runs/9360271371/job/25765482477
This PR skips 2 of them and make 4 pass: