-
Notifications
You must be signed in to change notification settings - Fork 511
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove np.bool
which is not supported in latest np versions
#1558
Conversation
np.bool
which is not supported in latest np versions
Our requirement is numpy<=1.23. |
I tried with |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* [Improvement] max_batches support to training log and tqdm progress bar. (#1554) * Added max_batches support to training log and tqdm progress bar. * Added changing string in accordance which parameter is used (len(loader) of max_batches) * Replaced stopping condition for the epoch with a smaller one (cherry picked from commit 749a9c7) * fix (#1558) Co-authored-by: Eugene Khvedchenya <[email protected]> (cherry picked from commit 8a1d255) * fix (#1564) (cherry picked from commit 24798b0) * Bugfix of model.export() to work correct with bs>1 (#1551) (cherry picked from commit 0515496) * Fixed incorrect automatic variable used (#1565) $@ is the name of the target being generated, and $^ are the dependencies Co-authored-by: Louis-Dupont <[email protected]> (cherry picked from commit 43f8bea) * fix typo in class documentation (#1548) Co-authored-by: Eugene Khvedchenya <[email protected]> Co-authored-by: Louis-Dupont <[email protected]> (cherry picked from commit ec21383) * Feature/sg 1198 mixed precision automatically changed with warning (#1567) * fix * work with tmpdir * minor change of comment * improve device_config (cherry picked from commit 34fda6c) * Fixed issue with torch 1.12 where _scale_fn_ref is missing in CyclicLR (#1575) (cherry picked from commit 23b4f7a) * Fixed issue with torch 1.12 issue with arange not supporting fp16 for CPU device. (#1574) (cherry picked from commit 1f15c76) --------- Co-authored-by: hakuryuu96 <[email protected]> Co-authored-by: Louis-Dupont <[email protected]> Co-authored-by: Alessandro Ros <[email protected]>
Issue
Exception when using SG with
numpy==1.25.2
Proposed solution
(torch.bool, np.bool),
->(torch.bool, bool),