-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Eager] Support div(scalar) in eager mode #42148
Conversation
你的PR提交成功,感谢你对开源项目的贡献! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
some comments
@@ -222,7 +224,7 @@ def __impl__(self, other_var): | |||
# so the calculation result here and the calculation result of numpy are | |||
# different after 6 decimal point. If necessary, we can also use float64 here. | |||
# torch's behavior here is consistent with ours | |||
if op_type == 'elementwise_div' and self.dtype in _supported_int_dtype_: | |||
if op_type in _supported_div_op_ and self.dtype in _supported_int_dtype_: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this May cause performance issue
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Done, Thx.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
* [Eager] Support div scalar in eager mode * Updated and remove debug logs * Remove list, use 'or' directly * Remove useless statement
* bind elementwise_mod_op_xpu *test=kunlun * add more supported dtypes and UTs *test=kunlun * fix datatype error * add op to in xpu1_op_list * Update Mac cmake version >=3.15 (#41456) * Update Mac cmake version >=3.15 * notest;read test1 notest;read test2 notest;read test3 * fix inference link error * fix inference link error * fix windows link error * fix cmake_policy * fix build big size * Add paddle::variant and replace paddle::any (#42139) * add variant and replace any * split attribute * disable unittest failed in eager CI in temporary (#42101) * test=py3-eager * test=py3-eager * test=py3-eager * combine graph_table and feature_table in graph_engine (#42134) * extract sub-graph * graph-engine merging * fix * fix * fix heter-ps config * test performance * test performance * test performance * test * test * update bfs * change cmake * test * test gpu speed * gpu_graph_engine optimization * add dsm sample method * add graph_neighbor_sample_v2 * Add graph_neighbor_sample_v2 * fix for loop * add cpu sample interface * fix kernel judgement * add ssd layer to graph_engine * fix allocation * fix syntax error * fix syntax error * fix pscore class * fix * change index settings * recover test * recover test * fix spelling * recover * fix * move cudamemcpy after cuda stream sync * fix linking problem * remove comment * add cpu test * test * add cpu test * change comment * combine feature table and graph table * test * test * pybind * test * test * test * test * pybind * pybind * fix cmake * pybind * fix * fix * add pybind * add pybind Co-authored-by: DesmonDay <[email protected]> * [CustomDevice] add eager mode support (#42034) * fix FlattenContiguousRangeOpConverter out dim error (#42087) * fix FlattenContiguousRangeOpConverter out dim error * update code * fix python3.10 compile bug on windows (#42140) * Optimize dygraph GetExpectedKernelType perf (#42154) * opt dygraph scheduling * revert part impl * fix incorrect usages of std::move and other compile errors (#41045) * fix bug of std::move and others * fix an compile error in debug mode * fix wrong copy assignment operator Signed-off-by: tiancaishaonvjituizi <[email protected]> * reformat Signed-off-by: tiancaishaonvjituizi <[email protected]> * reformat Signed-off-by: tiancaishaonvjituizi <[email protected]> * fix ArrayRef constructor following llvm * fix format * fix conflict with master * fix variant compile error (#42203) * [Eager] Support numpy.ndarry in CastNumpy2Scalar (#42136) * [Eager] Remove redundancy code, fix fp16 case (#42169) * [Eager] Support div(scalar) in eager mode (#42148) * [Eager] Support div scalar in eager mode * Updated and remove debug logs * Remove list, use 'or' directly * Remove useless statement * fix recompute (#42128) * fix recompute * modify return * add LICENSE in wheel dist-info package (#42187) * replace any by variant in infermeta (#42181) * 【PaddlePaddle Hackathon 2】24、为 Paddle 新增 nn.ChannelShuffle 组网 API (#40743) * Add infermeta for ChannelShuffle * Create channel_shuffle_grad_kernel.h * Create channel_shuffle_kernel.h * Create channel_shuffle_sig.cc * Create channel_shuffle_op.cc ChannelShuffle算子的描述 * Create channel_shuffle_kernel_impl.h ChannelShuffle核函数的实现 * Create channel_shuffle_grad_kernel_impl.h ChannelShuffle反向核函数的实现 * Add kernel register of channel shuffle and grad 注册ChannelShuffle及其反向的核函数 * add nn.functional.channel_shuffle * add nn.ChannelShuffle * Create test_channel_shuffle.py * Update example of ChannelShuffle in vision.py * Update test_channel_shuffle.py * 修改channel_shuffle核函数的实现位置 * 修正代码格式 * 删除多余空格 * 完善channel_shuffle的错误检查 * Update unary.cc * Update channel_shuffle_op.cc * Update test_channel_shuffle.py * Update unary.cc * add channel_shuffle * Update test_channel_shuffle.py * Update vision.py * 调整代码格式 * Update channel_shuffle_sig.cc * 更新ChannelShuffle的文档 * 更新channel_shuffle的文档 * remove ChannelShuffleOpArgumentMapping * add ChannelShuffleGradInferMeta * Update channel_shuffle_op.cc * 调整channel_shuffle及其梯度的核函数的位置 * Do not reset default stream for StreamSafeCUDAAllocator (#42149) * remove redundant computation in Categorical.probs (#42114) * Downloading data for test_analyzer_vit_ocr (#42041) * Change server URL * update config * add test to parallel UT rule * add checksum to ensure files are downloaded * change downloading target * reuse existing variable * change target directory * fix en docs of some Apis (gradients, scope_guard, cuda_places, name_scope, device_guard, load_program_state, scale, ParamAttr and WeightNormParamAttr) (#41604) * Update scope_guard; test=document_fix * gradients; test=document_fix * gradients; test=document_fix * name_scope; test=document_fix * cpu_places; test=document_fix * WeightNormParamAttr; test=document_fix * cuda_places; test=document_fix * load_program_state; test=document_fix * device_guard; test=document_fix * device_guard; test=document_fix * ParamAttr; test=document_fix * scale; test=document_fix * scale; test=document_fix * update code example;test=document_fix Co-authored-by: Chen Long <[email protected]> * fix datatype error add op to in xpu1_op_list *test=kunlun * fix elementwise_mod op path error *test=kunlun * fix elementwise_mod UT error *test=kunlun * fix datatype error add op to in xpu1_op_list *test=kunlun add op to in xpu1_op_list fix elementwise_mod op path error *test=kunlun fix elementwise_mod UT error *test=kunlun Co-authored-by: tianshuo78520a <[email protected]> Co-authored-by: Chen Weihang <[email protected]> Co-authored-by: pangyoki <[email protected]> Co-authored-by: seemingwang <[email protected]> Co-authored-by: DesmonDay <[email protected]> Co-authored-by: ronnywang <[email protected]> Co-authored-by: baoachun <[email protected]> Co-authored-by: Zhou Wei <[email protected]> Co-authored-by: tiancaishaonvjituizi <[email protected]> Co-authored-by: Weilong Wu <[email protected]> Co-authored-by: Roc <[email protected]> Co-authored-by: BrilliantYuKaimin <[email protected]> Co-authored-by: Ruibiao Chen <[email protected]> Co-authored-by: Feiyu Chan <[email protected]> Co-authored-by: Sławomir Siwek <[email protected]> Co-authored-by: Yilingyelu <[email protected]> Co-authored-by: Chen Long <[email protected]>
PR types
Bug fixes
PR changes
Others
Describe
Support test_tensor_scalar_type_promotion_dynamic.py switch to eager mode.