diff --git a/inference-engine/thirdparty/clDNN/src/impls/ocl/reduce.cpp b/inference-engine/thirdparty/clDNN/src/impls/ocl/reduce.cpp index ef412f5dc345b3..e63cf6dd315e70 100644 --- a/inference-engine/thirdparty/clDNN/src/impls/ocl/reduce.cpp +++ b/inference-engine/thirdparty/clDNN/src/impls/ocl/reduce.cpp @@ -101,6 +101,11 @@ attach_reduce_impl::attach_reduce_impl() { std::make_tuple(data_types::i32, format::b_fs_yx_fsv16), std::make_tuple(data_types::i8, format::b_fs_yx_fsv16), std::make_tuple(data_types::u8, format::b_fs_yx_fsv16), + std::make_tuple(data_types::f32, format::b_fs_yx_fsv32), + std::make_tuple(data_types::f16, format::b_fs_yx_fsv32), + std::make_tuple(data_types::i32, format::b_fs_yx_fsv32), + std::make_tuple(data_types::i8, format::b_fs_yx_fsv32), + std::make_tuple(data_types::u8, format::b_fs_yx_fsv32), }); } diff --git a/inference-engine/thirdparty/clDNN/src/layout_optimizer.cpp b/inference-engine/thirdparty/clDNN/src/layout_optimizer.cpp index 5b9f0fb311f275..65f6aa42112fa4 100644 --- a/inference-engine/thirdparty/clDNN/src/layout_optimizer.cpp +++ b/inference-engine/thirdparty/clDNN/src/layout_optimizer.cpp @@ -866,10 +866,8 @@ layout layout_optimizer::get_expected_layout(layout const& current_layout, } } else if ((_optimization_attributes.b_fs_yx_fsv16_network && convolution_b_fs_yx_fsv16_opt(input_layout, output_layout, weights_layout, prim)) && is_2d) { - if (is_dw) - expected_format = cldnn::format::b_fs_yx_fsv32; - else - expected_format = cldnn::format::b_fs_yx_fsv16; + // TODO: optimize clDNN kernels for good support of b_fs_yx_fsv32 format + expected_format = cldnn::format::b_fs_yx_fsv32; } else { expected_format = imad_case(node); } diff --git a/thirdparty/onednn_gpu b/thirdparty/onednn_gpu index 1ab3f50381ae85..75d978369d0c5b 160000 --- a/thirdparty/onednn_gpu +++ b/thirdparty/onednn_gpu @@ -1 +1 @@ -Subproject commit 1ab3f50381ae85ab5ab0a29d5565ebe72ce77030 +Subproject commit 75d978369d0c5be04ec36c3cea2e00a14da1ec83