Skip to content

Commit

Permalink
Fixed comments
Browse files Browse the repository at this point in the history
  • Loading branch information
ilyachur committed Dec 18, 2020
1 parent 21850b5 commit 29304c9
Show file tree
Hide file tree
Showing 5 changed files with 16 additions and 15 deletions.
2 changes: 1 addition & 1 deletion ngraph/core/src/op/multiply.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ bool op::v0::Multiply::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(
op_v0_Multiply_evaluate,
Multiply_evaluate,
return multiplyop::evaluate_multiply(inputs[0], inputs[1], outputs[0], get_autob()));
return false;
}
Expand Down
10 changes: 5 additions & 5 deletions ngraph/core/src/op/reduce_l1.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -67,11 +67,11 @@ namespace reduce_l1
bool rc = true;
switch (arg->get_element_type())
{
NGRAPH_TYPE_CASE(evaluate_sum, i32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, i64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, bf16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, f16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, f32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reducel1_sum, i32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reducel1_sum, i64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reducel1_sum, bf16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reducel1_sum, f16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reducel1_sum, f32, arg, out, axes, keep_dims);
default: rc = false; break;
}
return rc;
Expand Down
12 changes: 6 additions & 6 deletions ngraph/core/src/op/reduce_sum.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -68,12 +68,12 @@ namespace reduce_sum
bool rc = true;
switch (arg->get_element_type())
{
NGRAPH_TYPE_CASE(evaluate_sum, i32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, i64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, u32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, u64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, f16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_sum, f32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, i32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, i64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, u32, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, u64, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, f16, arg, out, axes, keep_dims);
NGRAPH_TYPE_CASE(evaluate_reduce_sum, f32, arg, out, axes, keep_dims);
default: rc = false; break;
}
return rc;
Expand Down
5 changes: 3 additions & 2 deletions ngraph/core/src/op/util/broadcast_base.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -361,7 +361,7 @@ bool op::util::BroadcastBase::evaluate(const HostTensorPtr& arg0,
const HostTensorPtr& out,
const AxisSet& broadcast_axes) const
{
NGRAPH_OP_SCOPE(op_util_BroadcastBase_evaluate,
NGRAPH_OP_SCOPE(util_BroadcastBase_evaluate_axes,
runtime::reference::broadcast(arg0->get_data_ptr<const char>(),
out->get_data_ptr<char>(),
arg0->get_shape(),
Expand Down Expand Up @@ -501,7 +501,8 @@ bool op::util::BroadcastBase::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(
op_util_BroadcastBase_evaluate, Shape target_shape = get_target_shape(inputs[1]);
util_BroadcastBase_evaluate,
Shape target_shape = get_target_shape(inputs[1]);

PartialShape result_shape;
std::pair<bool, AxisSet> pair_broadcast_axes;
Expand Down
2 changes: 1 addition & 1 deletion ngraph/core/src/op/xor.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -86,7 +86,7 @@ namespace logxor
bool op::v1::LogicalXor::evaluate(const HostTensorVector& outputs,
const HostTensorVector& inputs) const
{
NGRAPH_OP_SCOPE(op_v1_LogicalXor_evaluate,
NGRAPH_OP_SCOPE(v1_LogicalXor_evaluate,
return logxor::evaluate_logxor(inputs[0], inputs[1], outputs[0], get_autob()));
return false;
}
Expand Down

0 comments on commit 29304c9

Please sign in to comment.