-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[GNA] Remove extra FQ layers from the final network #10599
[GNA] Remove extra FQ layers from the final network #10599
Conversation
d7b91c5
to
f65e627
Compare
f65e627
to
8674cad
Compare
ngraph::opset8::Constant::create(ngraph::element::f32, {1}, {-10.}), | ||
ngraph::opset8::Constant::create(ngraph::element::f32, {1}, {10.}), | ||
static_cast<uint32_t>(std::numeric_limits<uint16_t>::max()) + 1); | ||
auto concat = ngraph::builder::makeConcat({fq1, fq2}, 1); | ||
function = std::make_shared<ngraph::Function>(concat, params, "WeighableLayerWithoutFq"); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you please explain why it is named WeighableLayerWithoutFq
? I don't understand why it is Weighable and why WithoutFq
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
As far as I remember this test was written for issue with quantization of ConcatAlignFilter. This layer is weightable and it's inserted by GNA plugin, so POT doesn't quantize it. So, I think, it's a reason of this test name. There is a PR where this test was added: #7834
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
In general, LGTM
@@ -74,7 +74,8 @@ static const char softSignLayersCounter[] = "numSoftSignLayers"; | |||
static void insertDiagonalLayerBetween(InferenceEngine::CNNLayerPtr prevLayer, | |||
InferenceEngine::CNNLayerPtr nextLayer, | |||
std::shared_ptr<IPassManager> passmanager, | |||
float fillValue) { | |||
float fillValue, | |||
size_t in_data_idx = invalid_data_idx) { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just on future, lets do not mix code styles in function signatures
…10599) * [GNA] Fuse all FakeQuantize layers with their previous layers * [GNA] Fuse FQ with previous layer if it's not required for precision change * [GNA] Fixed MatMulOverloadCorrectionTest
Details:
Tickets:
58806