Skip to content

Commit

Permalink
update compiled_model.export_model to file snippet
Browse files Browse the repository at this point in the history
  • Loading branch information
ilya-lavrenov committed Sep 17, 2023
1 parent 652f052 commit 6eeeb45
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 5 deletions.
1 change: 1 addition & 0 deletions docs/OV_Runtime_UG/deployment/local-distribution.md
Original file line number Diff line number Diff line change
Expand Up @@ -137,6 +137,7 @@ OpenVINO Runtime uses frontend libraries dynamically to read models in different
- ``openvino_tensorflow_lite_frontend`` is used to read the TensorFlow Lite file format.
- ``openvino_onnx_frontend`` is used to read the ONNX file format.
- ``openvino_paddle_frontend`` is used to read the Paddle file format.
- ``openvino_pytorch_frontend`` is used to convert PyTorch model via ``openvino.convert_model`` API.

Depending on the model format types that are used in the application in `ov::Core::read_model <classov_1_1Core.html#doxid-classov-1-1-core-1ae0576a95f841c3a6f5e46e4802716981>`__, select the appropriate libraries.

Expand Down
6 changes: 1 addition & 5 deletions docs/snippets/export_compiled_model.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -5,13 +5,9 @@ int main() {

//! [export_compiled_model]
ov::Core core;

std::stringstream stream;

ov::CompiledModel model = core.compile_model("modelPath", "deviceName");

std::fstream stream("compiled_model.blob");
model.export_model(stream);

//! [export_compiled_model]

return 0;
Expand Down

0 comments on commit 6eeeb45

Please sign in to comment.