-
Notifications
You must be signed in to change notification settings - Fork 0
GoogleSummerOfCode
Spend your summer doing something interesting and valuable for the open-source community and join Google Summer of Code. Read more information about how the program works on this page.
OpenVINO Toolkit is a mentoring organization for 2022!
We require one new notebook pull request sent to our OpenVINO notebooks repository from each potential GSoC contributor before accepting participation for GSoC. In addition, we would like to see if you know how to code and your coding style. To fulfill this requirement please:
- Fork OpenVINO Notebooks repository.
- Implement a new notebook as a 2xx notebook to demonstrate one of your favorite models from Open Model Zoo (Intel or public one) (please don't implement notebooks for models with demos already done or in progress - you can find them here and here).
- The notebook should be consistent with others in the same branch and should fit the requirements for the contribution.
- Your notebook should have descriptions and comments proving you know what you do (similar to descriptions and comments in e.g., 403 notebook).
- Create a pull request. Add the "gsoc" label to your PR.
Your application should consist of the following parts:
- About you
- Your full name
- Your university/current enrollment
- Short bio
- Your experience in programming (especially C++ and Python)
- Your experience in ML and DL
- About the project
- What is your choice?
- Why did you choose this specific idea?
- How much time do you plan to invest in the project?
- Provide an abstract of the solution
- Provide a detailed timeline of how you want to implement the project (include the main points you want to cover and dates)
- Answer the questions:
- How do you know OpenVINO?
- What do you know about OpenVINO?
- How could you apply it to your professional development?
- Describe any other career development plan you have for the summer in addition to GSoC.
- Why should we pick you?
- Tasks
- Screenshot of running object detection (401-object-detection-webcam notebook) – please take a screenshot for the last code cell
- Link to your pull request (for the prerequisite task – the top part of this document), even if it is already merged or closed
Proposal examples can be found here and here. Please contact early to discuss your application with the mentor.
Short description: ARM CPUs support has been added to Inference Engine via the dedicated ARM CPU plugin. ARM processors are widely used in Android smartphones, so we want to develop an Android demo application that demonstrates plugin possibilities on this platform. The demo should be written in Java and use Java wrappers to reach Inference Engine public API. We suggest reviewing the functionality of OMZ object detection demo and propagating its core functionality to the Android demo.
Expected outcomes: Android demo application with object detection functionality. Any model (or several models) from the list supported by the plugin could be used.
Skills required/preferred: Practical experience with developing Java apps for Android, at least a basic understanding of computer vision and deep learning technologies, enough to run a network and make it run at real-time speed
Mentors: Junwen Wu
Size of project: 350 hours
Difficulty: Medium
Short description: OpenVINO Inference Engine is a set of C++ libraries with C and Python bindings providing a common API to deliver high performance deep learning inference solutions on multi-platform including Windows, Linux, and Mac OS. We would like to extend these platforms by providing WASM build for the inference engine to enable the web/JS developers to use OpenVINO inference engine in the web browser.
Expected outcomes:
- WASM build of OpenVINO inference engine
- JavaScript and TypeScript friendly API for running inference in the browser using OpenVINO WASM build
- Sample Projects and documentation
Skills required/preferred: C++, JS, WASM and Emscripten
Mentors: Radwan Ibrahim
Size of project: 350 hours
Difficulty: Medium to hard
Short description: OpenVINO is an open-source product consisting of a set of tools and C++ runtime libraries which allows to perform inference on different hardware including Intel accelerators such as CPU, GPU, GNA, MYRIAD, ARM CPU, NVIDIA GPU, in a unified manner - write an application once using OpenVINO Runtime API and deploy on different OSes/hardware. The goal of this GSOC project is to introduce native support for Apple platforms (M1 and forward) by leveraging AI accelerators via the BNNS library-based plugin for OpenVINO. The work supposes the implementation of the plugin with all details: identify matching between OpenVINO operations and BNNS layers, adding graph transformations to convert between them, creation of BNNS graph, construction of optimal inference pipeline to execute the BNNS graph of operations, and take care of memory management.
Expected outcomes: Minimal working OpenVINO plugin which is able to perform inference via BNNS backend to offload heavy operations to Apple M1 accelerator, the plugin should demonstrate performance comparable to direct usage of Apple software, provide accurate results, have scalable architecture.
Skills required/preferred: C++, objective-c/swift, the interaction of C++ and objective-c/swift, system programming, deep learning frameworks, graph optimizations, multi-threading programming, gtest, cross-platform make
Mentors: Nico Galoppo
Size of project: 350 hours
Difficulty: Medium to hard
Short description: Mediapipe is a customizable platform for ML solutions for media processing. Intel OpenVINO is a set of C++ libraries with C and Python bindings providing a common API to deliver high-performance deep learning inference solutions. MediaPipe currently supports CNN inference using TensorFlow and TFLite frameworks. We want to add support for OpenVINO inference to MediaPipe. To do it, we need to develop new calculator classes supporting OpenVINO. Suggested calculator types:
- Minimal task (detects object according to provided model):
- Object detection calculator:
- Input: IMAGE:image
- Ouput: DETECTION:detections
- Object detection calculator:
- Maximal task
- BERT question answering calculator:
- Input: TENSOR: tokens
- Output: Start-end indexes
- Tokenizer calculator:
- Input: Text or URL (may be split it into 2 different calculators)
- Ouput: TENSOR: tokens
- Index_to_Text calculator:
- Input: Text or URL
- Input: start-end indexes
- Ouput: text
- BERT question answering calculator:
Inference calculators should be created using the ModelAPI framework (part of OpenVINO Model Zoo). Also, examples with a full pipeline should be created.
Expected outcomes: Working examples for Object detection and BERT question answering working with OpenVINO-based calculators.
Skills required/preferred: C++ knowledge, basic knowledge of OpenVINO framework, experience with Google MediaPipe is a plus
Mentors: Patrick Foley
Size of project: 350 hours
Difficulty: Medium to hard (depending on chosen scope)
Short description: The world is 3D but in AI development we often work on flat 2D displays with flat data visualization plots and charts. In machine learning, many of the tasks may be much better understood if we provide a 3D or 4D (space and time) perspective. In this project, we will answer this question by providing the beginner-friendly Jupyter 3D engine for machine learning visualization along with WebGL integration. Our main goal is not only to make 3D or 4D datasets easier to visualize, but also to make machine learning easier to understand in a more humanistic way.
Expected outcomes: Working 3D support for Jupyter Notebooks running OpenVINO (AI inference). For example, visualizing 3D body pose and characters, visualization 2D-3D mapping, and also providing a clean interface to set up these without a 3D engine or graphics programming background.
Skills required/preferred: Understanding of graphics pipelines and (GPU programming is a plus), software Engineer background and code releasing, Python, C++
Mentors: Raymond Lo
Size of project: 350 hours
Difficulty: Medium to hard
Short description: In most cases, DL model optimization requires the presence of real data that the user should provide for the optimization method (e.g. quantization or pruning). However, it was noticed that such methods can perform quite well on synthetic data which are not even relevant to the use case. In this task, we will create a synthetic GAN or VAE based DL model that is capable of generating synthetic images based on the text hints provided by the user. This data will be used to evaluate the 8-bit post-training quantization method on a wide range of Computer Vision models.
Expected outcomes: DL model that generates synthetic data based on the text input.
Skills required/preferred: DL model training, understanding of GAN, VAE architectures
Mentors: Mansi Sharma
Size of project: 350 hours
Difficulty: Medium
Short description: DL model inference performance is a hop trend of recent years that is aimed at bringing AI into real-world applications. And OpenVINO is a famous DL inference solution that delivers best-in-class performance on Intel Architectures. In this task, we will provide a Jupyter notebook/tutorial (similar to this one) where we showcase inference performance on the set of popular DL models that come from the PyTorch Image Models project on GitHub (timm). The tutorial will provide details on how to install OpenVINO, convert models to OpenVINO Intermediate Representation, and do the proper benchmarking under the various settings.
Expected outcomes: Pull-request with the tutorial into PyTorch Image Models.
Skills required/preferred: DL basics, understanding of ML model optimization
Mentors: Pradeep Sakhamoori
Size of project: 175 hours
Difficulty: Easy
Short description: OpenVINO integration with TensorFlow is an open-source product for TensorFlow developers who want to accelerate their inference on a variety of Intel compute devices such as CPU, iGPU, VPU (NCS2), etc. In this project, the developer will create and deploy OpenVINO integration with TensorFlow-based Jupyter notebooks for Vision/NLP (Huggingface etc) model on Intel devices. The goal is to demonstrate the acceleration of these models as well as ease of adaptation on different Intel devices. It is encouraged to create/extend docker images to deploy this solution on Intel devices in different environments.
Expected outcomes: Jupyter notebook containing inference scripts for selected Vision/NLP models. Docker files/images which are used to demonstrate model inference using OpenVINO integration with TensorFlow.
Skills required/preferred: Basic DL concepts, TensorFlow DL model inference, C++ and Python Programming knowledge
Mentors: Chandrakant Khandelwal, Ritesh Rajore
Size of project: 175 hours
Difficulty: Medium
10. Multi-channel streaming pipeline with inference application using OpenVINO integration with TensorFlow on Intel device
Short description: OpenVINO integration with TensorFlow is an open-source product for TensorFlow developers who want to accelerate their inference on a variety of Intel compute devices such as CPU, iGPU, VPU (NCS2) etc. In this project, developers are expected to create an end-to-end application that can demonstrate streaming from multiple image/video sources and run OpenVINO integration with TensorFlow-based inference on the same multi-stream input. It is expected to design and implement an application that can effectively use available multi-core Intel CPUs for latency/throughput mode. Task partitioning and scheduling inference applications will be important aspects of this project. It is encouraged to create or use existing docker files/images to demonstrate the scalability of the solution.
Expected outcomes: Python application for streaming from camera/video-files, a python script that can launch multiple inference instances, performance data collection, docket file/images
Skills required/preferred: DL basics, TensorFlow DL model inference, C++ and Python Programming knowledge
Mentors: Suryaprakash Shanmugam, Ritesh Rajore
Size of project: 175 hours
Difficulty: Medium
Contribution guidelines can be found here.
- Subscribe to our GSoC mailing list
- Start a new thread by sending an e-mail to [email protected]
Please contact us early to discuss your application with the mentor.
© Copyright 2018-2022, OpenVINO team
- Home
- General resources
- How to build
-
Developer documentation
- Inference Engine architecture
- OpenVINO Python API
- CPU plugin
- GPU plugin
- HETERO plugin architecture
- Snippets
- Sample for IE C++/C/Python API
- Proxy plugin (Concept)
- Tests