Skip to content

Here, I have prepared a Neural Networks Style Transfer Model.

License

Notifications You must be signed in to change notification settings

ThinamXx/NEURAL_STYLE_TRANSFER

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 

Repository files navigation

NEURAL NETWORKS STYLE TRANSFER

Neural Style Transfer Algorithms:

  • Neural Style Transfer is the task of changing the style of an image in one domain to the style of an image in another domain. Neural Style Transfer or NST refers to a class of software algorithms that manipulate digital images or videos in order to adopt the appearance or visual style of another image. NST algorithms are characterized by their use of deep neural networks for the sake of image transformation.

Libraries and Dependencies:

  • I have listed all the necessary Libraries and Dependencies required for this Project here:
from d2l import torch as d2l
from IPython import display
import matplotlib.pyplot as plt

import os
import torch                                     
import torchvision
from torchvision import transforms
from torch import nn
from torch.nn import functional as F

Preprocessing and Postprocessing:

  • I will initialize the composite image as a content image. The content image is the only variable that needs to be updated in the style transfer process. Now, I will read the content and style images. I will inspect the coordinate axes of images. I will define the functions for image preprocessing and postprocessing. The preprocess function normalizes each of the three RGB channels of the input images and transforms the results to a format that can be input to the CNN. The postprocess function restores the pixel values in the output image to their original values before normalization.

Image

Extracting Features:

  • I will use VGG-19 model pretrained on ImageNet dataset to extract image features. I will select the outputs of different layers from the VGG Neural Network for matching local and global styles. I will define the loss functions which are used for style transfer. The loss functions include content loss, style loss and total variation loss. I have presented the implementation of Function for Extracting Features and Square Error Loss Function using PyTorch here in the Snapshot.

Image

Creating and Initializing Composite Images:

  • In Style Transfer, the composite image is the only variable that needs to be updated. So, I will define a simple model in which composite image is treated as a model parameter.

Image

Training the Model:

  • During model training, I will constantly extract the content and style features of composite image and calculate the loss function.

Image

About

Here, I have prepared a Neural Networks Style Transfer Model.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published