Giter VIP home page Giter VIP logo

souravs17031999 / neuralstyletransfer-showcase Goto Github PK

View Code? Open in Web Editor NEW
4.0 4.0 3.0 26.84 MB

This is a social volunteering project for implementing neural style transfer algorithm for creating artistic styles with awesome textures to be printed on tees :)

Jupyter Notebook 100.00%
neural-style-transfer style-transfer neural-network style-loss content-loss neural-style-transfer-pytorch convolutional-neural-networks cnn deep-learning style-transfer-algorithms volunteering social-cause teespring charity udacity secure-and-private-ai spaic

neuralstyletransfer-showcase's Introduction

Project Objective :

This project aims to acheive neural style transfer using the algorithm given by Paper Image Style Transfer Using Convolutional Neural Networks by Gatys, implemented in Pytorch.

This is a volunteering project made for social cause for Project #t-shirt.

I along with my team of highly enthusiastic data scientists/machine learning/deep learning engineers collaborated on a project which aimed to design AI generated images using style transfer or GAN's and print it on tees and the sales produced through it is given to charity.
Tees/stickers/mugs etc.. printed with AI generated images were sold on TeeSpring.
(For more details about charity see below)

Theory and basic overview :

Basics of CNN :

Convolutional neural networks (CNN) are class of deep neural networks but differs from basic MLP (multi layer perceptron) in a way that unlike MLP , where all the neurons of the previous layers are connected to all the layers of next layers and this unnecessarily increases number of parameters whereas in CNN , at each layer only some of the neurons are connected to next layers , the neurons which are important to preserve the spatial information. As we can see in the below picture how CNN's help in preserving spatial informative , as how these neurons are interconnected in 2D and thus helps in identifying patterns , relations among different parts of a image.

Hence, helping in extracting features , and also called feature extractors.

This works on the idea of mathematical operation called convolution , in simple words it is simply superimposing one function over the other and here also , we are masking the image with matrices known as "kernels" or "filters" and the results produced are called "feature maps".

This graphics illustrates under the hood calculation of different feature maps which then combines over all the channels (1 for gray scale images and 3 for RGB images) produces the final result. We can detect edges, contours based on the results of feature maps produced as sudden change in light intensity pixels confirms there is a edge so there values in features maps will signifactly differ around that region.

Content and Style images

So, now we know how to extract features from a image using CNN. Now, we are going to see what are content images and style images which we will be using in our algorithm to generate some new awesome artistic images.

Content images : images from which we are going to extract contours, borders, edges.

Style images : images from which we are going to extract textures, styles, artistic effects.

The algorithmic process explained in simpler way

Suppose you start off with a empty canvas and starts painting it with your brush , and then you realized you are far away with what you imagined you would draw. Then a idea striked you and you feel amazed to understand that in order to go towards the right direction you need to draw like following those lines which you are imagining so that difference between what you imagined and what you are drawing is miniming and thereby taking towards your final image.

Now think about this in terms of our content image and style image. The algorithm requires that we start with our target image (which is our final image) , now there are two options for doing this :

  • either we can start off by taking a empty convas
  • or we can start from cloning (or copying content image)

Now let's understand this in geometrical way.

(image has been oversimplified for simplicity but mathematical explanations also lies ahead)

In the images , you can imagine a grid of pixels overlayed on them and try to understand.

euclidean distance :

We can clearly see from the image , euclidean distance between copied content image and target image is less as compared to previous one.

Okay so we start with copied content image.
But then how do we proceed further.
So, basically CNN's have convolutional layers with classifier layers but since we need not classify anything here , so we will just use CNN as a feature extractor to extract edges, contours from content image and mix them up with textures extracted from style image to create a whole new artistic target image. But how do we go about doing it.
So, here now we wanna understood formally the loss functions separately as content loss and style loss considering distance between content image and target image and distance between style image and target image.

content loss :

Let Fˡᵢⱼ(x)∈ Cₙₙ(x)and Pˡᵢⱼ(x) ∈ Cₙₙ(x) describe the respective intermediate feature representation of the network with inputs x and p at layer l

style loss :

correlation between different filter responses given by the Gram matrix Gˡ, where Gˡᵢⱼ is the inner product between the vectorized feature map i and j in layer l. We can see that Gˡᵢⱼ generated over the feature map for a given image represents the correlation between feature maps i and j.
Thus, the total style loss across each layer is

where we weight the contribution of each layer’s loss by some factor wl

where Gˡᵢⱼ and Aˡᵢⱼ are the respective style representation in layer l of input image x and style image

You just need to know how dot product works.
Dot product is simply multiplication of vectors component wise (corresponding elements multiplies).

Hence, total loss will be as follows :

where α and β are the weighting factors for content and style reconstruction, respectively.

Now our objective is to minimize the total loss.

Minimize total loss :

We can apply our gradient descent algorithm to minimize loss and backpropogation algorithm to update parameters.

Model architechture VGG19 and hyperparameters summary :

Hyperparameters setting :

  • total iteration steps : 20000
  • optimizer : Adam
  • Learning rate : 0.003
  • content_weight = 1 (alpha)
  • style_weight = 1e6 (beta)

The Results :

Combined results of different images

Some more photos :

This was my final designed artistic photo made for stickers on tees/mugs etc.. Buy at teespring
tees1
tees2

[NOTE] : Since the notebook file is large so there are chances it will not open up on Github, if that happens, i have also uploaded project report (PDF) which will open fast.
[NOTE] : Everything about the scope of project and charity : click for team repo

⭐️ this Project if you liked it !

neuralstyletransfer-showcase's People

Contributors

souravs17031999 avatar

Stargazers

 avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.