Style Transfer Training: Cyborg Faces

Inference on Pre-Trained Models:

To start off, I found a basic image of a street in Hong Kong, and played around with pre-trained models in ml5.js. 

Input: 

Style:

Output:

As you can see from the output, the style transfer is quite amazing; the wave texture is perfectly captured from the style image and transferred over to the input image, while still keeping the overall structure (such as the cars, signs, and roadway). Personally, I am still not sure how this pre-trained model is able to take the tidal curve texture  (which is not overly present in the style image, only on specific waves), and apply it across the input image so nicely. 

Training My Model:

First off, I followed the instructions on the powerpoint, and downloaded the data set accordingly. The process wasn’t too bad; took only around 2 hours or so:

However, after acquiring the data set, I ran into a few errors during the training step:

Turns out there were some issues with my directory and image path, so I had to spend some time fixing those issues. 

Idea:

I wanted to experiment with style transfer using faces as the input image, since human faces are one of the most prominent, recognizable features to us.    Interestingly, I came across this article which touched upon the strange phenomenon where humans possess the tendency to find ‘faces’ on inanimate objects, especially cars. Personally, I’ve always thought that the fronts of cars have always had a certain personality to them (ie, some cars look happy, some cars sleepy, etc). In the article, it talks about a study published in the Proceedings of the National Academy of Sciences, where researchers had auto experts look at the fronts of cars, and they found that the same area of the brain involved in facial recognition was activated (fusiform face area). I thought this was especially interesting, in that car faces were essentially triggering the same responses as human faces in the brain. Therefore, for my style transfer project, I wanted to have human faces as the input image, and a car face as the style image. I thought it would be fascinating to see how the model would combine the perceived style of the car face into the human face, and what the results would be. Would the output image just be a photo where the machinelike style of the car is textured across the human? Or would the output face be more like a cyborg, where certain qualities of the car is fixed into the face, making it seem part human, part machine?

Results:

 +  =  

  +  = 

  +   = 

Judging from these results, I would say that my first run with the male test subject turned out the best; I think it may be because the prominent circles on the car translated over to the man’s pupils, giving him that ‘cyborg’ like feel I had hoped for. Interestingly, the model managed to position the headlights of each car into the eyes of the human faces, which I thought was quite fitting. The model also took the gridline patterns of the front of the cars and translated them across the background of the output images (which is most noticeable with the male subject and third female subject). Ultimately, my output images were quite striking in my opinion; it seemed as if they really were like ‘cyborgs’. It would also be interesting to see human faces styled with other face-like objects as well. 

Sources:

https://www.smithsonianmag.com/smart-news/for-experts-cars-really-do-have-faces-57005307/

Week 09 – Style Transfer Exercise – Alison Frank

***I am still working on training the model and will update this documentation once complete***

For my style image, I chose the following piece by Josef Albers due to its geometric qualities and  variations in color. I feel that combining this with another image would create an interesting result and I wanted to see how the model would interpret this input image. Along with this, I feel that this image has a very specific style to it which would make it very distinguishable once transferred.

While trying to train the model, I ran into many issues. Most of the errors I received were due to me not changing the directory path in the train.sh file, but after this was fixed I received another error:

error file result

After conversing with Aven, I found out that the training image set was not unzipped and downloaded properly.  Currently,  I am trying to properly download the training set and will try again to train the model, at which point I will update this documentation to reflect changes.

Week 06 – Midterm Proposal – Jenny

The idea

I really hope to combine what I have learned in class with my interest area in daily lives and as I said at the first class of this semester, I am interest in using photography or video in art forms to express myself. At the same time, I want to make this interactive process fun and easier for people to use. It could be better if my project could have some real-life usage. 

Then I landed on the idea to create an outfit generator. I would like to create an either real-time or non real-time platform where people can select the favorite  style they want to have. The detailed form could be as follows. The first way is letting the user simply type in the type of tops and bottoms they want to have and then the system will generate a predicted outfit image. The second way is letting the user input an image and then generate the predicted outfit image onto the user’s body. More detailed plan could be generated later on.

I think in everyone’s daily life, there is a need to create some visualized fitting system to help us select outfits. In reality, some online shopping website like has already implemented this kind of function. In the future, I hope everyone can have this kind of tools by empowering AI technology.

21

Action Plan

First of all, I planned to search similar work online including their models, their data set and any special techniques they used for reference. Then based on my initial research, I am going to do some deep dive into the areas I am interested in. 

I scanned the major fashion-related interactive machine learning project and find one research paper highly related to my initial idea. I requested for the dataset and find out that the model they used is called fashionGAN. There may have other similar models that could probably generate similar results and that’s the part I am going to research further.

1

Initial Findings

After taking a closer look at the models I found, I found out it is based on Torch and PyTorch, which I did not have any experience before. That could be hard for me to implement this project and I would like to find more related sources or models and also rethink of the whole workflow for this project.

Reference

https://github.com/zhusz/ICCV17-fashionGAN#complete-demo

http://mmlab.ie.cuhk.edu.hk/projects/DeepFashion/FashionSynthesis.html

Week 9 – Vietnamese Lunar New Year – Jarred van de Voort

Vietnamese Lunar New Year Portraits

Vietnam shares its own set of lunar new year traditions, cultures, and practices that have survived thousands of years. In light of the past lunar year year, I wanted to use the neural style transfer to generate stylized portraits for different years.

Staring with the source of the style transfer: 

This painting style is famous amongst traditional Vietnamese artists, with very distinct strokes and colors making it a good candidate for style transfer. Using the intel devCloud we can train using this as the source image to generate a model to apply to other images.

Here are some examples:

Year of the Rat – 2020

Year of the Monkey – 2016

Year of the Horse – 2014

You’ll notice that the context of the last image best matches that of the source image, making it a better candidate for generating a portrait. The two prior examples were isolated images that don’t have a background which may lessen the consistency of our style transfer. 

Style Transfer Assignment

Topic: Create a Galaxy

Inspiration: Inspired by One of Memo Akten’s work as below:
With the input video of an eye, it’s output is a Stardust. He mentioned that this is not done by style transfer.
What’s more, I’ve tried to make galaxy like images in Processing before, and it is very troublesome, especially to mimic the foggy and cloudy like texture. I used noise() in Processing for mimicking, but the result is not very good. Here’s an example:
Continue reading “Style Transfer Assignment”