Week12 Assignment: Final Concept Documentation–Crystal Liu

Background+Motivation

My final project is mainly inspired by the re-created famous paintings, especially the portraits. Some people replace the Mona Lisa’s face with the Mr. Bean’s face and the painting is really weird but interesting. 

Image result for mr bean and mona lisa

Also, I found that some people tend to motivate the poses of the characters in the painting, such as The Scream:

Image result for people imitate the screamImage result for people imitate the scream

Therefore, I want to build a project to let the users add their creativity to the famous painting and personalize the paintings to recreate these paintings. It reminds me my previous assignment for style-transfer. For that assignment I use a painting from Picasso to train the model, so that everyone or everything showing in the video can be changed into Picasso’s style. Even though the result is not that good, it still shows a way to personalize the painting or to let the users create their own version of paintings.

My idea is that the user can trigger a famous painting by imitating the pose of the characters in that painting. For example, if the user wants to trigger The Scream, he or she needs to make the pose like this: 😱. After the painting showing up, the user can choose to transfer the style of the live camera to the style of The Scream. If the users want to change to another painting, they just need to do the corresponding pose to trigger the expected painting.

Reference

My reference is the project called moving mirror. The basic idea is that when the user makes a certain pose, there will be lots of images with people making the same or similar pose.

What attracts me most is the connection between images and human poses. It displays a new way of interaction between human and computer or machine. Users can use certain poses to trigger things they want, and in my project it is the painting. 

The second one is style-transfer. It reminds me some artistic filters in Meituxiuxiu, a popular Chinese photo beautification application. These filters can change the style of the picture to sketch style, watercolor style or crayon style.

But the filter is only for still picture. I want to use style-transfer model to add this filter to the dynamic video so that the user can see their style-changed motions in a real time.

Week 12 Assignment: Document Final Concept —— Lishan Qin

Background

When I was young, I was fascinated by the magic world created by J.K. Rowling in Harry Potter. She has created so many bizarre objects in that world of magic that I still find very remarkable today. “The Daily Prophet”, a form of newspaper in the Harry Potter world, is the main inspiration of my final project. “The Daily Prophet” is a series of printed newspaper that contains magic which allows the image on the printed paper to appear as if it’s moving. It inspires me to create an interactive newspaper with an “AI editor” where not only the images on the newspaper will update every second according to the video captured by the webcam, but also the passage on it will change according to the image. In my final project, I will use Style Transfer to make the users’ face appear on the newspaper and utilize im2txt to change the words of the passages on the newspaper according to what the user is doing. I will build an interactive newspaper that constantly reports the users’ action. 

       

Motivation

Even with the development of social medias which allow new information to be spread almost every moment and every second, it still requires human people behind the screen to type, collect and then post those news. However, if there is an AI editor that could document, write and edit the news on the newspaper for us, the real-time capability of spreading information of the newspaper would be even better. Thus, I want to create an interactive self-edited newspaper that asks an AI to write the news about the action of the people it sees by generating sentences on their own. 

Reference

I’ll refer to the im2txt model on github https://github.com/runwayml/p5js/tree/master/im2txt here to create the video caption. This model will generate sentences according to the object and action the webcam video caught. I will run this model on the runway and then it will sent the result of the caption to html so that I can manipulate the outcome. Since some of the captions aren’t that accurate, I still need to find some ways to improve on that.

Week 12 Assignment: Document Final Concept – Ziying Wang (Jamie)

Background

Dancing With Strangers is a development of my midterm project Dancing With a Stranger. In my midterm project, I used Posenet model to mirror human movements on the screen and exchange the controls of their legs. With my final project Dancing With Strangers, I’m hoping to create a communal dancing platform that enables every user to use their own terminal to log onto the platform and mirror all the movements on the same platform. In terms of the figure that will be displayed on the screen, I’m planning to build an abstract figure based on the coordinates provided by Posenet, The figure will illustrate the movements of the human body but will not look like the skeleton or the contour.

Motivation

My motivation for this final project is similar to my midterm project: interaction with electronic devices can pull us closer, but can also drift us apart. Using them to strengthen the connections between people become necessary. Dancing, in every culture, is the best way to pull together different people, a communal dancing platform can achieve this goal. Compared with my midterm project, the stick figure I create was too specific, in a way, being specific means assimilation. Yet people are very different, they move different, Therefore, I don’t want to use the common-sense stick figure to illustrate body movement. Abstract provides us with diversity, without the boundary of the human torso, people can express themselves freer.

Reference

For building the communal dancing platform, I’m using the Firebase server as a data collector that records live data sent by different users from different terminals.

For the inspiration of my abstract figure, I’m deeply inspired by the artist and designer Zach Lieberman. One of his series depicts the human body movement in a very abstract way, it tracks the speed of the movements and the patterns illustrate this change by changing its size. With simple lines, bezier and patterns, he creates various dancing shapes that are aesthetically pleasing. I plan to achieve similar results in my final project.

Some works by Zach Lieberman

AI ARTS- Week 12- Final Proposal- Eszter Vigh

AI
How Users Think of AI

I went back to the drawing board with this idea SO many times. It took visiting McaM to really get an idea solidified. I’ve been working towards that idea since that field trip last Saturday. 

So what’s the big idea? 

It’s an AI Chat Bot. But what I want to do with it… is kind of cool. I want to ask users afterwards to provide feedback on it. So, the core of my project is to have a chat bot that is as human as I can make it (using a Google API).

NOW… here is my big issue. I don’t want to collect the chat history or any data really from the users. Rather, I just want user-tester feedback. The google API unfortunately, “By default, bots can only read the basic identity of users that invoke them. This information includes the user’s display name, user ID, email address, and avatar image.” I don’t even want that. 

So, I may end up using a slightly different API. I found this tutorial, and I think it will honestly be better.

This tutorial requires my own data. I’m going to do some experimentation and research on what other AI conversation bots use as their training data and then maybe include some of my own Eszter-isms. What I still really need to narrow down based on this blog post I found, is what the purpose of my AI chat bot is. Do I just want to make it an all-around bot and have users find out what doesn’t work? Like… try to see the limit of knowledge the bot has? 

Survey
This is a potential hypothesis

That would be cool. (At least, I think it would be). For the User Test survey I want to ask the following questions:

  1. How human is this Chat Bot ? (Scale 0-5, Five = Practically Human, Zero= Obviously Coded Computer)
  2. What did you find most human about the chat bot?
  3. How un-human is this Chat Bot ? (Scale 0-5, Five = Undeniably a Computer, Zero= Human)
  4. What did you find most un-human about the chat bot?
  5. Would you feel comfortable using this Chat Bot again? Why or why not?

I want to set up a computer in the IMA lab with my experiment with a little survey collection box. I plan on shredding all of the survey results after to maintain privacy of the users. The goal is make this fully anonymous and voluntary. 

My hypothesis for this project is that most users will feel uncomfortable using the chat bot because of the same logic they use to reason through not signing up for the face-scanner to get into the building. I think they will view this user testing session will reveal some conflicts between human and machine. 

The design of the chatbot will be super simple, very basic chat room-esque is what I am going for.