vikz-arsculpture

Untitled Duck AR by Meijie and Vicky is an augmented reality duck that appears on your tongue when you open your mouth, and is triggered to yell with you when you stick out your tongue.

https://www.youtube.com/watch?v=xt1FgOcHXko&feature=youtu.be

Process 

We initially started off by using a bunch of our limited developer builds (heads up for future builds: there is a limit of 10 per week, per free developer account lol) by testing the numerous different types of build templates that we could use to implement AR over our mouth, most particularly image target, face feature tracker, and face feature detector.

We actually got to successfully have an image tracker work for Meijie's open mouth, however, it was a very finicky system because she would have to force her mouth to be in the same exact shape, and very similar lighting, in order for it to register. We plugged in an apple prefab, and thought it was quite humorous as it almost was like being a big stuffed with an apple.

With this, we initially wanted to explore having an animation of some sort take place in the mouth. However, that proved difficult due to the lack of accuracy with small differences in depth, and also the amount of lighting that would need to be taken into consideration. Also, because the image target had some issues with detecting the mouth, we decided to migrate to the face mesh and facial feature detector.

We combined both the face mesh and feature detector, to trigger a duck to appear on the tongue when the mouth is open.

12.4.19 Updated Prototype

Having the duck appear (within grass and more refined detail) when mouth is first opened, and then having a raw duck (yum yum!) appear the second time mouth is opened.

vikz-arwo-justaline

For the Justaline project Arden and I played around with different possibilities and ideas. Ultimately, we really enjoyed the depth effect we could make in the app by drawing multiple "doorways" to move through. We also took advantage of the potential to create transformative effects by having shaped frames change into other shapes while passing through. This augmented reality creates an anticipation for the viewer, kind of like a tunnel or a rabbit hole, as they travel through the floating framed shapes. Here we chose to transform a triangle into a circle:

viks-UnityScripting

For my Unity Scripting tutorial, I decided to start at the basic fundamentals and dive into the Course with Code. I went through the first tutorial of the series, Start Your 3D Engine, in which a simple car game "world" and assets are being created and imported, and then I went through the second tutorial, in which we create and apply the first C# script.

Part 1.1

Part 1.2 with scripts

vikz-SituatedEye

We were really inspired by Google's drawing machine learning, and the ability to play around with the different types of applications that machine learning has in with drawing. In order to most quickly and accurately iterate over and over again, we started our explorations by playing around with the whiteboard. We started off playing around with the program to see if machine learning was able to detect the difference between written text and drawings. From this, we were also thinking of maybe incorporating mathematical graphs and/or equations as a possible third scope; an example that lives between text and drawing.

From our experiments, we saw that computer could usually detect between drawings and text, presumptuously mostly dependent on the text. The diversity of drawings was differed widely, as we literally began to draw just about everything that first came to mind, whereas text was definitely more limited in terms of aesthetic, and was visually more uniform. However, we came upon an interesting discovery when drawing stars, but in a linear form. Despite being drawings, it was detected as text, because of its linear nature. This propelled us into thinking about the possible implications for using machine learning to detect the differences between languages.

The stars that sparked off our stars.

Our final exploration dealt with exploring the detecting the difference between western and eastern languages; western being more letter-based, and eastern being more pictorial-based characters.

Western Languages

Eastern Languages

Training our model with white background, western text, and eastern text.

We decided to map the result out visually through three colors:

  • White indicates that there is no hand written text being shown. (We fed a series of white background images to train this part)
  • Blue indicates that it is more-so western text. (We fed a series of handwritten phrases and words in English, Spanish, French, and Latin to train this part)
  • Red indicates that is more-so eastern text. (We fed a series of handwritten phrases and words in Chinese, Japanese, and Korean to train this part)

From our results, we've discovered a couple things.

The program is relatively good at detecting a "blank background" though a couple times, when our paper was shifted, the program recognized it as "western".

But most importantly, the program was very accurate in detecting western text, but significantly less so with eastern text.

This observation has led us to a couple hypotheses:

  • Our data is lacking. We took about 100 photos for western and eastern each, but this may have been not enough for the machine learning to generative a conclusive enough database.
  • The photos that we took could also have been of not high enough quality.
  • In our sample data, we only wrote horizontally for western text, where as eastern had both horizontal and vertical.

Future thoughts...

To test the machine learning program to see if could simply tell the difference between western and eastern languages, we could do away with the "varied handwriting" completely and use very strict criteria (for handwriting style) in writing our sample text. When we tested the learned program, we could continue to write in that same style between the eastern and western texts. This could help isolate our variables to test out our above hypothesis.

vikz-ML

A. Pix2Pix

edges2cats

For edges2cats, I found it most fun and interesting when I would try to create things completely unlike cats -- for example, a teacup, and see what kind of "cat" would be created. I found it odd, however, that when I did strive to make a cat, it didn't seem to register it all that well (ie. my double headed cats)

facades 

Facades was also interesting to play with in that it allowed me to create this unique entities that, in my opinion, are barely recognizable as buildings and architecture.

B. GANPaint Studio

GANPaint studio, in my experience, was a bit more frustrating to use. I had some difficulty in painting over certain areas and altering the landscape. The "dome feature" was the only one that I could paint over in random areas and have be the most notably.

C. ArtBreeder

General 

Portrait 

The most intriguing part of ArtBreeder for my was by far toggling between the different percentage of genes in each photo. Even merely scaling one gene on one side of the spectrum to the other side of the spectrum produced crazy diverse results. I found it interesting how I managed to create something that looks of a bread / boiled crab / cheesey tomato dish without using a single element of anything of that sort.

D. Infinite Patterns by Alexander Mordvintsev

Original Image

 

Tiling Image

E. GPT-2

How to make a chicken pot pie in under 5 minutes:  This recipe makes over 8 chicken pot pie, or 6 for a larger meal.  Make sure your soup is very soupy and thick before pouring your soup into a pie shell.  This chicken pot pie is also great with a crust or rice pudding.  

3 lbs bone in chicken, skin on for quick cooking for lunch/dinner  

3 cups chicken stock (or broth)

1/2 tsp salt

1/2 tsp pepper

1/2 tsp garlic powder

1/2 cup onions, chopped

12 eggs, whisked and peeled

1 egg white, whisked

a splash of lemon juice

1 Tbsp olive oil

1 1/2 cups baby carrots, peeled and chopped

1 tsp vanilla extract

1 cup chicken stock (or broth) 

a dash of red wine vinegar

2 tsp salt

1 egg, beaten with about 1/3 cup of the stock (or broth) in a small bowl (preferably more than your own cooking water)

2 Tbsp extra virgin olive oil

It all started when I first met her at Lion King