If this is your first visit, be sure to
check out the FAQ by clicking the
link above. You may have to register
before you can post: click the register link above to proceed. To start viewing messages,
select the forum that you want to visit from the selection below.

Recently, you may have noticed people sharing stylized photos and videos that resemble famous paintings, like "Starry Night" by Van Gogh and "The Scream" by Munch. But how do the apps that make these images work? How can the style of a painting be transferred to a selfie or a photo of your dog? Turns out, a whole lot of A.I. using machine learning and deep neural networks. In this video, we break down the basics of how style transfer works, and demo some recent style transfer experiments created by research scientists at Google.

The world is filled with things that most of us are able to understand and react to without much thought… a stop sign partially covered by snow is still a stop sign… a chair that’s five times bigger than usual, is still a place to sit. But for computers, the world is often messy and complicated. Google engineers and researchers discuss how machine learning is beginning to make computers, and many of the things we use them for (maps, search, recommending videos, translations), better.

With all of the exciting A.I. stuff happening, there are lots of people eager to start tinkering with machine learning technology. That's why we've created A.I. Experiments, a site that showcases simple experiments that let anyone play with this technology hands-on, and resources for creating your own experiments.

Various visualizations featured in video made by Gene Kogan. Additional footage by Sarah Riazati.