What is DeepDream?
DeepDream is a computer vision program created at Google that uses a convoluted neural network to sequentially “interpret” an input, tweaking it until it matches a trained model. For example, turning static into bananas:
Google has a good article on what the tool is capable of, but it went viral for the cool artwork people were making with it.
Here’s an example of using a model trained on abstract expressionism:
What really captured my interest was how you can feed the same image into different models and get beautiful artwork styles. Here’s an example using abstract expressionism:
There are tons more examples here.
I went through some install instruction, installing Caffe (a deep learning framework) and CUDA (to use my GPU), and started playing around with images.
As the process runs, it spits out an image at every Nth level of processing. You can decide how many steps of processing you want. For fun, I strung these images together using ImageMagick to form a GIF.
The results were pretty neat!
I used an image of a bison from my friend’s trip to Yosemite. Here, you can see the “dog slugs” - an artifact of the default model. It quickly loses any semblance of the original image.
This one starts from a picture of ivy in an abandoned building. It evolves into cars, buildings, and trees.
I didn’t get as far into this project as I’d have liked, but it was an interesting learning experience.