Course»Course 9»Spring 2019»9.60»Homepage

9.60  Machine-Motivated Human Vision

Spring 2019

Instructors: Atissa Banuazizi, Pawan Sinha

TA: Michael J Lee

Lecture:  TR11-12.30  (46-4062)        

Information: 

Announcements

An example of Deep Dream

As mentioned in class, this music video uses a Deep Dream-like algorithm to certain parts of the video. It's pretty interesting to watch:

https://www.youtube.com/watch?v=dJ1VorN9Cl0

The way it works that a DNN processes an individual frame from the video (which is processed in a series of daisy chained functions, ultimately culminating in labels at the end). Then a human artist selects some neuron in a DNN it would like to optimize for - for example, it looks like the artist chooses a "dog" category neuron for many of the frames.

The human then asks the network to optimize the *pixels* of the image such that the activation for this neuron is increased. This is possible because you can take the derivative of the neuron with respect to the pixels, so he can find out what direction to "move" in pixel space to slightly increase the output of the neuron. This happens a bunch of times.

What results is an altered image that "evolves" dog-like features. It "grows" these features out of parts of the image that sort of look like a dog already (i.e., near a local maxima in pixel space for a dog feature). This is similar to the perception of visual hallucinations, no? To do this for a video, one can apply this process to individual frames independently (as far as I know, this is how it was done. What might be another way to do this?)

This is also how adversarial images are produced. The idea is that you can run just a few steps of the optimization for some category *not* originally in an image. The image doesn't change much, but in practice the output of the neuron changes dramatically, leading to an erroneous prediction.

Announced on 15 March 2019  9:01  a.m. by Michael J Lee

Slides from today uploaded & upcoming tutorials

Hey all, the slides from today are uploaded on Stellar under Materials.

At some point in the coming weeks I'll be uploading tutorials on some things you'll need to do to perform your final project.

I've already uploaded one tutorial on using MATLAB to generate features and perform transfer learning (among other things that probably aren't necessary for your project, like running the "Deep Dream" algorithm on a deep net).

I will also provide more information on using Mechanical Turk soon. The plan is to make sure most of your time on the final project is spent on careful hypothesis generation, experiment design, and data analysis - not messing around with annoying details on Mechanical Turk. So no worries on learning how to use Turk from scratch - stay tuned for more details soon.

Announced on 07 March 2019  2:56  p.m. by Michael J Lee

Welcome - and classroom changes starting next week

Hello all,

Welcome to 9.60! I'm looking forward to a semester of discussing and learning about one of the most exciting perspectives in neuroscience today - the view that the brain (and specifically vision) can be understood as a computing system, and potentially reverse engineered.

In general, feel free to send me a quick email about any issue or question you have with the class (e.g. course expectations, homework help, grading, logistics etc) at mil@mit.edu

As Professor Sinha mentioned today in class, our Thursday class will be held in the same location, 46-3015.

However, starting from Tuesday 2/12/2019, our class will be held in 46-4062.

See you then!

Michael

Announced on 05 February 2019  3:43  p.m. by Michael J Lee