 When my son was two years old, one day I showed him the character Yi on a piece of paper. The next day he will point to the different Yi in the street, including this huge upside-down Yi painted on the football field. I was amazed that he can learn and generalize from just one example. When my little daughter saw this Picasso painting, she screamed, phase, right away, even though she has never seen such a distorted phase before. Of course, being my children, they are naturally very smart. But how is that possible? Computers today aren't this intelligent. Computers can only do things that they are trained for. For example, after the Boston Marathon bombing, human FBI agents have to come in to watch our surveillance tape to identify the bombers. Computers cannot do that because they don't even know who and what to look for. We can train computers to learn to recognize objects by giving them millions of examples with the correct answers. A human baby, on the other hand, learn to recognize many concepts and objects all by themselves, simply by interacting with a few examples in the real world. My research at Carnegie Mellons involved looking inside the brain to study what is going on at the levels of individual brain cells and circuits when the brain is seeing and learning to recognize objects. We want to use this knowledge to make computers see and learn like humans. For humans, seeing is a creative process. There is a distinction between what our eyes take in, which are fragments of the world, and what we perceive. We rely heavily on our experience and knowledge to make up the image that we see in our mind. These pictures illustrate what I mean. On the left, you see a red translucent surface, but really there's no red surface, only fragments of the back ring have been turned red. On the right, you see a white triangle, but in reality it's not there, it is an illusion. What we see in our mind is our interpretation of the world. The brain fills in a lot of missing details to make up the most probable mental image that can explain what comes into our eyes. We see with our imagination, and we create an image that we perceive in our mind. Creating this image in our brain involves the interactions of many levels of brain circuits in the visual cortex, the part of the brain that is responsible for processing visual information. During perceptions, information flows up and down across the different levels to integrate global and local information. We have observed that at the high level, neurons can see the white triangles, but only firstly. Neurons at the low level can see more clearly, but initially they only see fragmented wheel of the world. But after a brief moment, they start to represent the white triangle as well. We believe that the brain created this image as a way to check whether its interpretation of the world is correct. This is the result of a computer program we roll based on the same principles. Given an image, the program has to interpret its 3D structures. For each interpretation, it can imagine what it expects to see. And if the imagined image matches the input image or explains the input image, the program knows he got the right answer. So the brain, when it is not interpreting and perceiving the world using this process, it can use the same circuit to imagine how an object might look like under different situations. So in this way, it can actually generate a huge amount of big data based on a few examples to train itself. So our natural impulse to daydream, to imagine the future and to do creative things and our innate need for artistic expression and art-making could actually be all byproducts of the same program and process running constantly in our brain. So these activities could be key to the development of our abilities to learn from a few examples and to recognize objects that we have never seen before. So we want to give this capacity of creativity and imaginations to computer so that they can learn automatically like a baby. As a part of this large-scale Apollo project of the brain, recently launched, we at Carnegie Mellons are studying the actual neural circuits that enable our visual system to see and imagine with imaginations and we want to put this to work in computers. So we hope this work will not only help us to understand the brain better but to make more flexible and intelligent robots. And the question I want to leave with you is can you imagine what machines can do if they have the powers of imagination? Thank you.