HCI Seminar
Speaker
Host
Arvind Satyanarayan
Assistant Professor, EECS
Interpretability and Safety
How can we understand the inner workings of neural networks? Neural networks greatly exceed anything humans can design directly at computer vision by building up their own internal hierarchy of internal visual concepts. So, what are they detecting? How do they implement these detectors? How do they fit together to create the behavior of the network as a whole?
At a more practical level, can we use these techniques to audit neural networks? Or find cases where the right decision is made for bad reasons? To allow human feedback on the decision process, rather than just the final decision? Or to improve our ability to design models?
Bio: Chris Olah is best known for DeepDream, the Distill journal, and his blog. He spent five years at Google Brain, where he focused on neural network interpretability and safety. He's also worked on various other projects, including early TensorFlow, generative models, and NLP. Prior to Google Brain, Chris dropped out of university and did deep learning research independently as a Thiel Fellow. Chris will be joining OpenAI in October to start a new interpretability team there.
How can we understand the inner workings of neural networks? Neural networks greatly exceed anything humans can design directly at computer vision by building up their own internal hierarchy of internal visual concepts. So, what are they detecting? How do they implement these detectors? How do they fit together to create the behavior of the network as a whole?
At a more practical level, can we use these techniques to audit neural networks? Or find cases where the right decision is made for bad reasons? To allow human feedback on the decision process, rather than just the final decision? Or to improve our ability to design models?
Bio: Chris Olah is best known for DeepDream, the Distill journal, and his blog. He spent five years at Google Brain, where he focused on neural network interpretability and safety. He's also worked on various other projects, including early TensorFlow, generative models, and NLP. Prior to Google Brain, Chris dropped out of university and did deep learning research independently as a Thiel Fellow. Chris will be joining OpenAI in October to start a new interpretability team there.