Quest Symposium on Robust, Interpretable AI

Tuesday, November 20, 2018 at 3:00pm to 6:00pm

Building 46, Atrium and Auditorium
43 VASSAR ST, Cambridge, MA 02139

To advance further, deep learning systems need to become more transparent. They will have to prove they are reliable, can withstand malicious attacks, and can explain the reasoning behind their decisions, especially in safety-critical applications like self-driving cars.

The Quest Symposium on Robust, Interpretable AI will explore the latest techniques for making AI more trustworthy. Join us for posters by MIT students and postdocs, and talks by MIT faculty. Research topics will include attack and defense methods for deep neural networks, visualizations, interpretable modeling, and other methods for revealing deep network behavior, structure, sensitivities, and biases. 

2:30 pm. Aleksander Madry, "Robustness and Interpretability."
2:55 pm. Tommi Jaakkola, "Co-operative games of interpretability."
3:25 pm. Stefanie Jegelka, "Robustness in GANs."
3:50 pm. Poster Session A
4:30 pm. David Sontag, "Challenges and Dangers of Machine Learning in Health Care."
4:55 pm. Luca Daniel, "Evaluating the Robustness of Neural Networks."
5:15 pm. Antonio Torralba, "Dissecting Neural Networks."
5:40 pm. Poster Session B 

All are encouraged to attend. Light refreshments will be served. This symposium is part of the Robust Intelligence Initiative at the Computer Science and Artificial Intelligence Lab (CSAIL), funded by Microsoft and the MIT-IBM Watson AI Lab.

Event Type

Conferences/Seminars/Lectures, Meetings/Gatherings

Events By Interest


Events By Audience

MIT Community, Students, Faculty

Events By School

School of Engineering (SoE), School of Science


MIT Quest for Intelligence

#AI, machine learning, posters

Contact Email

Add to my calendar

Recent Activity