Skip directly to content

Minimize RSR Award Detail

Research Spending & Results

Award Detail

Awardee:MASSACHUSETTS INSTITUTE OF TECHNOLOGY
Doing Business As Name:Massachusetts Institute of Technology
PD/PI:
  • Tomaso Poggio
  • (617) 253-5230
  • tp@ai.mit.edu
Co-PD(s)/co-PI(s):
  • Constantinos Daskalakis
  • Stefanie Jegelka
  • Aleksander Madry
Award Date:09/15/2021
Estimated Total Award Amount: $ 600,000
Funds Obligated to Date: $ 600,000
  • FY 2021=$600,000
Start Date:12/01/2021
End Date:11/30/2024
Transaction Type:Grant
Agency:NSF
Awarding Agency Code:4900
Funding Agency Code:4900
CFDA Number:47.049
Primary Program Source:040100 NSF RESEARCH & RELATED ACTIVIT
Award Title or Description:Collaborative Research: Foundations of Deep Learning: Theory, Robustness, and the Brain​
Federal Award ID Number:2134108
DUNS ID:001425594
Parent DUNS ID:001425594
Program:CDS&E-MSS
Program Officer:
  • Tracy Kimbrel
  • (703) 292-7924
  • tkimbrel@nsf.gov

Awardee Location

Street:77 MASSACHUSETTS AVE
City:Cambridge
State:MA
ZIP:02139-4301
County:Cambridge
Country:US
Awardee Cong. District:07

Primary Place of Performance

Organization Name:Massachusetts Institute of Technology
Street:77 Massachusetts Avenue
City:Cambridge
State:MA
ZIP:02139-4301
County:Cambridge
Country:US
Cong. District:07

Abstract at Time of Award

A truly comprehensive theory of machine learning has the potential of informing science and engineering in the same profound way Maxwell’s equations did. It was the development of that theory by Maxwell that truly unleashed the potential of electricity, leading to radio, radars, computers, and the Internet. In an analogy, deep learning (DL) has found over the past decade many applications, so far without a comprehensive theory. An eventual theory of learning that explains why and how deep networks work and what their limitations are may thus enable the development of even more powerful learning approaches – especially if the goal of reconnecting DL to brain research bears fruit. In the long term, the ability to develop and build better intelligent machines will be essential to any technology-based economy. After all, even in its current – still highly imperfect – state, DL is impacting or about to impact just about every aspect of our society and life. The investigators also plan to complement their theoretical research with the educational goal of training a diverse population of young researchers from mathematics, computer science, statistics, electrical engineering, and computational neuroscience in the field of machine learning and of its theoretical underpinnings. The investigators propose to join forces in a multi-pronged and collaborative assault on the profound mysteries of DL, informed by the sum of their experience, expertise, ideas, and insight. The research goals are threefold: to develop a sound foundational/mathematical understanding of DL; in doing so to advance the foundational understanding of learning more generally; and to advance the practice of DL by addressing its above-mentioned weaknesses. Of six foundational thrusts, the first two focus on the standard decomposition of the prediction error in approximation and sample (or estimation) error. Their goal is to extend classical results in approximation theory and theory of learnability to DL. These two are then supported by a research project that is specific to deep learning: analysis of the dynamics of gradient descent in training a network. The fourth theme is about robustness against adversaries and shifts, a powerful test for theories which is also important for practical deployment of learning systems. The fifth thrust is about developing the theory of control through DL, as well as exploring dynamical systems aspects of deep reinforcement learning. The final topic connects research on DL to its origins - and possibly its future: networks of neurons in the brain. The proposed research also promises to advance the foundations of learning theory. Success in this project will result in sharper mathematical techniques for machine learning and comprehensive foundations of machine learning robustness, broadly construed. It will also ultimately enable development of learning algorithms that transcend deep learning and guide the way towards creating more intelligent machines, and shed new light on our own intelligence. This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

For specific questions or comments about this information including the NSF Project Outcomes Report, contact us.