Award Abstract # 2031883
Collaboration on the Theoretical Foundations of Deep Learning

NSF Org: DMS
Division Of Mathematical Sciences
Recipient: REGENTS OF THE UNIVERSITY OF CALIFORNIA, THE
Initial Amendment Date: August 24, 2020
Latest Amendment Date: August 30, 2023
Award Number: 2031883
Award Instrument: Continuing Grant
Program Manager: Christopher Stark
cstark@nsf.gov
 (703)292-4869
DMS
 Division Of Mathematical Sciences
MPS
 Direct For Mathematical & Physical Scien
Start Date: September 1, 2020
End Date: August 31, 2025 (Estimated)
Total Intended Award Amount: $5,000,000.00
Total Awarded Amount to Date: $4,000,000.00
Funds Obligated to Date: FY 2020 = $2,000,000.00
FY 2021 = $500,000.00

FY 2022 = $500,000.00

FY 2023 = $1,000,000.00
History of Investigator:
  • Peter Bartlett (Principal Investigator)
    bartlett@stat.berkeley.edu
  • Roman Vershynin (Co-Principal Investigator)
  • Bin Yu (Co-Principal Investigator)
  • Andrea Montanari (Co-Principal Investigator)
  • Alexander Rakhlin (Co-Principal Investigator)
Recipient Sponsored Research Office: University of California-Berkeley
1608 4TH ST STE 201
BERKELEY
CA  US  94710-1749
(510)643-3891
Sponsor Congressional District: 12
Primary Place of Performance: University of California-Berkeley
Sponsored Projects Office
Berkeley
CA  US  94710-1749
Primary Place of Performance
Congressional District:
12
Unique Entity Identifier (UEI): GS3YEVSS12N6
Parent UEI:
NSF Program(s): OFFICE OF MULTIDISCIPLINARY AC,
Special Projects - CCF,
MATHEMATICAL SCIENCES RES INST,
EPCN-Energy-Power-Ctrl-Netwrks
Primary Program Source: 01002021DB NSF RESEARCH & RELATED ACTIVIT
01002122DB NSF RESEARCH & RELATED ACTIVIT

01002223DB NSF RESEARCH & RELATED ACTIVIT

01002324DB NSF RESEARCH & RELATED ACTIVIT
Program Reference Code(s): 075Z, 079Z
Program Element Code(s): 125300, 287800, 733300, 760700
Award Agency Code: 4900
Fund Agency Code: 4900
Assistance Listing Number(s): 47.041, 47.049, 47.070

ABSTRACT

The success of deep learning has had a major impact across industry, commerce, science and society. But there are many aspects of this technology that are very different from classical methodology and that are poorly understood. Gaining a theoretical understanding will be crucial for overcoming its drawbacks. The Collaboration on the Theoretical Foundations of Deep Learning aims to address these challenges: understanding the mathematical mechanisms that underpin the practical success of deep learning, using this understanding to elucidate the limitations of current methods and extending them beyond the domains where they are currently applicable, and initiating the study of the array of mathematical problems that emerge. The team has planned a range of mechanisms to facilitate collaboration, including teleconference and in-person research meetings, a centrally organized postdoc program, and a program for visits between institutions by postdocs and graduate students. Research outcomes from the collaboration have strong potential to directly impact the many application domains for deep learning. The project will also have broad impacts through its education, human resource development and broadening participation programs, in particular through training a diverse cohort of graduate students and postdocs using an approach that emphasizes strong mentorship, flexibility, and breadth of collaboration opportunities; through an annual summer school that will deliver curriculum in the theoretical foundations of deep learning to a diverse group of graduate students, postdocs, and junior faculty; and through targeting broader participation in the collaboration?s research workshops and summer schools.

The collaboration?s research agenda is built on the following hypotheses: that overparametrization allows efficient optimization; that interpolation with implicit regularization enables generalization; and that depth confers representational richness through compositionality. The team aims to formulate and rigorously study these hypotheses as general mathematical phenomena, with the objective of understanding deep learning, extending its applicability, and developing new methods. Beyond enabling the development of improved deep learning methods based on principled design techniques, understanding the mathematical mechanisms that underlie the success of deep learning will also have repercussions on statistics and mathematics, including a new point of view of classical statistical methods, such as reproducing kernel Hilbert spaces and decision forests, and new research directions in nonlinear matrix theory and in understanding random landscapes. In addition, the research workshops that the collaboration will organize will be open to the public and will serve the broader research community in addressing these key challenges.

This award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.

PUBLICATIONS PRODUCED AS A RESULT OF THIS RESEARCH

Note:  When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

(Showing: 1 - 10 of 46)
Stephan, Ludovic and Zhu, Yizhe "Sparse random hypergraphs: Non-backtracking spectra and community detection" ArXivorg , 2022 Citation Details
Zhu, Yizhe "On the Second Eigenvalue of Random Bipartite Biregular Graphs" Journal of Theoretical Probability , 2022 https://doi.org/10.1007/s10959-022-01190-0 Citation Details
Wang, Zhichao and Zhu, Yizhe "Deformed semicircle law and concentration of nonlinear random matrices for ultra-wide neural networks" ArXivorg , 2022 Citation Details
Dover, Kathryn and Cang, Zixuan and Ma, Anna and Nie, Qing and Vershynin, Roman "AVIDA: Alternating method for Visualizing and Integrating Data" Journal of computational science , v.68 , 2023 Citation Details
Ma, Anna and Stöger, Dominik and Zhu, Yizhe "Robust recovery of low-rank matrices and low-tubal-rank tensors from noisy sketches" ArXivorg , 2022 https://doi.org/10.48550/arXiv.2206.00803 Citation Details
Lybrand, Eric and Ma, Anna and Saab, Rayan "On the number of faces and radii of cells induced by Gaussian spherical tessellations" ArXivorg , 2022 Citation Details
Ferber, Asaf and Sah, Ashwin and Sawhney, Mehtaab and Zhu, Yizhe "Sparse recovery properties of discrete random matrices" Combinatorics probability and computing , v.32 , 2023 Citation Details
Boedihardjo, March and Strohmer, Thomas and Vershynin, Roman "Privacy of synthetic data: A statistical framework" ArXivorg , 2021 Citation Details
Boedihardjo, March and Strohmer, Thomas and Vershynin, Roman "Covariance's Loss is Privacy's Gain: Computationally Efficient, Private and Accurate Synthetic Data" ArXivorg , 2021 Citation Details
Ronen, Omer and Saarinen, Theo and Tan, Yan Shuo and Duncan, James and Yu, Bin "A Mixing Time Lower Bound for a Simplified Version of BART" arXivorg , 2022 Citation Details
Tan, Yan Shuo and Singh, Chandan and Nasseri, Keyan and Agarwal, Abhineet and Duncan, James and Ronen, Omer and Epland, Matthew and Kornblith, Aaron and Yu, Bin "Fast Interpretable Greedy-Tree Sums (FIGS)" ArXivorg , 2023 Citation Details
(Showing: 1 - 10 of 46)

Please report errors in award information by writing to: awardsearch@nsf.gov.

Print this page

Back to Top of page