Jason D. Lee
Jason D. Lee
Assistant Professor of Electrical Engineering, Princeton University
Verified email at princeton.edu - Homepage
Cited by
Cited by
Gradient descent only converges to minimizers
JD Lee, M Simchowitz, MI Jordan, B Recht
Conference on learning theory, 1246-1257, 2016
Exact post-selection inference, with application to the lasso
JD Lee, DL Sun, Y Sun, JE Taylor
The Annals of Statistics 44 (3), 907-927, 2016
Gradient descent finds global minima of deep neural networks
SS Du, JD Lee, H Li, L Wang, X Zhai
arXiv preprint arXiv:1811.03804, 2018
Matrix completion has no spurious local minimum
R Ge, JD Lee, T Ma
arXiv preprint arXiv:1605.07272, 2016
Matrix completion and low-rank SVD via fast alternating least squares
T Hastie, R Mazumder, J Lee, R Zadeh
Journal of Machine Learning Research, 2014
Theoretical insights into the optimization landscape of over-parameterized shallow neural networks
M Soltanolkotabi, A Javanmard, JD Lee
IEEE Transactions on Information Theory 65 (2), 742-769, 2018
A kernelized Stein discrepancy for goodness-of-fit tests
Q Liu, J Lee, M Jordan
International conference on machine learning, 276-284, 2016
Proximal Newton-type methods for minimizing composite functions
JD Lee, Y Sun, MA Saunders
SIAM Journal on Optimization 24 (3), 1420-1443, 2014
Learning one-hidden-layer neural networks with landscape design
R Ge, JD Lee, T Ma
arXiv preprint arXiv:1711.00501, 2017
Implicit bias of gradient descent on linear convolutional networks
S Gunasekar, J Lee, D Soudry, N Srebro
arXiv preprint arXiv:1806.00468, 2018
Characterizing implicit bias in terms of optimization geometry
S Gunasekar, J Lee, D Soudry, N Srebro
International Conference on Machine Learning, 1832-1841, 2018
Communication-efficient distributed statistical inference
MI Jordan, JD Lee, Y Yang
Journal of the American Statistical Association, 2018
Learning the structure of mixed graphical models
JD Lee, TJ Hastie
Journal of Computational and Graphical Statistics 24 (1), 230-253, 2015
Gradient descent learns one-hidden-layer cnn: Don’t be afraid of spurious local minima
S Du, J Lee, Y Tian, A Singh, B Poczos
International Conference on Machine Learning, 1339-1348, 2018
On the power of over-parametrization in neural networks with quadratic activation
S Du, J Lee
International Conference on Machine Learning, 1329-1338, 2018
Optimality and approximation with policy gradient methods in markov decision processes
A Agarwal, SM Kakade, JD Lee, G Mahajan
Conference on Learning Theory, 64-66, 2020
Practical large-scale optimization for max-norm regularization
J Lee, B Recht, RR Salakhutdinov, N Srebro, JA Tropp
Neural Information Processing Systems, 2010
Gradient descent can take exponential time to escape saddle points
SS Du, C Jin, JD Lee, MI Jordan, B Poczos, A Singh
arXiv preprint arXiv:1705.10412, 2017
Solving a class of non-convex min-max games using iterative first order methods
M Nouiehed, M Sanjabi, T Huang, JD Lee, M Razaviyayn
arXiv preprint arXiv:1902.08297, 2019
Stochastic subgradient method converges on tame functions
D Davis, D Drusvyatskiy, S Kakade, JD Lee
Foundations of computational mathematics 20 (1), 119-154, 2020
The system can't perform the operation now. Try again later.
Articles 1–20