Yuan Cao

Affiliations:
  • University of California, Los Angeles, Department of Computer Science, CA, USA
  • Princeton University, Department of Operations Research and Financial Engineering, NJ, USA (PhD)


According to our database1, Yuan Cao authored at least 29 papers between 2018 and 2023.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2023
Benign Overfitting in Two-Layer ReLU Convolutional Neural Networks for XOR Data.
CoRR, 2023

Per-Example Gradient Regularization Improves Learning Signals from Noisy Data.
CoRR, 2023

Benign Overfitting in Adversarially Robust Linear Classification.
Proceedings of the Uncertainty in Artificial Intelligence, 2023

The Benefits of Mixup for Feature Learning.
Proceedings of the International Conference on Machine Learning, 2023

Understanding Train-Validation Split in Meta-Learning with Neural Networks.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

Understanding the Generalization of Adam in Learning Neural Networks with Proper Regularization.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

How Does Semi-supervised Learning with Pseudo-labelers Work? A Case Study.
Proceedings of the Eleventh International Conference on Learning Representations, 2023

The Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks.
Proceedings of the Thirty Sixth Annual Conference on Learning Theory, 2023

2022
Benign Overfitting in Two-layer Convolutional Neural Networks.
CoRR, 2022

Benign Overfitting in Two-layer Convolutional Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

2021
Risk Bounds for Over-parameterized Maximum Margin Classification on Sub-Gaussian Mixtures.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Towards Understanding the Spectral Bias of Deep Learning.
Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, 2021

Provable Generalization of SGD-trained Neural Networks of Any Width in the Presence of Adversarial Label Noise.
Proceedings of the 38th International Conference on Machine Learning, 2021

Agnostic Learning of Halfspaces with Gradient Descent via Soft Margins.
Proceedings of the 38th International Conference on Machine Learning, 2021

How Much Over-parameterization Is Sufficient to Learn Deep ReLU Networks?
Proceedings of the 9th International Conference on Learning Representations, 2021

2020
Gradient descent optimizes over-parameterized deep ReLU networks.
Mach. Learn., 2020

Mean-Field Analysis of Two-Layer Neural Networks: Non-Asymptotic Rates and Generalization Bounds.
CoRR, 2020

Agnostic Learning of a Single Neuron with Gradient Descent.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Closing the Generalization Gap of Adaptive Gradient Methods in Training Deep Neural Networks.
Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, 2020

Accelerated Factored Gradient Descent for Low-Rank Matrix Factorization.
Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, 2020

Generalization Error Bounds of Gradient Descent for Learning Over-Parameterized Deep ReLU Networks.
Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, 2020

2019
A Generalization Theory of Gradient Descent for Learning Over-parameterized Deep ReLU Networks.
CoRR, 2019

Algorithm-Dependent Generalization Bounds for Overparameterized Deep Residual Networks.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

2018
Stochastic Gradient Descent Optimizes Over-parameterized Deep ReLU Networks.
CoRR, 2018

On the Convergence of Adaptive Gradient Methods for Nonconvex Optimization.
CoRR, 2018

The Edge Density Barrier: Computational-Statistical Tradeoffs in Combinatorial Inference.
Proceedings of the 35th International Conference on Machine Learning, 2018


  Loading...