Aaron Defazio

Orcid: 0000-0002-8764-3986

According to our database1, Aaron Defazio authored at least 36 papers between 2012 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

Online presence:

On csauthors.net:

Bibliography

2024
Directional Smoothness and Gradient Methods: Convergence and Adaptivity.
CoRR, 2024

2023
When, Why and How Much? Adaptive Learning Rate Scheduling by Refinement.
CoRR, 2023

Prodigy: An Expeditiously Adaptive Parameter-Free Learner.
CoRR, 2023

MoMo: Momentum Models for Adaptive Learning Rates.
CoRR, 2023

Mechanic: A Learning Rate Tuner.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Learning-Rate-Free Learning by D-Adaptation.
Proceedings of the International Conference on Machine Learning, 2023

2022
A scaling calculus for the design and initialization of ReLU networks.
Neural Comput. Appl., 2022

A Momentumized, Adaptive, Dual Averaged Gradient Method.
J. Mach. Learn. Res., 2022

Compressed sensing with a jackknife and a bootstrap.
J. Data Sci. Stat. Vis., 2022

Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method.
CoRR, 2022

2021
Stochastic Polyak Stepsize with a Moving Target.
CoRR, 2021

Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization.
CoRR, 2021

Almost sure convergence rates for Stochastic Gradient Descent and Stochastic Heavy Ball.
Proceedings of the Conference on Learning Theory, 2021

The Power of Factorial Powers: New Parameter settings for (Stochastic) Optimization.
Proceedings of the Asian Conference on Machine Learning, 2021

2020
Dual Averaging is Surprisingly Effective for Deep Learning Optimization.
CoRR, 2020

Understanding the Role of Momentum in Non-Convex Optimization: Practical Insights from a Lyapunov Analysis.
CoRR, 2020

On the convergence of the Stochastic Heavy Ball Method.
CoRR, 2020

Factorial Powers for Stochastic Optimization.
CoRR, 2020

Advancing machine learning for MR image reconstruction with an open competition: Overview of the 2019 fastMRI challenge.
CoRR, 2020

MRI Banding Removal via Adversarial Training.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

End-to-End Variational Networks for Accelerated MRI Reconstruction.
Proceedings of the Medical Image Computing and Computer Assisted Intervention - MICCAI 2020, 2020

GrappaNet: Combining Parallel Imaging With Deep Learning for Multi-Coil MRI Reconstruction.
Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020

2019
Offset Masking Improves Deep Learning based Accelerated MRI Reconstructions.
CoRR, 2019

Scaling Laws for the Principled Design, Initialization and Preconditioning of ReLU Networks.
CoRR, 2019

On the Ineffectiveness of Variance Reduced Optimization for Deep Learning.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

On the Curved Geometry of Accelerated Optimization.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

2018
Controlling Covariate Shift using Equilibrium Normalization of Weights.
CoRR, 2018

fastMRI: An Open Dataset and Benchmarks for Accelerated MRI.
CoRR, 2018

2016
A Simple Practical Accelerated Method for Finite Sums.
Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, 2016

2015
New Optimisation Methods for Machine Learning.
CoRR, 2015

Non-Uniform Stochastic Average Gradient Method for Training Conditional Random Fields.
Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, 2015

2014
A Comparison of learning algorithms on the Arcade Learning Environment.
CoRR, 2014

SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives.
Proceedings of the Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, 2014

Finito: A faster, permutable incremental gradient method for big data problems.
Proceedings of the 31th International Conference on Machine Learning, 2014

2012
A Convex Formulation for Learning Scale-Free Networks via Submodular Relaxation.
Proceedings of the Advances in Neural Information Processing Systems 25: 26th Annual Conference on Neural Information Processing Systems 2012. Proceedings of a meeting held December 3-6, 2012

A Graphical Model Formulation of Collaborative Filtering Neighbourhood Methods with Fast Maximum Entropy Training.
Proceedings of the 29th International Conference on Machine Learning, 2012


  Loading...