Samuel S. Schoenholz

Affiliations:
  • OpenAI, San Francisco, CA, USA


According to our database1, Samuel S. Schoenholz authored at least 37 papers between 2009 and 2023.

Collaborative distances:

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2023
Temperature check: theory and practice for training models with softmax-cross-entropy losses.
Trans. Mach. Learn. Res., 2023

Scaling deep learning for materials discovery.
Nat., 2023

End-to-End Differentiable Reactive Molecular Dynamics Simulations Using JAX.
Proceedings of the High Performance Computing - 38th International Conference, 2023

2022
∂<i>PV</i>: An end-to-end differentiable solar-cell simulator.
Comput. Phys. Commun., 2022

What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries.
CoRR, 2022

Fast Finite Width Neural Tangent Kernel.
Proceedings of the International Conference on Machine Learning, 2022

Deep equilibrium networks are sensitive to initialization statistics.
Proceedings of the International Conference on Machine Learning, 2022

2021
Gradients are Not All You Need.
CoRR, 2021

Rapid training of deep neural networks without skip connections or normalization layers using Deep Kernel Shaping.
CoRR, 2021

Whitening and Second Order Optimization Both Make Information in the Dataset Unusable During Training, and Can Reduce or Prevent Generalization.
Proceedings of the 38th International Conference on Machine Learning, 2021

Tilting the playing field: Dynamical loss functions for machine learning.
Proceedings of the 38th International Conference on Machine Learning, 2021

Learn2Hop: Learned Optimization on Rough Landscapes.
Proceedings of the 38th International Conference on Machine Learning, 2021

2020
Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible.
CoRR, 2020

On the infinite width limit of neural networks with a standard parameterization.
CoRR, 2020

JAX MD: A Framework for Differentiable Physics.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Finite Versus Infinite Neural Networks: an Empirical Study.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Disentangling Trainability and Generalization in Deep Neural Networks.
Proceedings of the 37th International Conference on Machine Learning, 2020

Neural Tangents: Fast and Easy Infinite Neural Networks in Python.
Proceedings of the 8th International Conference on Learning Representations, 2020

2019
Disentangling trainability and generalization in deep learning.
CoRR, 2019

Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent.
CoRR, 2019

Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs.
CoRR, 2019

Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

MetaInit: Initializing learning by learning to initialize.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

A Mean Field Theory of Batch Normalization.
Proceedings of the 7th International Conference on Learning Representations, 2019

2018
Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10, 000-Layer Vanilla Convolutional Neural Networks.
Proceedings of the 35th International Conference on Machine Learning, 2018

Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks.
Proceedings of the 35th International Conference on Machine Learning, 2018

Deep Neural Networks as Gaussian Processes.
Proceedings of the 6th International Conference on Learning Representations, 2018

Adversarial Spheres.
Proceedings of the 6th International Conference on Learning Representations, 2018

Intriguing Properties of Adversarial Examples.
Proceedings of the 6th International Conference on Learning Representations, 2018

The emergence of spectral universality in deep networks.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2018

2017
A Correspondence Between Random Neural Networks and Statistical Field Theory.
CoRR, 2017

Mean Field Residual Networks: On the Edge of Chaos.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Resurrecting the sigmoid in deep learning through dynamical isometry: theory and practice.
Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017

Neural Message Passing for Quantum Chemistry.
Proceedings of the 34th International Conference on Machine Learning, 2017

Deep Information Propagation.
Proceedings of the 5th International Conference on Learning Representations, 2017

Explaining the Learning Dynamics of Direct Feedback Alignment.
Proceedings of the 5th International Conference on Learning Representations, 2017

2009
Specialization as an optimal strategy under varying external conditions.
Proceedings of the 2009 IEEE International Conference on Robotics and Automation, 2009


  Loading...