Picture for Samuel S. Schoenholz

Samuel S. Schoenholz

Shammie

What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries

Add code
Oct 11, 2022
Figure 1 for What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries
Figure 2 for What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries
Figure 3 for What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries
Figure 4 for What does a deep neural network confidently perceive? The effective dimension of high certainty class manifolds and their low confidence boundaries
Viaarxiv icon

Deep equilibrium networks are sensitive to initialization statistics

Add code
Jul 19, 2022
Figure 1 for Deep equilibrium networks are sensitive to initialization statistics
Figure 2 for Deep equilibrium networks are sensitive to initialization statistics
Figure 3 for Deep equilibrium networks are sensitive to initialization statistics
Figure 4 for Deep equilibrium networks are sensitive to initialization statistics
Viaarxiv icon

Fast Finite Width Neural Tangent Kernel

Add code
Jun 17, 2022
Figure 1 for Fast Finite Width Neural Tangent Kernel
Figure 2 for Fast Finite Width Neural Tangent Kernel
Figure 3 for Fast Finite Width Neural Tangent Kernel
Figure 4 for Fast Finite Width Neural Tangent Kernel
Viaarxiv icon

Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models

Add code
Jun 10, 2022
Viaarxiv icon

Gradients are Not All You Need

Add code
Nov 10, 2021
Figure 1 for Gradients are Not All You Need
Figure 2 for Gradients are Not All You Need
Figure 3 for Gradients are Not All You Need
Figure 4 for Gradients are Not All You Need
Viaarxiv icon

Rapid training of deep neural networks without skip connections or normalization layers using Deep Kernel Shaping

Add code
Oct 05, 2021
Viaarxiv icon

Tilting the playing field: Dynamical loss functions for machine learning

Add code
Feb 13, 2021
Figure 1 for Tilting the playing field: Dynamical loss functions for machine learning
Figure 2 for Tilting the playing field: Dynamical loss functions for machine learning
Figure 3 for Tilting the playing field: Dynamical loss functions for machine learning
Figure 4 for Tilting the playing field: Dynamical loss functions for machine learning
Viaarxiv icon

Finite Versus Infinite Neural Networks: an Empirical Study

Add code
Sep 08, 2020
Figure 1 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 2 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 3 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 4 for Finite Versus Infinite Neural Networks: an Empirical Study
Viaarxiv icon

Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible

Add code
Aug 25, 2020
Figure 1 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 2 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 3 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 4 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Viaarxiv icon

On the infinite width limit of neural networks with a standard parameterization

Add code
Jan 25, 2020
Figure 1 for On the infinite width limit of neural networks with a standard parameterization
Figure 2 for On the infinite width limit of neural networks with a standard parameterization
Figure 3 for On the infinite width limit of neural networks with a standard parameterization
Figure 4 for On the infinite width limit of neural networks with a standard parameterization
Viaarxiv icon