Loss Spikes in Gradient Descent
Loss spikes aren’t noise. They’re gradient descent briefly exceeding the edge of stability and snapping back. Here’s why.
Loss spikes aren’t noise. They’re gradient descent briefly exceeding the edge of stability and snapping back. Here’s why.
L1 regression is more robust to outliers than least squares, but harder to solve. We walk through four algorithms, each addressing a limitation of the previo...
Golden section search reuses objective evaluations to efficiently minimize 1D functions. Learn how this classical algorithm connects to the golden ratio and ...
Derive and interpret the dual form of an optimization problem.
Solve constrained optimization problems using your favorite unconstrained solver.
Fit nonlinear models using Gauss-Newton and Levenberg-Marquardt algorithms.
A ground-up look at how floating-point and fixed-point numbers are represented and how arithmetic works for each
Interpolate equally-spaced data efficiently and discover its connection to Taylor series.
Tackle tricky integrals with endpoint singularities using a clever variable transformation.
Approximate functions using polynomial interpolation without solving linear systems.
Compute logarithms and exponentials without a floating point unit.
Compute sine, cosine, and exponentials using only addition, subtraction, and bit shifts.
Fit nonlinear models using Gauss-Newton and Levenberg-Marquardt algorithms.
Plot ellipses using conic, quadratic, and parametric representations.
Implement a regression tree from scratch using only numpy.
Create cleaner 3D surface plots using radial and elliptical grids in matplotlib.
Interpolate equally-spaced data efficiently and discover its connection to Taylor series.
Fit rational functions to data with poles and discontinuities where polynomials fail.
Approximate functions using polynomial interpolation without solving linear systems.
Interpolate equally-spaced data efficiently and discover its connection to Taylor series.
Fit rational functions to data with poles and discontinuities where polynomials fail.
Approximate functions using polynomial interpolation without solving linear systems.
A ground-up look at how floating-point and fixed-point numbers are represented and how arithmetic works for each
Compute logarithms and exponentials without a floating point unit.
Compute sine, cosine, and exponentials using only addition, subtraction, and bit shifts.
A ground-up look at how floating-point and fixed-point numbers are represented and how arithmetic works for each
Compute logarithms and exponentials without a floating point unit.
Compute sine, cosine, and exponentials using only addition, subtraction, and bit shifts.
Fit rational functions to data with poles and discontinuities where polynomials fail.
Fit nonlinear models using Gauss-Newton and Levenberg-Marquardt algorithms.
L1 regression is more robust to outliers than least squares, but harder to solve. We walk through four algorithms, each addressing a limitation of the previo...
Fit nonlinear models using Gauss-Newton and Levenberg-Marquardt algorithms.
Plot ellipses using conic, quadratic, and parametric representations.
Create cleaner 3D surface plots using radial and elliptical grids in matplotlib.
Plot ellipses using conic, quadratic, and parametric representations.
Create cleaner 3D surface plots using radial and elliptical grids in matplotlib.
Derive and interpret the dual form of an optimization problem.
Solve constrained optimization problems using your favorite unconstrained solver.
Loss spikes aren’t noise. They’re gradient descent briefly exceeding the edge of stability and snapping back. Here’s why.
L1 regression is more robust to outliers than least squares, but harder to solve. We walk through four algorithms, each addressing a limitation of the previo...
Implement a regression tree from scratch using only numpy.
Implement a regression tree from scratch using only numpy.
Plot ellipses using conic, quadratic, and parametric representations.
Derive and interpret the dual form of an optimization problem.
Tackle tricky integrals with endpoint singularities using a clever variable transformation.
Tackle tricky integrals with endpoint singularities using a clever variable transformation.
Fit rational functions to data with poles and discontinuities where polynomials fail.
Compute sine, cosine, and exponentials using only addition, subtraction, and bit shifts.
Interpolate equally-spaced data efficiently and discover its connection to Taylor series.
Detect nonlinear relationships that Pearson and Spearman miss.
Detect nonlinear relationships that Pearson and Spearman miss.
Detect nonlinear relationships that Pearson and Spearman miss.
A ground-up look at how floating-point and fixed-point numbers are represented and how arithmetic works for each
Loss spikes aren’t noise. They’re gradient descent briefly exceeding the edge of stability and snapping back. Here’s why.