Learning digits with InPCA

Training a neural network: animation

Stages of training a convolutional neural network (CNN). Each point in this animation represents one of the 10,000 handwritten digits in the MNIST test data set. Each of the ten colors represents one of the ten digits (dark blue = 0, ..., yellow = 9, see summary figure below). The time evolution, or epoch, reflects the training of the neural network. At the initial frame, the untrained network is unable to classify images -- it has a 90% error rate, just as one would expect from random guessing. As the training progresses and the error rate decreases, the digits in the cloud begin to cluster. (The network was trained on a separate set of 55,000 images, so the clustering represents success in recognizing new digits.)

This animation makes use of a new tool, InPCA, for visualizing high-dimensional data (see Visualizing probabilistic models: Intensive Principal Component Analysis, by Katherine N. Quinn, Colin B. Clement, Francesco De Bernardis, Michael D. Niemack, and James P. Sethna.) Our method bypasses the `curse of dimensionality' that arises for large models with too much information by taking the limit of zero information using the 'replica trick'.

Training a neural network:Summary

Summary figure, showing digits.

References and Videos

More on sloppiness:

Short course on information geometry, sloppy models, and visualizing behavior in high dimensions

Last Modified: May 9, 2019

James P. Sethna, sethna@lassp.cornell.edu; This work supported by the Division of Materials Research of the U.S. National Science Foundation, through grant DMR-1719490.

Statistical Mechanics: Entropy, Order Parameters, and Complexity, now available at Oxford University Press (USA, Europe).