Announcement

Acknowledgement

Dr. Hua Zhou’s slides

In the next two lectures, we discuss a general framework for learning, neural networks.

History and recent surge

From Wang and Raj (2017) “On the Origin of Deep Learning”:

The current AI wave came in 2012 when AlexNet (60 million parameters) cuts the error rate of ImageNet competition (classify 1.2 million natural images) by half.

Learning sources

Single layer neural network (SLP)

image source

Multi-layer neural network (MLP)

Expressivity of neural network

Universal approximation properties

Practical issues

Neural networks are not a fully automatic tool, as they are sometimes advertised; as with all statistical models, subject matter knowledge should and often be used to improve their performance.

Convolutional neural networks (CNN)

Sources: https://colah.github.io/posts/2014-07-Conv-Nets-Modular/

Example: handwritten digit recognition

Results (320 training cases, 160 test cases):

network links weights accuracy
net 1 2570 2570 80.0%
net 2 3124 3214 87.0%
net 3 1226 1226 88.5%
net 4 2266 1131 94.0%
net 5 5194 1060 98.4%

Net-5 and similar networks were state-of-the-art in early 1990s.

Example: image classification

Source: http://cs231n.github.io/convolutional-networks/