Data science

In the last two lecturs, we discuss a general framework for learning, neural networks.

History and recent surge

From Wang and Raj (2017):

The current AI wave came in 2012 when AlexNet (60 million parameters) cuts the error rate of ImageNet competition (classify 1.2 million natural images) by half.

Learning sources

Single layer neural network (SLP)

image source

Multi-layer neural network (MLP)

Expressivity of neural network

Universal approximation properties

Practical issues

Neural networks are not a fully automatic tool, as they are sometimes advertised; as with all statistical models, subject matter knowledge should and often be used to improve their performance.

Convolutional neural networks (CNN)

Sources: https://colah.github.io/posts/2014-07-Conv-Nets-Modular/

Example: handwritten digit recognition

Results (320 training cases, 160 test cases):   

| network | links | weights | accuracy |
|---------|-------|---------|----------|
| net 1   | 2570  | 2570    | 80.0%    |
| net 2   | 3124  | 3214    | 87.0%    |
| net 3   | 1226  | 1226    | 88.5%    |
| net 4   | 2266  | 1131    | 94.0%    |
| net 5   | 5194  | 1060    | 98.4%    |  

Net-5 and similar networks were start-of-the-art in early 1990s.

Example: image classification

Source: http://cs231n.github.io/convolutional-networks/

Recurrent neural networks (RNN)

Generative Adversarial Networks (GANs)

The coolest idea in deep learning in the last 20 years.
- Yann LeCun on GANs.