• Aucun résultat trouvé

Deep Learning: Past, present and future

N/A
N/A
Protected

Academic year: 2021

Partager "Deep Learning: Past, present and future"

Copied!
30
0
0

Texte intégral

(1)

Deep Learning

Past, present and future

Prof. Gilles Louppe

(2)

Past

(3)

How to write a computer program that performs tasks what we can all easily do, yet all fail to describe precisely how?

(4)

Perceptron

The Perceptron (Rosenblatt, 1957)

f

(x; w, b) = σ(w x + b)

T

(5)
(6)

Convolutional networks

Hubel and Wiesel, 1962

(7)
(8)

Learning

θ

t+1

= θ

t

− γ∇

θ

L

t

)

(9)
(10)

Present

(11)
(12)
(13)

Algorithms Data

Software Compute engines

(14)

Depth

Szegedy et al, 2014

(15)

Pixel data

(e.g., visual recognition)

Audio data

(e.g., speech recognition and synthesis)

Text data

(e.g., machine translation)

System applications (e.g., databases)

(16)

Adversarial training Generative models

Few-shot learning Learning to learn

Beyond supervised learning

(17)

NVIDIA DRIVE Autonomous Vehicle Platform

(18)

DQN Breakout

Learning to play video games (Mnih et al, 2013)

(19)
(20)

Andrej Karpathy (Director of AI, Tesla, 2017) Neural networks are not just another classi er, they represent the beginning of a fundamental shift in how we write software. They are Software 2.0.

(21)

Software 1.0

Software 1.0

Programs are written in languages such as Python, C or Java.

They consist of explicit instructions to the computer written by a programmer. The programmer identi es a speci c point in program space with some

(22)

Software 1.0 Software 2.0

Software 2.0

Programs are written in neural network weights

No human is involved in writing those weights!

Instead, specify constraints on the behavior of a desirable program (e.g., through data).

Search the program space through optimization.

(23)

For many real-world problems, it is often signi cantly easier to collect the data than to explicitly write the program.

Therefore,

programmers of tomorrow do not maintain complex software repositories, write intricate programs or analyze their running times.

Instead, programmers become teachers. They collect, clean, manipulate, label, analyze and visualize the data that feeds neural nets.

(24)

(Jeff Dean, Lead of Google.ai, 2017)

(25)

Bene ts

Computationally homogeneous Simple to bake in silicon

Constant running time and memory use It is highly portable

It is very agile

(26)

Modules can meld into an optimal whole (Jeff Dean, Lead of Google.ai, 2017)

(27)

How do you trust systems made of opaque neural networks, for which domain knowledge seems to have disappeared?

interpretability issues accountability issues security issues

(28)

Domain knowledge should not be abandoned. Instead, use it to design neural networks, thereby gaining in understanding and trust.

(29)

Summary

Past: Deep Learning has a long history, fueled by contributions from neuroscience, control and computer science.

Present: It is now mature enough to enable applications with super-human

level performance, as already illustrated in many engineering disciplines.

Future: Neural networks are not just another classi er. Sooner than later, they will take over increasingly large portions of what Software 1.0 is responsible for today.

(30)

The end.

Références

Documents relatifs

Biotech- nological strategies for the recovery of valuable and critical raw materials from waste electrical and electronic equipment (WEEE) – A review... van Hullebusch a , c ,

Analyser le niveau de risque des activités illégales dans l’Unité Forestière d'Aménagement (UFA) 10009 appartenant à la Société d'Exploitation de BOIS d'Afrique

2013 made an assessment of impact of projected climate change on the hydrologic regime and climate extremes of the Congo River basin.. This specific river basin, despite its

multiple folded paths. the surface wave energy where needed to compensate for propagation losses by the method suggested in figure 13. Here some of the surface wave

The software receives from the data analyst the training session parameters (mean, standard deviation, number of epochs, number of images per mini batch, learning rate

The creation of the large TIPSTER collection in 1990, followed by the first Text REtrieval Conference (TREC) in 1992 reframed the shared concept to mean not only using the same

Today, we are growing fast in Asia with full industry chain.. Decathlon Asia IT is located

Tense is the grammatical term that refers to the time when the action of the verb occurs: past, present, future.. The time frame of an action is usually established by referring to