DeepLearning L1 Intro
DeepLearning L1 Intro
A.U : 2022/2023
What is Deep Learning ?
Why Deep Learning and Why Now ?
Why Deep Learning ?
Hand engineered features are time consuming, brittle, and not scalable in practice.
𝑥2 = Hours
spent on the
final project
𝑥2 = Hours
spent on the
final project
How does a small change in one weight (ex. 𝑤2 ) affect the final loss 𝐽 𝑊 ?
Computing gradients: Backpropagation
Computing gradients: Backpropagation
Computing gradients: Backpropagation
Computing gradients: Backpropagation
Repeat this for every weight in the network using gradients from later layers
Computing gradients: Backpropagation
Neural Networks in Practice:
Optimization
Training Neural Networks is difficult
Idea 1:
Try lots of different learning rates and see what works « just right »
Comment gérer cela?
Idée 1:
essayez de nombreux taux d'apprentissage différents et voyez ce qui fonctionne « juste »
Idée 2:
Do something smarter!
Design and adaptive learning rate that adapts to the landscape
Adaptive Learning Rates
what is it?
Technique that constrains our optimization problem to discourage complex models
Regularization
C’est quoi?
Technique qui contraint notre problème d'optimisation à décourager les modèles complexes
Why?
Improve generalisation of our model on unseen data
Regularization 1: Dropout
Regularization 1: Dropout
Regularization 1: Dropout
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Regularization 2: Early Stopping
Résumé: les fondations de base