Skip to content

richakbee/Deep-Learning-Specialization

Repository files navigation

Deep-Learning-Specialization

deep learning specialization on coursera by deeplearning.ai

Course 1: Neural Networks and Deep Learning

Week 1: Introduction to deep learning

  • Welcome
  • What is a neural network?
  • Supervised Learning with Neural Networks
  • Why is Deep Learning taking off?
  • Geoffrey Hinton interview

Week 2:Neural Networks Basics

Learn to set up a machine learning problem with a neural network mindset. Learn to use vectorization to speed up your models.

  • Binary Classification
  • Logistic Regression
  • Logistic Regression Cost Function
  • Gradient Descent
  • Derivatives,More Derivative Examples
  • Computation graph ,Derivatives with a Computation Graph
  • Logistic Regression Gradient Descent
  • Gradient Descent on m Examples
  • Vectorization, More Vectorization Examples6m
  • Vectorizing Logistic Regression
  • Vectorizing Logistic Regression's Gradient Output
  • Broadcasting in Python
  • A note on python/numpy vectors
  • Explanation of logistic regression cost function
  • Pieter Abbeel interview
  • Assignment : Logistic Regression with a Neural Network mindset

Week 3:Shallow neural networks

Learn to build a neural network with one hidden layer, using forward propagation and backpropagation.

  • Neural Networks Overview, Neural Network Representation
  • Computing a Neural Network's Output
  • Vectorizing across multiple examples
  • Activation functions
  • Why do you need non-linear activation functions? ,Derivatives of activation functions
  • Gradient descent for Neural Networks
  • Backpropagation intuition
  • Random Initialization
  • Assignment :Planar data classification with one hidden layer

Week 4 :Deep Neural Networks

Understand the key computations underlying deep learning, use them to build and train deep neural networks, and apply it to computer vision.

Course 2: Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization

Week 1: Practical aspects of Deep Learning

Week 2:Optimization algorithms

  • Mini-batch gradient descent
  • Understanding mini-batch gradient descent
  • Exponentially weighted averages
  • Understanding exponentially weighted averages
  • Bias correction in exponentially weighted averages
  • Gradient descent with momentum
  • RMSprop
  • Adam optimization algorithm
  • Learning rate decay
  • The problem of local optima
  • Assignment : Optimization Methods for Neural Networkfrom scratch

Week 3:Hyperparameter tuning, Batch Normalization and Programming Frameworks (tensorflow)

  • Tuning process
  • Using an appropriate scale to pick hyperparameters
  • Hyperparameters tuning in practice: Pandas vs. Caviar
  • Normalizing activations in a network
  • Fitting Batch Norm into a neural network
  • Why does Batch Norm work?
  • Batch Norm at test time
  • Softmax Regression
  • Training a softmax classifier
  • Deep learning frameworks
  • TensorFlow
  • Assignment :Neural network using tensorflow

Course 3: Structuring Machine Learning Projects

Week 1: ML Strategy (1)

  • Why ML Strategy
  • Orthogonalization
  • Single number evaluation metric
  • Satisficing and Optimizing metric
  • Train/dev/test distributions
  • Size of the dev and test sets
  • When to change dev/test sets and metrics
  • Why human-level performance?
  • Avoidable bias
  • Understanding human-level performance,Surpassing human-level performance
  • Improving your model performance
  • Quiz (case study) : Bird recognition in the city of Peacetopia (case study)

Week 2: ML Strategy (2)

  • Carrying out error analysis
  • Cleaning up incorrectly labeled data
  • Build your first system quickly, then iterate
  • Training and testing on different distributions
  • Bias and Variance with mismatched data distributions
  • Addressing data mismatch
  • Transfer learning
  • Multi-task learning
  • What is end-to-end deep learning?
  • Whether to use end-to-end deep learning
  • Quiz (case study) :Autonomous driving (case study)

Course 4: Convolutional Neural Networks

Week 1: Foundations of Convolutional Neural Networks

Learn to implement the foundational layers of CNNs (pooling, convolutions) and to stack them properly in a deep network to solve multi-class image classification problems.

  • Computer Vision
  • Edge Detection Example
  • More Edge Detection
  • Padding
  • Strided Convolutions
  • Convolutions Over Volume
  • One Layer of a Convolutional Network
  • Simple Convolutional Network Example
  • Pooling Layers
  • CNN Example
  • Why Convolutions?
  • Assignment :Convolutional Neural Networks: Step by Step (from scratch)

Week 2:Deep convolutional models: case studies

Learn about the practical tricks and methods used in deep CNNs straight from the research papers.

Week 3:Object detection

Learn how to apply your knowledge of CNNs to one of the toughest but hottest field of computer vision: Object detection.

  • Object Localization
  • Landmark Detection
  • Object Detection
  • Convolutional Implementation of Sliding Windows
  • Bounding Box Predictions
  • Intersection Over Union
  • Non-max Suppression
  • Anchor Boxes
  • YOLO Algorithm
  • Assignment :Autonomous driving - Car detection

Week 4 :Special applications: Face recognition & Neural style transfer

Discover how CNNs can be applied to multiple fields, including art generation and face recognition. Implement your own algorithm to generate art and recognize faces!

Course 5: Sequence Models

Week 1: Recurrent Neural Networks

Learn about recurrent neural networks. This type of model has been proven to perform extremely well on temporal data. It has several variants including LSTMs, GRUs and Bidirectional RNNs, which you are going to learn about in this section.

  • Why sequence models
  • Notation
  • Recurrent Neural Network Model
  • Backpropagation through time
  • Different types of RNNs
  • Language model and sequence generation
  • Sampling novel sequences
  • Vanishing gradients with RNNs
  • Gated Recurrent Unit (GRU)
  • Long Short Term Memory (LSTM)
  • Bidirectional RNN
  • Deep RNNs5
  • Assignment :Building your Recurrent Neural Network - Step by Step(from scratch)

Week 2:Natural Language Processing & Word Embeddings

Natural language processing with deep learning is an important combination. Using word vector representations and embedding layers you can train recurrent neural networks with outstanding performances in a wide variety of industries. Examples of applications are sentiment analysis, named entity recognition and machine translation.

  • Word Representation
  • Using word embeddings
  • Properties of word embeddings
  • Embedding matrix
  • Learning word embeddings
  • Word2Vec
  • Negative Sampling
  • GloVe word vectors
  • Sentiment Classification
  • Debiasing word embeddings
  • Assignment :Operations on word vectors

Week 3:Sequence models & Attention mechanism

Sequence models can be augmented using an attention mechanism. This algorithm will help your model understand where it should

  • Tuning process
  • Basic Models
  • Picking the most likely sentence
  • Beam Search
  • Refinements to Beam Search
  • Error analysis in beam search
  • Bleu Score (optional)
  • Attention Model Intuition
  • Attention Model
  • Speech recognition
  • Trigger Word Detection

About

deep learning specialization on coursera by deeplearning.ai

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published