Deep Learning Specialization - Coursera
main
main
  • Introduction
  • Neural Networks and Deep Learning
    • Introduction to Deep Learning
    • Logistic Regression as a Neural Network (Neural Network Basics)
    • Shallow Neural Network
    • Deep Neural Network
  • Improving Deep Neural Networks: Hyperparameter tuning, Regularization and Optimization
    • Practical Aspects of Deep Learning
    • Optimization Algorithms
    • Hyperparameter Tuning, Batch Normalization and Programming Frameworks
  • Structuring Machine Learning Projects
    • Introduction to ML Strategy
    • Setting Up Your Goal
    • Comparing to Human-Level Performance
    • Error Analysis
    • Mismatched Training and Dev/Test Set
    • Learning from Multiple Tasks
    • End-to-End Deep Learning
  • Convolutional Neural Networks
    • Foundations of Convolutional Neural Networks
    • Deep Convolutional Models: Case Studies
      • Classic Networks
      • ResNets
      • Inception
    • Advice for Using CNNs
    • Object Detection
      • Object Localization
      • Landmark Detection
      • Sliding Window Detection
      • The YOLO Algorithm
      • Intersection over Union
      • Non-Max Suppression
      • Anchor Boxes
      • Region Proposals
    • Face Recognition
      • One-Shot Learning
      • Siamese Network
      • Face Recognition as Binary Classification
    • Neural Style Transfer
  • Sequence Models
    • Recurrent Neural Networks
      • RNN Structure
      • Types of RNNs
      • Language Modeling
      • Vanishing Gradient Problem in RNNs
      • Gated Recurrent Units (GRUs)
      • Long Short-Term Memory Network (LSTM)
      • Bidirectional RNNs
    • Natural Language Processing & Word Embeddings
      • Introduction to Word Embeddings
      • Learning Word Embeddings: Word2Vec and GloVe
      • Applications using Word Embeddings
      • De-Biasing Word Embeddings
    • Sequence Models & Attention Mechanisms
      • Sequence to Sequence Architectures
        • Basic Models
        • Beam Search
        • Bleu Score
        • Attention Model
      • Speech Recognition
Powered by GitBook
On this page
  • What is a Neural Network?
  • Supervised Learning with Neural Networks
  • Structured vs. Unstructured Data
  • Why is Deep Learning Taking Off?

Was this helpful?

  1. Neural Networks and Deep Learning

Introduction to Deep Learning

PreviousNeural Networks and Deep LearningNextLogistic Regression as a Neural Network (Neural Network Basics)

Last updated 4 years ago

Was this helpful?

Deep Learning refers to the training of large neural networks (with multiple hidden layers) using vast amounts of training data.

What is a Neural Network?

Neural Networks are Machine Learning models inspired by the vast interconnection of biological neurons in the human brain. These networks consist of input and output layers, and one or more hidden layers. Each layer contains several nodes/neurons. Every node in the input layer is connected to every node in the hidden layer, and so on.

Deep Neural Networks have multiple hidden layers.

Neural Networks are incredibly good at mapping the input variables x to the target variable y, if given enough training data and sufficient information about x and y.

The image below depicts a simple neural network to predict house prices, given certain features of the house:

Supervised Learning with Neural Networks

Supervised Learning is a type of Machine Learning paradigm wherein the output labels are known in advance i.e. before we train the model. There are mainly two Supervised Learning techniques:

  • Classification: The target variable has discrete values (yes/no, 0/1, etc.)

  • Regression: The target variable has continuous values (home prices, etc.)

Some common applications of Supervised Learning techniques:

Input (x)

Output (y)

Application

Type of Neural Network

Home Features

Price

Real Estate

Standard NN

Ad, User Info

Click on Ad? (0/1)

Online Advertising

Standard NN

Image

Object (1, ..., 1000)

Photo Tagging

Convolutional NN

Audio

Text Transcript

Speech Recognition

Recurrent NN

English

Chinese

Machine Translation

Recurrent NN

Image, Radar Info

Position of other Cars

Autonomous Driving

Hybrid/Custom NN

Note: RNNs are mainly used for sequence data/time series data, i.e. data that exhibits a temporal property.

Structured vs. Unstructured Data

Structured data refers to data that has a well-defined meaning and can usually be stored in organized databases. Examples include prices, ages, etc.

Unstructured data refers to data that doesn't have a well-defined meaning. Examples include pixel data, audio data, etc.

Machines usually find it harder to make sense of unstructured data, when compared to structured data, but advances in neural networks are making computers capable of understanding unstructured data as well.

Why is Deep Learning Taking Off?

In the last few years, we have seen a tremendous increase in the amount of data available for training our ML models. The performance of traditional ML algorithms increases upto a point, when we increase the amount of data, but then plateaus because they are not able to take advantage of the large amounts of training data.

For Neural Networks, however, the performance increases as either of the following increase:

  • Size of the training data

  • Size of the Neural Network

As depicted by the following graph, we can see a change in performance only when there are large amounts of data, because for small amounts of data, performance depends more on feature engineering than the model architecture.

Another important factor that has led to improved performance of ML models is the improvements made to training algorithms. For example, using the ReLU activation function instead of the Sigmoid activation function speeds up the training of neural networks. Also, advances in GPU computing have accelerated the training of efficient Deep Learning models.

To summarize, the main drivers of Deep Learning success are:

  • More data

  • More efficient computation

  • Better algorithms