Deep Learning

What is deep learning?

Deep Learning is a subset of machine learning concerned with algorithms inspired by the structure and function of the brain called artificial neural networks.

ML vs DL

Deep LearningMachine Learning
DL is a subset of ML. It is based on neural networks, a conceptual model of the brain It is a subset of AI. It focuses on statistical methods to enable machine to improve with experience.
It learns the features by itself. Very good with processing on image data We have to explicitly tell what features to identify.
Can implement hand crafted features Can’t implement hand crafted features
Algorithms like CNN, RNN, GANs Algorithms like Linear Regression, KNN, Decision Tree, ANN

Why deep learning algorithms is called neural networks?

Computing Systems inspired by Biological Neural Networks.

Biological Neural Network

So basically it works on the same idea. The most basic network of deep learning is Artificial Neural Network it is also called as vanilla network. As you can see in blow picture there are inputs to the nodes of the network, hidden layer nodes and some output nodes. All these nodes are connected with some edges and some nodes receives inputs from biases. So neural network consist of main 3 layers

  • Input Layer
  • Hidden Layer
  • Output Layer
Artificial Neural Network (Fully Connected Network)

There are the terms called activation function and biases. So an activation function is the mathematical exponential or linear function which is to be applied on the output of a particular node and its output goes to next node in the network. There are many reasons behind the usage of the activation functions.

Let’s take an example to understand the use of activation function.

When our brain is fed with a lot of information simultaneously, it tries hard to understand and classify the information between useful and not-so-useful information. We need a similar mechanism to classify incoming information as useful or less-useful in case of Neural Networks. This is a very important in the way a network learns because not all information is equally useful. Some of it is just noise. Well, activation functions help the network do this segregation. They help the network use the useful information and suppress the irrelevant data points.

Popular Activation Functions

  • Identity
  • Binary Step
  • Sigmoid
  • Tanh
  • ReLU
  • Leaky ReLU
  • Softmax

Bias

The bias value allows the activation function to be shifted to the left or right, to better fit the data. Hence changes to the weights alter the steepness of the sigmoid curve, whilst the bias offsets it, shifting the entire curve so it fits better. Note also how the bias only influences the output values, it doesn’t interact with the actual input data

Major Architectures

  • Unsupervised Pretrained Networks
  • Convolutional Neural Networks
  • Recurrent Neural Networks
  • Recursive Neural Networks

UPNs

  • Autoencoders
  • Deep Belief Networks(DBNs)
  • Generative Adversarial Networks(GANs)

Use cases of UPN architectures

  • Feature Extraction
  • Initialization
  • Synthesizing

CNNs

  • Lenet-5
  • AlexNet
  • VGGNet
  • GoogleNet(Inception)
  • ResNet
  • ResNext
  • DenseNet
  • RCNN(RegionBasedCNN)
  • YOLO(YouOnlyLookOnce)
  • SqueezeNet
  • SegNet

Use cases of CNN architecture

  • Computer Vision
  • Natural Language Processing

RNNs

  • Hopfield Network
  • Long Short-Term Memory (LSTM)
  • Gated Recurrent Unit (GRU)

Use cases of RNN architecture

  • Sentiment Classification
  • Image Captioning
  • Language Translation
  • Video Captioning

Recursive Neural Network

  • Recursive Autoencoder
  • Recursive Neural Tensor Network

Use cases of Recursive NN architecture

  • Image scene decomposition
  • NLP
  • Audio-to-text transcription

References

Leave a comment

Design a site like this with WordPress.com
Get started