Skip to content
Adam Gibson edited this page Dec 29, 2013 · 11 revisions

Overview

  • Deep Learning is a newer architecture of neural networks that has come in to favor as of recent years (starting with Hinton 2006) . A Deep Network consists of (usually) 3 or more layers that each have a binomial sampling component as a way of doing feature extraction followed by a Logistic Regression Softmax Layer (there are other examples that will be described later). See Denoising AutoEncoders or Restricted Boltzmann Machines as the typical examples. Their multilayer counterparts counterparts are Stacked Denoising AutoEncoders and Deep Belief Networks respectively.

  • Training - 2 phases * Pretraining: feature extraction via sampling by minimizing reconstruction error. Each layer is trained with a supervised objective function of the input minimizing reconstruction error.

    * Finetuning: From there stochastic gradient descent/LBFGS/Conjugate Gradient are run as a way of updating the weights on the final Logistic Regression output layer. The Logistic Regression layer in this case is learning from the features extracted via unsupervised means in the previous layers. From there you can run a classifier.
    

Resources

Use Cases