×

Discriminative pretraining of deep neural networks

  • US 10,325,200 B2
  • Filed: 10/01/2015
  • Issued: 06/18/2019
  • Est. Priority Date: 11/26/2011
  • Status: Active Grant
First Claim
Patent Images

1. A computer-implemented process for pretraining deep neural network (DNN), comprising:

  • using a computer to perform the following process actions;

    (a) training a single hidden layer neural network (NN) comprising an input layer into which training data is input, a multi-neuron output layer from which an output is generated, and a first multi-neuron hidden layer which is interconnected with the input and output layers with randomly initialized weights, said hidden layer having a fixed number of neurons, wherein said training comprises,accessing a set of training data entries, each data entry of which has a corresponding label assigned thereto,inputting each data entry of said set one by one into the input layer until all the data entries have been input at least once to produce an initial NN, such that after the inputting of each data entry, said weights associated with a first hidden layer are set via an error back propagation (BP) procedure so that the output generated from the multi-neuron output layer matches the label associated with the training data entry;

    (b) discarding a current multi-neuron output layer and adding a new multi-neuron hidden layer which is interconnected with a last previously trained hidden layer and a new multi-neuron output layer with randomly initialized weights to produce a new multiple hidden layer deep neural network, said new hidden layer having a fixed number of neurons;

    (c) inputting each data entry of said set one by one into the input layer until all the data entries have been input at least once to produce a revised multiple hidden layer deep neural network, such that after the inputting of each data entry, said weights associated with the new hidden layer and each previously trained hidden layer are set via the error BP procedure to produce an output from the new multi-neuron output layer that matches the label associated with the training data entry;

    (d) repeating actions (b) and (c) until a prescribed number of hidden layers have been added;

    (e) designating the last produced revised multiple layer DNN to be said pretrained DNN; and

    (f) iteratively training the pretrained DNN to produce a trained DNN.

View all claims
  • 2 Assignments
Timeline View
Assignment View
    ×
    ×