MLP TRAINING: BACK PROPAGATION

By incorporating a nonlinear transfer function that is differentiable, we can derive an iterative gradient descent training algorithm for a multi-layer perceptron (MLP). This algorithm is known as back propagation:


The MLP network has been the most popular architecture for speech processing applications due to the existence of robust training algorithms and its powerful classification properties.