Back Propagation
An algorithm that looks for the minimum value of the error function in weight space.
Marcus Kelly avatar
Written by Marcus Kelly
Updated over a week ago

The Backpropagation algorithm looks for the minimum value of the error function in weight space using a technique called the delta rule or gradient descent. The weights that minimize the error function is then considered to be a solution to the learning problem.

Did this answer your question?