The Backpropagation algorithm looks for the minimum value of the error function in weight space using a technique called the delta rule or gradient descent. The weights that minimize the error function is then considered to be a solution to the learning problem.

An algorithm that looks for the minimum value of the error function in weight space.

Written by Marcus Kelly

Updated over a week ago