A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2017; you can also visit the original URL.
The file type is application/pdf
.
Beyond Backpropagation
1999
Journal of Organizational and End User Computing
The vast majority of neural network research relies on a gradient algorithm, typically a variation of backpropagation, to obtain the weights of the model. Because of the enigmatic nature of complex nonlinear optimization problems, such as training artificial neural networks, this technique has often produced inconsistent and unpredictable results. To go beyond backpropagation's typical selection of local solutions, simulated annealing is suggested as an alternative training technique that will
doi:10.4018/joeuc.1999070101
fatcat:t2bluizlzbaafj4z6duh56vsqy