TY - GEN
T1 - Improving generalization capability of neural networks based on simulated annealing
AU - Lee, Yeejin
AU - Lee, Jong Seok
AU - Lee, Sun Young
AU - Park, Cheol Hoon
PY - 2007
Y1 - 2007
N2 - This paper presents a single-objective and a multiobjective stochastic optimization algorithms for global training of neural networks based on simulated annealing. The algorithms overcome the limitation of local optimization by the conventional gradient-based training methods and perform global optimization of the weights of the neural networks. Especially, the multiobjective training algorithm is designed to enhance generalization capability of the trained networks by minimizing the training error and the dynamic range of the network weights simultaneously. For fast convergence and good solution quality of the algorithms, we suggest the hybrid simulated annealing algorithm with the gradient-based local optimization method. Experimental results show that the performance of the trained networks by the proposed methods is better than that by the gradient-based local training algorithm and, moreover, the generalization capability of the networks is significantly improved by preventing overfitting phenomena.
AB - This paper presents a single-objective and a multiobjective stochastic optimization algorithms for global training of neural networks based on simulated annealing. The algorithms overcome the limitation of local optimization by the conventional gradient-based training methods and perform global optimization of the weights of the neural networks. Especially, the multiobjective training algorithm is designed to enhance generalization capability of the trained networks by minimizing the training error and the dynamic range of the network weights simultaneously. For fast convergence and good solution quality of the algorithms, we suggest the hybrid simulated annealing algorithm with the gradient-based local optimization method. Experimental results show that the performance of the trained networks by the proposed methods is better than that by the gradient-based local training algorithm and, moreover, the generalization capability of the networks is significantly improved by preventing overfitting phenomena.
UR - http://www.scopus.com/inward/record.url?scp=78751624990&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78751624990&partnerID=8YFLogxK
U2 - 10.1109/CEC.2007.4424918
DO - 10.1109/CEC.2007.4424918
M3 - Conference contribution
AN - SCOPUS:78751624990
SN - 1424413400
SN - 9781424413409
T3 - 2007 IEEE Congress on Evolutionary Computation, CEC 2007
SP - 3447
EP - 3453
BT - 2007 IEEE Congress on Evolutionary Computation, CEC 2007
T2 - 2007 IEEE Congress on Evolutionary Computation, CEC 2007
Y2 - 25 September 2007 through 28 September 2007
ER -