期刊名称:Journal of Artificial Intelligence and Soft Computing Research
电子版ISSN:2083-2567
出版年度:2021
卷号:11
期号:4
页码:287-306
DOI:10.2478/jaiscr-2021-0017
语种:English
出版社:Walter de Gruyter GmbH
摘要:In this paper1 a new neural networks training algorithm is presented. The algorithm originates from the Recursive Least Squares (RLS) method commonly used in adaptive filtering. It uses the QR decomposition in conjunction with the Givens rotations for solving a normal equation - resulting from minimization of the loss function. An important parameter in neural networks is training time. Many commonly used algorithms require a big number of iterations in order to achieve a satisfactory outcome while other algorithms are effective only for small neural networks. The proposed solution is characterized by a very short convergence time compared to the well-known backpropagation method and its variants. The paper contains a complete mathematical derivation of the proposed algorithm. There are presented extensive simulation results using various benchmarks including function approximation, classification, encoder, and parity problems. Obtained results show the advantages of the featured algorithm which outperforms commonly used recent state-of-the-art neural networks training algorithms, including the Adam optimizer and the Nesterov’s accelerated gradient.
关键词:neural network training algorithm;QR decomposition;Givens rotations;approximation;classification