RUS  ENG
Full version
JOURNALS // Informatics and Automation // Archive

Informatics and Automation, 2021 Issue 20, volume 2, Pages 463–490 (Mi trspy1150)

Artificial Intelligence, Knowledge and Data Engineering

Hybrid method of conventional neural network training

A. Golubinskiya, A. Tolstykhb

a JSC “Concern “Sozvezdie”
b Moscow University of the Ministry of Internal Affairs of Russia

Abstract: The paper proposes a hybrid method for training convolutional neural networks. The method consists of combining second and first-order methods for different elements of the architecture of a convolutional neural network. The hybrid convolution neural network training method allows to achieve significantly better convergence compared to Adam; however, it requires fewer computational operations to implement. Using the proposed method, it is possible to train networks on which learning paralysis occurs when using first-order methods. Moreover, the proposed method could adjust its computational complexity to the hardware on which the computation is performed; at the same time, the hybrid method allows using the mini-packet learning approach.
The analysis of the ratio of computations between convolutional neural networks and fully connected artificial neural networks is presented. The mathematical apparatus of error optimization of artificial neural networks is considered, including the method of backpropagation of the error, the Levenberg-Marquardt algorithm. The main limitations of these methods that arise when training a convolutional neural network are analyzed.
The analysis of the stability of the proposed method when the initialization parameters are changed. The results of the applicability of the method in various problems are presented.

Keywords: convolutional neural networks, training methods for artificial neural networks, optimization methods.

UDC: 004.032.26

Received: 08.12.2020

DOI: 10.15622/ia.2021.20.2.8



© Steklov Math. Inst. of RAS, 2024