RUS  ENG
Full version
JOURNALS // Informatika i Ee Primeneniya [Informatics and its Applications] // Archive

Inform. Primen., 2018 Volume 12, Issue 4, Pages 63–69 (Mi ia564)

This article is cited in 1 paper

Optimal recurrent neural network model in paraphrase detection

A. N. Smerdova, O. Yu. Bakhteeva, V. V. Strijovab

a Moscow Institute of Physics and Technology, 9 Institutskiy Per., Dolgoprudny, Moscow Region 141700, Russian Federation
b A. A. Dorodnicyn Computing Center, Federal Research Center “Computer Science and Control” of the Russian Academy of Sciences, 40 Vavilov Str., Moscow 119333, Russian Federation

Abstract: This paper addresses the problem of optimal recurrent neural network selection. It asserts the neural network evidence lower bound as the optimal criterion for selection. It investigates variational inference methods to approximate the posterior distribution of the network parameters. As a particular case, the normal distribution of the parameters with different types of the covariance matrix is investigated. The authors propose a method of pruning parameters with the highest probability density in zero to increase the model marginal likelihood. As an illustrative example, a computational experiment of multiclass classification on the SemEval 2015 dataset was carried out.

Keywords: deep learning, recurrent neural network, neural network pruning, variational approach.

Received: 05.05.2018

DOI: 10.14357/19922264180409



Bibliographic databases:


© Steklov Math. Inst. of RAS, 2024