RUS  ENG
Full version
JOURNALS // Zhurnal Vychislitel'noi Matematiki i Matematicheskoi Fiziki // Archive

Zh. Vychisl. Mat. Mat. Fiz., 2021 Volume 61, Number 5, Pages 800–812 (Mi zvmmf11239)

This article is cited in 5 papers

General numerical methods

Reduced-order modeling of deep neural networks

J. V. Gusaka, T. K. Daulbaeva, I. V. Oseledetsab, E. S. Ponomareva, A. S. Cichockia

a Skolkovo Institute of Science and Technology, Moscow, Russia
b Marchuk Institute of Numerical Mathematics of the Russian Academy of Sciences, Moscow

Abstract: We introduce a new method for speeding up the inference of deep neural networks. It is somewhat inspired by the reduced-order modeling techniques for dynamical systems. The cornerstone of the proposed method is the maximum volume algorithm. We demonstrate efficiency on neural networks pre-trained on different datasets. We show that in many practical cases it is possible to replace convolutional layers with much smaller fully-connected layers with a relatively small drop in accuracy.

Key words: acceleration of neural networks, MaxVol, machine learning, component analysis.

UDC: 519.65

Received: 24.12.2020
Revised: 24.12.2020
Accepted: 14.01.2021

DOI: 10.31857/S0044466921050100


 English version:
Computational Mathematics and Mathematical Physics, 2021, 61:5, 774–785

Bibliographic databases:


© Steklov Math. Inst. of RAS, 2024