RUS  ENG
Full version
JOURNALS // Zhurnal Vychislitel'noi Matematiki i Matematicheskoi Fiziki // Archive

Zh. Vychisl. Mat. Mat. Fiz., 2021 Volume 61, Number 5, Pages 865–877 (Mi zvmmf11244)

This article is cited in 1 paper

Optimal control

TT-QI: Faster value iteration in tensor train format for stochastic optimal control

A. I. Boykoa, I. V. Oseledetsab, G. Ferrera

a Skolkovo Institute of Science and Technology, 121205, Moscow, Russia
b Marchuk Institute of Numerical Mathematics of the Russian Academy of Sciences, Moscow

Abstract: The problem of general non-linear stochastic optimal control with small Wiener noise is studied. The problem is approximated by a Markov Decision Process. Bellman Equation is solved using Value Iteration (VI) algorithm in the low rank Tensor Train format (TT-VI). In this paper a modification of the TT-VI algorithm called TT-Q-Iteration (TT-QI) is proposed by authors. In it, the nonlinear Bellman Optimality Operator is iteratively applied to the solution as a composition of internal Tensor Train algebraic operations and TT-CROSS algorithm. We show that it has lower asymptotic complexity per iteration than the method existing in the literature, provided that TT-ranks of transition probabilities are small. In test examples of an underpowered inverted pendulum and Dubins cars our method shows up to 3–10 times faster convergence in terms of wall clock time compared with the original method.

Key words: dynamic programming, optimal control, Markov decision process, MDP, Markov chain approximation, MCA, low rank decomposition.

UDC: 517.977.54

Received: 24.11.2020
Revised: 24.11.2020
Accepted: 14.01.2021

DOI: 10.31857/S0044466921050045


 English version:
Computational Mathematics and Mathematical Physics, 2021, 61:5, 836–846

Bibliographic databases:


© Steklov Math. Inst. of RAS, 2024