RUS  ENG
Full version
JOURNALS // Doklady Rossijskoj Akademii Nauk. Mathematika, Informatika, Processy Upravlenia // Archive

Dokl. RAN. Math. Inf. Proc. Upr., 2025 Volume 527, Pages 346–353 (Mi danma692)

SPECIAL ISSUE: ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING TECHNOLOGIES

Reinforcement learning for delta-hedging in illiquid markets

I. R. Dineev, P. Lukianchenko

National Research University Higher School of Economics, Moscow

Abstract: This paper addresses the problem of delta-hedging in illiquid markets, where transaction costs, limited depth of the limit order book, and market impact of large trades play a significant role. Classical approaches based on the Black–Scholes model assume continuous trading and infinite liquidity, which leads to significant distortions in practice. To overcome these limitations, we propose a reinforcement learning approach with a risk-averse Bellman operator. As the training environment, we employ an agent-based exchange simulator with support for trading the underlying asset and options, which reproduces market microstructure and limit order book dynamics. A DeepLOB convolutional encoder is used to extract order book features and capture hidden liquidity characteristics. Numerical experiments show that the proposed method produces a realized PnL distribution centered around zero with lighter tails compared to the classical Black–Scholes delta-hedger. Furthermore, the risk-aversion parameter $\lambda$ enables control over the trade-off between mean profitability and tail-risk management. The results demonstrate the efficiency of the approach and its applicability for constructing robust hedging strategies in illiquid markets.

Keywords: delta-hedging, deep hedging, reinforcement learning, risk-averse Bellman operator, agent-based modeling, limit order book, option derivatives.

UDC: 517.54

Received: 20.08.2025
Accepted: 29.09.2025

DOI: 10.7868/S2686954325070306



Bibliographic databases:


© Steklov Math. Inst. of RAS, 2025