Some applications of Multi-Agent Reinforcement Learning in Over-the-Counter Market Simulations and Local Stochastic Volatility Model Calibration.
In the first part of the talk, we study a game between liquidity provider and liquidity taker agents interacting in an over-the-counter market, for which the typical example is foreign exchange. We show how a suitable design of parameterized families of reward functions coupled with associated shared policy learning constitutes an efficient solution to this problem. Precisely, we show that our reinforcement-learning-driven agents learn emergent behaviors relative to a wide spectrum of incentives encompassing profit-and-loss, optimal execution and market share, by playing against each other. In particular, we find that liquidity providers naturally learn to balance hedging and skewing as a function of their incentives, where the latter refers to setting their buy and sell prices asymmetrically as a function of their inventory. Further, we introduce a novel RL-based calibration algorithm which we found performed well at imposing constraints on the game equilibrium, both on toy and real market data.
In the second part of the talk, we present a multi-agent RL-based algorithm to calibrate Local Stochastic Volatility models to market prices of exotics. In our experiments, we show that we are able to learn a path-dependent volatility that minimizes the price of a Bermudan option while remaining calibrated to vanillas.