Logarithmic regret in the ergodic Avellaneda-Stoikov market making model

ArXiv ID: 2409.02025 “View on arXiv”

Authors: Unknown

Abstract

We analyse the regret arising from learning the price sensitivity parameter $κ$ of liquidity takers in the ergodic version of the Avellaneda-Stoikov market making model. We show that a learning algorithm based on a maximum-likelihood estimator for the parameter achieves the regret upper bound of order $\ln^2 T$ in expectation. To obtain the result we need two key ingredients. The first is the twice differentiability of the ergodic constant under the misspecified parameter in the Hamilton-Jacobi-Bellman (HJB) equation with respect to $κ$, which leads to a second–order performance gap. The second is the learning rate of the regularised maximum-likelihood estimator which is obtained from concentration inequalities for Bernoulli signals. Numerical experiments confirm the convergence and the robustness of the proposed algorithm.

Keywords: Avellaneda-Stoikov market making, regret analysis, maximum-likelihood estimator, Hamilton-Jacobi-Bellman (HJB), ergodic control, Market Making

Complexity vs Empirical Score

  • Math Complexity: 9.2/10
  • Empirical Rigor: 3.5/10
  • Quadrant: Lab Rats
  • Why: The paper is mathematically dense, featuring ergodic stochastic control, Hamilton-Jacobi-Bellman equations, and logarithmic regret analysis in continuous time. However, empirical validation is limited to numerical experiments without specific backtest metrics, code, or real-market datasets.
  flowchart TD
    A["Research Goal:<br>Regret analysis for learning κ<br>in ergodic Avellaneda-Stoikov model"] --> B["Methodology"]
    
    B --> C["Key Ingredients & Analysis"]
    C --> D["1. HJB Differentiability:<br>Twice differentiable ergodic constant<br>wrt κ (misspecified parameter)"]
    C --> E["2. Learning Theory:<br>Regularised Maximum-Likelihood Estimator<br>for Bernoulli signals"]
    
    B --> F["Computational Process:<br>Simulation of ergodic market<br>with adaptive spreads based on learned κ"]
    
    D & E & F --> G["Main Theoretical Result:<br>Regret upper bound O(ln²T) in expectation"]
    
    H["Numerical Experiments"] --> I["Key Findings:<br>- Convergence of regret<br>- Robustness of algorithm<br>- Validation of theory"]
    
    G --> I