false

Convergence of the generalization error for deep gradient flow methods for PDEs

Convergence of the generalization error for deep gradient flow methods for PDEs ArXiv ID: 2512.25017 “View on arXiv” Authors: Chenguang Liu, Antonis Papapantoleon, Jasper Rou Abstract The aim of this article is to provide a firm mathematical foundation for the application of deep gradient flow methods (DGFMs) for the solution of (high-dimensional) partial differential equations (PDEs). We decompose the generalization error of DGFMs into an approximation and a training error. We first show that the solution of PDEs that satisfy reasonable and verifiable assumptions can be approximated by neural networks, thus the approximation error tends to zero as the number of neurons tends to infinity. Then, we derive the gradient flow that the training process follows in the ``wide network limit’’ and analyze the limit of this flow as the training time tends to infinity. These results combined show that the generalization error of DGFMs tends to zero as the number of neurons and the training time tend to infinity. ...

December 31, 2025 · 2 min · Research Team

DeepSVM: Learning Stochastic Volatility Models with Physics-Informed Deep Operator Networks

DeepSVM: Learning Stochastic Volatility Models with Physics-Informed Deep Operator Networks ArXiv ID: 2512.07162 “View on arXiv” Authors: Kieran A. Malandain, Selim Kalici, Hakob Chakhoyan Abstract Real-time calibration of stochastic volatility models (SVMs) is computationally bottlenecked by the need to repeatedly solve coupled partial differential equations (PDEs). In this work, we propose DeepSVM, a physics-informed Deep Operator Network (PI-DeepONet) designed to learn the solution operator of the Heston model across its entire parameter space. Unlike standard data-driven deep learning (DL) approaches, DeepSVM requires no labelled training data. Rather, we employ a hard-constrained ansatz that enforces terminal payoffs and static no-arbitrage conditions by design. Furthermore, we use Residual-based Adaptive Refinement (RAR) to stabilize training in difficult regions subject to high gradients. Overall, DeepSVM achieves a final training loss of $10^{"-5"}$ and predicts highly accurate option prices across a range of typical market dynamics. While pricing accuracy is high, we find that the model’s derivatives (Greeks) exhibit noise in the at-the-money (ATM) regime, highlighting the specific need for higher-order regularization in physics-informed operator learning. ...

December 8, 2025 · 2 min · Research Team

The deep multi-FBSDE method: a robust deep learning method for coupled FBSDEs

The deep multi-FBSDE method: a robust deep learning method for coupled FBSDEs ArXiv ID: 2503.13193 “View on arXiv” Authors: Unknown Abstract We introduce the deep multi-FBSDE method for robust approximation of coupled forward-backward stochastic differential equations (FBSDEs), focusing on cases where the deep BSDE method of Han, Jentzen, and E (2018) fails to converge. To overcome the convergence issues, we consider a family of FBSDEs that are equivalent to the original problem in the sense that they satisfy the same associated partial differential equation (PDE). Our algorithm proceeds in two phases: first, we approximate the initial condition for the FBSDE family, and second, we approximate the original FBSDE using the initial condition approximated in the first phase. Numerical experiments show that our method converges even when the standard deep BSDE method does not. ...

March 17, 2025 · 2 min · Research Team

The ATM implied skew in the ADO-Heston model

The ATM implied skew in the ADO-Heston model ArXiv ID: 2309.15044 “View on arXiv” Authors: Unknown Abstract In this paper similar to [“P. Carr, A. Itkin, 2019”] we construct another Markovian approximation of the rough Heston-like volatility model - the ADO-Heston model. The characteristic function (CF) of the model is derived under both risk-neutral and real measures which is an unsteady three-dimensional PDE with some coefficients being functions of the time $t$ and the Hurst exponent $H$. To replicate known behavior of the market implied skew we proceed with a wise choice of the market price of risk, and then find a closed form expression for the CF of the log-price and the ATM implied skew. Based on the provided example, we claim that the ADO-Heston model (which is a pure diffusion model but with a stochastic mean-reversion speed of the variance process, or a Markovian approximation of the rough Heston model) is able (approximately) to reproduce the known behavior of the vanilla implied skew at small $T$. We conclude that the behavior of our implied volatility skew curve ${"\cal S"}(T) \propto a(H) T^{“b\cdot (H-1/2)”}, , b = const$, is not exactly same as in rough volatility models since $b \ne 1$, but seems to be close enough for all practical values of $T$. Thus, the proposed Markovian model is able to replicate some properties of the corresponding rough volatility model. Similar analysis is provided for the forward starting options where we found that the ATM implied skew for the forward starting options can blow-up for any $s > t$ when $T \to s$. This result, however, contradicts to the observation of [“E. Alos, D.G. Lorite, 2021”] that Markovian approximation is not able to catch this behavior, so remains the question on which one is closer to reality. ...

September 26, 2023 · 3 min · Research Team