Meta-Learning Neural Process for Implied Volatility Surfaces with SABR-induced Priors

ArXiv ID: 2509.11928 “View on arXiv”

Authors: Jirong Zhuang, Xuan Wu

Abstract

We treat implied volatility surface (IVS) reconstruction as a learning problem guided by two principles. First, we adopt a meta-learning view that trains across trading days to learn a procedure that maps sparse option quotes to a full IVS via conditional prediction, avoiding per-day calibration at test time. Second, we impose a structural prior via transfer learning: pre-train on SABR-generated dataset to encode geometric prior, then fine-tune on historical market dataset to align with empirical patterns. We implement both principles in a single attention-based Neural Process (Volatility Neural Process, VolNP) that produces a complete IVS from a sparse context set in one forward pass. On SPX options, the VolNP outperforms SABR, SSVI, and Gaussian process. Relative to an ablation trained only on market data, the SABR-induced prior reduces RMSE by about 40% and suppresses large errors, with pronounced gains at long maturities where quotes are sparse. The resulting model is fast (single pass), stable (no daily recalibration), and practical for deployment at scale.

Keywords: implied volatility surface, meta-learning, transfer learning, neural processes, SABR model, Equity Derivatives

Complexity vs Empirical Score

  • Math Complexity: 8.5/10
  • Empirical Rigor: 8.0/10
  • Quadrant: Holy Grail
  • Why: The paper employs advanced mathematics including neural process theory, attention mechanisms, and meta-learning frameworks, scoring high on math complexity. It provides concrete empirical results with specific metrics (40% RMSE reduction), compares to baselines (SABR, SSVI, Gaussian process), and discusses practical deployment, indicating strong empirical rigor.
  flowchart TD
    A["Research Goal: Learn IVS from sparse options data<br/>Avoid per-day calibration, incorporate structural prior"] --> B["Methodology: Volatility Neural Process VolNP<br/>Attention-based meta-learning + SABR transfer learning"]
    B --> C["Data & Inputs: SPX Options Data"]
    B --> D["SABR-generated synthetic dataset<br/>(Geometric Prior Encoding)"]
    C --> E["Computational Process: Meta-training across trading days<br/>Conditional prediction for full IVS"]
    D --> E
    E --> F["Key Findings/Outcomes"]
    F --> G["Accuracy: 40% RMSE reduction vs. market-only ablation"]
    F --> H["Robustness: Stable, fast (single pass)<br/>No daily recalibration needed"]
    F --> I["Generalization: Strong gains at long maturities<br/>Outperforms SABR, SSVI, Gaussian Process"]
```mermaid
flowchart TD
    A["Research Goal: Learn IVS from sparse options data<br/>Avoid per-day calibration, incorporate structural prior"] --> B["Methodology: Volatility Neural Process VolNP<br/>Attention-based meta-learning + SABR transfer learning"]
    B --> C["Data & Inputs: SPX Options Data"]
    B --> D["SABR-generated synthetic dataset<br/>(Geometric Prior Encoding)"]
    C --> E["Computational Process: Meta-training across trading days<br/>Conditional prediction for full IVS"]
    D --> E
    E --> F["Key Findings/Outcomes"]
    F --> G["Accuracy: 40% RMSE reduction vs. market-only ablation"]
    F --> H["Robustness: Stable, fast (single pass)<br/>No daily recalibration needed"]
    F --> I["Generalization: Strong gains at long maturities<br/>Outperforms SABR, SSVI, Gaussian Process"]