Fourier Neural Network Approximation of Transition Densities in Finance

ArXiv ID: 2309.03966 “View on arXiv”

Authors: Unknown

Abstract

This paper introduces FourNet, a novel single-layer feed-forward neural network (FFNN) method designed to approximate transition densities for which closed-form expressions of their Fourier transforms, i.e. characteristic functions, are available. A unique feature of FourNet lies in its use of a Gaussian activation function, enabling exact Fourier and inverse Fourier transformations and drawing analogies with the Gaussian mixture model. We mathematically establish FourNet’s capacity to approximate transition densities in the $L_2$-sense arbitrarily well with finite number of neurons. The parameters of FourNet are learned by minimizing a loss function derived from the known characteristic function and the Fourier transform of the FFNN, complemented by a strategic sampling approach to enhance training. We derive practical bounds for the $L_2$ estimation error and the potential pointwise loss of nonnegativity in FourNet for $d$-dimensions ($d\ge 1$), highlighting its robustness and applicability in practical settings. FourNet’s accuracy and versatility are demonstrated through a wide range of dynamics common in quantitative finance, including Lévy processes and the Heston stochastic volatility models-including those augmented with the self-exciting Queue-Hawkes jump process.

Keywords: FourNet, Gaussian activation function, Characteristic function, Heston model, Lévy processes, Equity Derivatives

Complexity vs Empirical Score

  • Math Complexity: 8.5/10
  • Empirical Rigor: 6.5/10
  • Quadrant: Holy Grail
  • Why: The paper demonstrates high mathematical complexity through rigorous $L_2$-analysis, universal approximation theorems, and multi-dimensional error bounds, while maintaining strong empirical rigor with validated demonstrations across diverse financial models (Lévy, Heston, Queue-Hawkes) and practical error controls for real-world applications.
  flowchart TD
    A["Research Goal"] --> B["FourNet Architecture"]
    A --> C["Input Data"]
    subgraph B ["Neural Network Structure"]
        B1["Gaussian Activation"]
        B2["Fourier Transform"]
    end
    subgraph C ["Financial Models"]
        C1["Levy Processes"]
        C2["Heston Model"]
        C3["Queue-Hawkes Jumps"]
    end
    B --> D["Training Process"]
    C --> D
    D --> E["Loss Function"]
    D --> F["Strategic Sampling"]
    E --> G["Minimization"]
    F --> G
    G --> H["Key Outcomes"]
    subgraph H ["Findings"]
        H1["Arbitrary L2 Approximation"]
        H2["Pointwise Error Bounds"]
        H3["Preserves Nonnegativity"]
    end