Fast Learning in Quantitative Finance with Extreme Learning Machine

ArXiv ID: 2505.09551 “View on arXiv”

Authors: Liexin Cheng, Xue Cheng, Shuaiqiang Liu

Abstract

A critical factor in adopting machine learning for time-sensitive financial tasks is computational speed, including model training and inference. This paper demonstrates that a broad class of such problems, especially those previously addressed using deep neural networks, can be efficiently solved using single-layer neural networks without iterative gradient-based training. This is achieved through the extreme learning machine (ELM) framework. ELM utilizes a single-layer network with randomly initialized hidden nodes and output weights obtained via convex optimization, enabling rapid training and inference. We present various applications in both supervised and unsupervised learning settings, including option pricing, intraday return prediction, volatility surface fitting, and numerical solution of partial differential equations. Across these examples, ELM demonstrates notable improvements in computational efficiency while maintaining comparable accuracy and generalization compared to deep neural networks and classical machine learning methods. We also briefly discuss theoretical aspects of ELM implementation and its generalization capabilities.

Keywords: Extreme Learning Machine (ELM), Computational Efficiency, Option Pricing, Volatility Surface Fitting, Single-layer Neural Networks, Multi-Asset

Complexity vs Empirical Score

  • Math Complexity: 6.5/10
  • Empirical Rigor: 4.0/10
  • Quadrant: Lab Rats
  • Why: The paper introduces advanced neural network theory and convex optimization with several mathematical proofs and bounds, scoring high in math complexity, but lacks detailed code, datasets, or backtest results, resulting in moderate empirical rigor.
  flowchart TD
    A["Research Goal: Develop a computationally efficient alternative to DNNs for time-sensitive financial tasks"] --> B["Methodology: Extreme Learning Machine<br/>(Single-layer network with random weights)"]
    B --> C["Data: Financial time-series & surfaces<br/>(Option prices, volatility, intraday returns)"]
    C --> D["Computational Process<br/>1. Randomly initialize hidden nodes<br/>2. Convex optimization for output weights<br/>3. Rapid training & inference"]
    D --> E["Key Outcomes<br/>• High computational efficiency<br/>• Comparable accuracy to DNNs<br/>• Strong generalization<br/>• Applicable to option pricing, PDEs, forecasting"]