false

Hybrid Quantum-Classical Ensemble Learning for S&P 500 Directional Prediction

Hybrid Quantum-Classical Ensemble Learning for S&P 500 Directional Prediction ArXiv ID: 2512.15738 “View on arXiv” Authors: Abraham Itzhak Weinberg Abstract Financial market prediction is a challenging application of machine learning, where even small improvements in directional accuracy can yield substantial value. Most models struggle to exceed 55–57% accuracy due to high noise, non-stationarity, and market efficiency. We introduce a hybrid ensemble framework combining quantum sentiment analysis, Decision Transformer architecture, and strategic model selection, achieving 60.14% directional accuracy on S&P 500 prediction, a 3.10% improvement over individual models. Our framework addresses three limitations of prior approaches. First, architecture diversity dominates dataset diversity: combining different learning algorithms (LSTM, Decision Transformer, XGBoost, Random Forest, Logistic Regression) on the same data outperforms training identical architectures on multiple datasets (60.14% vs.\ 52.80%), confirmed by correlation analysis ($r>0.6$ among same-architecture models). Second, a 4-qubit variational quantum circuit enhances sentiment analysis, providing +0.8% to +1.5% gains per model. Third, smart filtering excludes weak predictors (accuracy $<52%$), improving ensemble performance (Top-7 models: 60.14% vs.\ all 35 models: 51.2%). We evaluate on 2020–2023 market data across seven instruments, covering diverse regimes including the COVID-19 crash and inflation-driven correction. McNemar’s test confirms statistical significance ($p<0.05$). Preliminary backtesting with confidence-based filtering (6+ model consensus) yields a Sharpe ratio of 1.2 versus buy-and-hold’s 0.8, demonstrating practical trading potential. ...

December 6, 2025 · 2 min · Research Team

Statistical Arbitrage in Polish Equities Market Using Deep Learning Techniques

Statistical Arbitrage in Polish Equities Market Using Deep Learning Techniques ArXiv ID: 2512.02037 “View on arXiv” Authors: Marek Adamczyk, Michał Dąbrowski Abstract We study a systematic approach to a popular Statistical Arbitrage technique: Pairs Trading. Instead of relying on two highly correlated assets, we replace the second asset with a replication of the first using risk factor representations. These factors are obtained through Principal Components Analysis (PCA), exchange traded funds (ETFs), and, as our main contribution, Long Short Term Memory networks (LSTMs). Residuals between the main asset and its replication are examined for mean reversion properties, and trading signals are generated for sufficiently fast mean reverting portfolios. Beyond introducing a deep learning based replication method, we adapt the framework of Avellaneda and Lee (2008) to the Polish market. Accordingly, components of WIG20, mWIG40, and selected sector indices replace the original S&P500 universe, and market parameters such as the risk free rate and transaction costs are updated to reflect local conditions. We outline the full strategy pipeline: risk factor construction, residual modeling via the Ornstein Uhlenbeck process, and signal generation. Each replication technique is described together with its practical implementation. Strategy performance is evaluated over two periods: 2017-2019 and the recessive year 2020. All methods yield profits in 2017-2019, with PCA achieving roughly 20 percent cumulative return and an annualized Sharpe ratio of up to 2.63. Despite multiple adaptations, our conclusions remain consistent with those of the original paper. During the COVID-19 recession, only the ETF based approach remains profitable (about 5 percent annual return), while PCA and LSTM methods underperform. LSTM results, although negative, are promising and indicate potential for future optimization. ...

November 20, 2025 · 2 min · Research Team

Attention Factors for Statistical Arbitrage

Attention Factors for Statistical Arbitrage ArXiv ID: 2510.11616 “View on arXiv” Authors: Elliot L. Epstein, Rose Wang, Jaewon Choi, Markus Pelger Abstract Statistical arbitrage exploits temporal price differences between similar assets. We develop a framework to jointly identify similar assets through factors, identify mispricing and form a trading policy that maximizes risk-adjusted performance after trading costs. Our Attention Factors are conditional latent factors that are the most useful for arbitrage trading. They are learned from firm characteristic embeddings that allow for complex interactions. We identify time-series signals from the residual portfolios of our factors with a general sequence model. Estimating factors and the arbitrage trading strategy jointly is crucial to maximize profitability after trading costs. In a comprehensive empirical study we show that our Attention Factor model achieves an out-of-sample Sharpe ratio above 4 on the largest U.S. equities over a 24-year period. Our one-step solution yields an unprecedented Sharpe ratio of 2.3 net of transaction costs. We show that weak factors are important for arbitrage trading. ...

October 13, 2025 · 2 min · Research Team

Graph Learning for Foreign Exchange Rate Prediction and Statistical Arbitrage

Graph Learning for Foreign Exchange Rate Prediction and Statistical Arbitrage ArXiv ID: 2508.14784 “View on arXiv” Authors: Yoonsik Hong, Diego Klabjan Abstract We propose a two-step graph learning approach for foreign exchange statistical arbitrages (FXSAs), addressing two key gaps in prior studies: the absence of graph-learning methods for foreign exchange rate prediction (FXRP) that leverage multi-currency and currency-interest rate relationships, and the disregard of the time lag between price observation and trade execution. In the first step, to capture complex multi-currency and currency-interest rate relationships, we formulate FXRP as an edge-level regression problem on a discrete-time spatiotemporal graph. This graph consists of currencies as nodes and exchanges as edges, with interest rates and foreign exchange rates serving as node and edge features, respectively. We then introduce a graph-learning method that leverages the spatiotemporal graph to address the FXRP problem. In the second step, we present a stochastic optimization problem to exploit FXSAs while accounting for the observation-execution time lag. To address this problem, we propose a graph-learning method that enforces constraints through projection and ReLU, maximizes risk-adjusted return by leveraging a graph with exchanges as nodes and influence relationships as edges, and utilizes the predictions from the FXRP method for the constraint parameters and node features. Moreover, we prove that our FXSA method satisfies empirical arbitrage constraints. The experimental results demonstrate that our FXRP method yields statistically significant improvements in mean squared error, and that the FXSA method achieves a 61.89% higher information ratio and a 45.51% higher Sortino ratio than a benchmark. Our approach provides a novel perspective on FXRP and FXSA within the context of graph learning. ...

August 20, 2025 · 2 min · Research Team

Empirical Analysis of the Model-Free Valuation Approach: Hedging Gaps, Conservatism, and Trading Opportunities

Empirical Analysis of the Model-Free Valuation Approach: Hedging Gaps, Conservatism, and Trading Opportunities ArXiv ID: 2508.16595 “View on arXiv” Authors: Zixing Chen, Yihan Qi, Shanlan Que, Julian Sester, Xiao Zhang Abstract In this paper we study the quality of model-free valuation approaches for financial derivatives by systematically evaluating the difference between model-free super-hedging strategies and the realized payoff of financial derivatives using historical option prices from several constituents of the S&P 500 between 2018 and 2022. Our study allows in particular to describe the realized gap between payoff and model-free hedging strategy empirically so that we can quantify to which degree model-free approaches are overly conservative. Our results imply that the model-free hedging approach is only marginally more conservative than industry-standard models such as the Heston-model while being model-free at the same time. This finding, its statistical description and the model-independence of the hedging approach enable us to construct an explicit trading strategy which, as we demonstrate, can be profitably applied in financial markets, and additionally possesses the desirable feature with an explicit control of its downside risk due to its model-free construction preventing losses pathwise. ...

August 9, 2025 · 2 min · Research Team

CTBench: Cryptocurrency Time Series Generation Benchmark

CTBench: Cryptocurrency Time Series Generation Benchmark ArXiv ID: 2508.02758 “View on arXiv” Authors: Yihao Ang, Qiang Wang, Qiang Huang, Yifan Bao, Xinyu Xi, Anthony K. H. Tung, Chen Jin, Zhiyong Huang Abstract Synthetic time series are essential tools for data augmentation, stress testing, and algorithmic prototyping in quantitative finance. However, in cryptocurrency markets, characterized by 24/7 trading, extreme volatility, and rapid regime shifts, existing Time Series Generation (TSG) methods and benchmarks often fall short, jeopardizing practical utility. Most prior work (1) targets non-financial or traditional financial domains, (2) focuses narrowly on classification and forecasting while neglecting crypto-specific complexities, and (3) lacks critical financial evaluations, particularly for trading applications. To address these gaps, we introduce \textsf{“CTBench”}, the first comprehensive TSG benchmark tailored for the cryptocurrency domain. \textsf{“CTBench”} curates an open-source dataset from 452 tokens and evaluates TSG models across 13 metrics spanning 5 key dimensions: forecasting accuracy, rank fidelity, trading performance, risk assessment, and computational efficiency. A key innovation is a dual-task evaluation framework: (1) the \emph{“Predictive Utility”} task measures how well synthetic data preserves temporal and cross-sectional patterns for forecasting, while (2) the \emph{“Statistical Arbitrage”} task assesses whether reconstructed series support mean-reverting signals for trading. We benchmark eight representative models from five methodological families over four distinct market regimes, uncovering trade-offs between statistical fidelity and real-world profitability. Notably, \textsf{“CTBench”} offers model ranking analysis and actionable guidance for selecting and deploying TSG models in crypto analytics and strategy development. ...

August 3, 2025 · 2 min · Research Team

An Application of the Ornstein-Uhlenbeck Process to Pairs Trading

An Application of the Ornstein-Uhlenbeck Process to Pairs Trading ArXiv ID: 2412.12458 “View on arXiv” Authors: Unknown Abstract We conduct a preliminary analysis of a pairs trading strategy using the Ornstein-Uhlenbeck (OU) process to model stock price spreads. We compare this approach to a naive pairs trading strategy that uses a rolling window to calculate mean and standard deviation parameters. Our findings suggest that the OU model captures signals and trends effectively but underperforms the naive model on a risk-return basis, likely due to non-stationary pairs and parameter tuning limitations. ...

December 17, 2024 · 2 min · Research Team

A Deep Learning Approach for Trading Factor Residuals

A Deep Learning Approach for Trading Factor Residuals ArXiv ID: 2412.11432 “View on arXiv” Authors: Unknown Abstract The residuals in factor models prevalent in asset pricing presents opportunities to exploit the mis-pricing from unexplained cross-sectional variation for arbitrage. We performed a replication of the methodology of Guijarro-Ordonez et al. (2019) (G-P-Z) on Deep Learning Statistical Arbitrage (DLSA), originally applied to U.S. equity data from 1998 to 2016, using a more recent out-of-sample period from 2016 to 2024. Adhering strictly to point-in-time (PIT) principles and ensuring no information leakage, we follow the same data pre-processing, factor modeling, and deep learning architectures (CNNs and Transformers) as outlined by G-P-Z. Our replication yields unusually strong performance metrics in certain tests, with out-of-sample Sharpe ratios occasionally exceeding 10. While such results are intriguing, they may indicate model overfitting, highly specific market conditions, or insufficient accounting for transaction costs and market impact. Further examination and robustness checks are needed to align these findings with the more modest improvements reported in the original study. (This work was conducted as the final project for IEOR 4576: Data-Driven Methods in Finance at Columbia University.) ...

December 16, 2024 · 2 min · Research Team

A Comparison between Financial and Gambling Markets

A Comparison between Financial and Gambling Markets ArXiv ID: 2409.13528 “View on arXiv” Authors: Unknown Abstract Financial and gambling markets are ostensibly similar and hence strategies from one could potentially be applied to the other. Financial markets have been extensively studied, resulting in numerous theorems and models, while gambling markets have received comparatively less attention and remain relatively undocumented. This study conducts a comprehensive comparison of both markets, focusing on trading rather than regulation. Five key aspects are examined: platform, product, procedure, participant and strategy. The findings reveal numerous similarities between these two markets. Financial exchanges resemble online betting platforms, such as Betfair, and some financial products, including stocks and options, share speculative traits with sports betting. We examine whether well-established models and strategies from financial markets could be applied to the gambling industry, which lacks comparable frameworks. For example, statistical arbitrage from financial markets has been effectively applied to gambling markets, particularly in peer-to-peer betting exchanges, where bettors exploit odds discrepancies for risk-free profits using quantitative models. Therefore, exploring the strategies and approaches used in both markets could lead to new opportunities for innovation and optimization in trading and betting activities. ...

September 20, 2024 · 2 min · Research Team

Market information of the fractional stochastic regularity model

Market information of the fractional stochastic regularity model ArXiv ID: 2409.07159 “View on arXiv” Authors: Unknown Abstract The Fractional Stochastic Regularity Model (FSRM) is an extension of Black-Scholes model describing the multifractal nature of prices. It is based on a multifractional process with a random Hurst exponent $H_t$, driven by a fractional Ornstein-Uhlenbeck (fOU) process. When the regularity parameter $H_t$ is equal to $1/2$, the efficient market hypothesis holds, but when $H_t\neq 1/2$ past price returns contain some information on a future trend or mean-reversion of the log-price process. In this paper, we investigate some properties of the fOU process and, thanks to information theory and Shannon’s entropy, we determine theoretically the serial information of the regularity process $H_t$ of the FSRM, giving some insight into one’s ability to forecast future price increments and to build statistical arbitrages with this model. ...

September 11, 2024 · 2 min · Research Team