false

Adaptive Alpha Weighting with PPO: Enhancing Prompt-Based LLM-Generated Alphas in Quant Trading

Adaptive Alpha Weighting with PPO: Enhancing Prompt-Based LLM-Generated Alphas in Quant Trading ArXiv ID: 2509.01393 “View on arXiv” Authors: Qizhao Chen, Hiroaki Kawashima Abstract This paper proposes a reinforcement learning framework that employs Proximal Policy Optimization (PPO) to dynamically optimize the weights of multiple large language model (LLM)-generated formulaic alphas for stock trading strategies. Formulaic alphas are mathematically defined trading signals derived from price, volume, sentiment, and other data. Although recent studies have shown that LLMs can generate diverse and effective alphas, a critical challenge lies in how to adaptively integrate them under varying market conditions. To address this gap, we leverage the deepseek-r1-distill-llama-70b model to generate fifty alphas for five major stocks: Apple, HSBC, Pepsi, Toyota, and Tencent, and then use PPO to adjust their weights in real time. Experimental results demonstrate that the PPO-optimized strategy achieves strong returns and high Sharpe ratios across most stocks, outperforming both an equal-weighted alpha portfolio and traditional benchmarks such as the Nikkei 225, S&P 500, and Hang Seng Index. The findings highlight the importance of reinforcement learning in the allocation of alpha weights and show the potential of combining LLM-generated signals with adaptive optimization for robust financial forecasting and trading. ...

September 1, 2025 · 2 min · Research Team

Controllable Generation of Implied Volatility Surfaces with Variational Autoencoders

Controllable Generation of Implied Volatility Surfaces with Variational Autoencoders ArXiv ID: 2509.01743 “View on arXiv” Authors: Jing Wang, Shuaiqiang Liu, Cornelis Vuik Abstract This paper presents a deep generative modeling framework for controllably synthesizing implied volatility surfaces (IVSs) using a variational autoencoder (VAE). Unlike conventional data-driven models, our approach provides explicit control over meaningful shape features (e.g., volatility level, slope, curvature, term-structure) to generate IVSs with desired characteristics. In our framework, financially interpretable shape features are disentangled from residual latent factors. The target features are embedded into the VAE architecture as controllable latent variables, while the residual latent variables capture additional structure to preserve IVS shape diversity. To enable this control, IVS feature values are quantified via regression at an anchor point and incorporated into the decoder to steer generation. Numerical experiments demonstrate that the generative model enables rapid generation of realistic IVSs with desired features rather than arbitrary patterns, and achieves high accuracy across both single- and multi-feature control settings. For market validity, an optional post-generation latent-space repair algorithm adjusts only the residual latent variables to remove occasional violations of static no-arbitrage conditions without altering the specified features. Compared with black-box generators, the framework combines interpretability, controllability, and flexibility for synthetic IVS generation and scenario design. ...

September 1, 2025 · 2 min · Research Team

Is All the Information in the Price? LLM Embeddings versus the EMH in Stock Clustering

Is All the Information in the Price? LLM Embeddings versus the EMH in Stock Clustering ArXiv ID: 2509.01590 “View on arXiv” Authors: Bingyang Wang, Grant Johnson, Maria Hybinette, Tucker Balch Abstract This paper investigates whether artificial intelligence can enhance stock clustering compared to traditional methods. We consider this in the context of the semi-strong Efficient Markets Hypothesis (EMH), which posits that prices fully reflect all public information and, accordingly, that clusters based on price information cannot be improved upon. We benchmark three clustering approaches: (i) price-based clusters derived from historical return correlations, (ii) human-informed clusters defined by the Global Industry Classification Standard (GICS), and (iii) AI-driven clusters constructed from large language model (LLM) embeddings of stock-related news headlines. At each date, each method provides a classification in which each stock is assigned to a cluster. To evaluate a clustering, we transform it into a synthetic factor model following the Arbitrage Pricing Theory (APT) framework. This enables consistent evaluation of predictive performance in a roll forward, out-of-sample test. Using S&P 500 constituents from from 2022 through 2024, we find that price-based clustering consistently outperforms both rule-based and AI-based methods, reducing root mean squared error (RMSE) by 15.9% relative to GICS and 14.7% relative to LLM embeddings. Our contributions are threefold: (i) a generalizable methodology that converts any equity grouping: manual, machine, or market-driven, into a real-time factor model for evaluation; (ii) the first direct comparison of price-based, human rule-based, and AI-based clustering under identical conditions; and (iii) empirical evidence reinforcing that short-horizon return information is largely contained in prices. These results support the EMH while offering practitioners a practical diagnostic for monitoring evolving sector structures and provide academics a framework for testing alternative hypotheses about how quickly markets absorb information. ...

September 1, 2025 · 3 min · Research Team

The Impact of Sequential versus Parallel Clearing Mechanisms in Agent-Based Simulations of Artificial Limit Order Book Exchanges

The Impact of Sequential versus Parallel Clearing Mechanisms in Agent-Based Simulations of Artificial Limit Order Book Exchanges ArXiv ID: 2509.01683 “View on arXiv” Authors: Matej Steinbacher, Mitja Steinbacher, Matjaz Steinbacher Abstract This study examines the impact of different computing implementations of clearing mechanisms on multi-asset price dynamics within an artificial stock market framework. We show that sequential processing of order books introduces a systematic and significant bias by affecting the allocation of traders’ capital within a single time step. This occurs because applying budget constraints sequentially grants assets processed earlier preferential access to funds, distorting individual asset demand and consequently their price trajectories. The findings highlight that while the overall price level is primarily driven by macro factors like the money-to-stock ratio, the market’s microstructural clearing mechanism plays a critical role in the allocation of value among individual assets. This underscores the necessity for careful consideration and validation of clearing mechanisms in artificial markets to accurately model complex financial behaviors. ...

September 1, 2025 · 2 min · Research Team

Multi Scale Analysis of Nifty 50 Return Characteristics Valuation Dynamics and Market Complexity 1990 to 2024

Multi Scale Analysis of Nifty 50 Return Characteristics Valuation Dynamics and Market Complexity 1990 to 2024 ArXiv ID: 2509.00697 “View on arXiv” Authors: Chandradew Sharma Abstract This study presents a unified, distribution-aware, and complexity-informed framework for understanding equity return dynamics in the Indian market, using 34 years (1990 to 2024) of Nifty 50 index data. Addressing a key gap in the literature, we demonstrate that the price to earnings ratio, as a valuation metric, may probabilistically map return distributions across investment horizons spanning from days to decades. Return profiles exhibit strong asymmetry. One-year returns show a 74 percent probability of gain, with a modal return of 10.67 percent and a reward-to-risk ratio exceeding 5. Over long horizons, modal CAGRs surpass 13 percent, while worst-case returns remain negative for up to ten years, defining a historical trapping period. This horizon shortens to six years in the post-1999 period, reflecting growing market resilience. Conditional analysis of the P/E ratio reveals regime-dependent outcomes. Low valuations (P/E less than 13) historically show zero probability of loss across all horizons, while high valuations (P/E greater than 27) correspond to unstable returns and extended breakeven periods. To uncover deeper structure, we apply tools from complexity science. Entropy, Hurst exponents, and Lyapunov indicators reveal weak persistence, long memory, and low-dimensional chaos. Information-theoretic metrics, including mutual information and transfer entropy, confirm a directional and predictive influence of valuation on future returns. These findings offer actionable insights for asset allocation, downside risk management, and long-term investment strategy in emerging markets. Our framework bridges valuation, conditional distributions, and nonlinear dynamics in a rigorous and practically relevant manner. ...

August 31, 2025 · 2 min · Research Team

Prospects of Imitating Trading Agents in the Stock Market

Prospects of Imitating Trading Agents in the Stock Market ArXiv ID: 2509.00982 “View on arXiv” Authors: Mateusz Wilinski, Juho Kanniainen Abstract In this work we show how generative tools, which were successfully applied to limit order book data, can be utilized for the task of imitating trading agents. To this end, we propose a modified generative architecture based on the state-space model, and apply it to limit order book data with identified investors. The model is trained on synthetic data, generated from a heterogeneous agent-based model. Finally, we compare model’s predicted distribution over different aspects of investors’ actions, with the ground truths known from the agent-based model. ...

August 31, 2025 · 2 min · Research Team

FinFlowRL: An Imitation-Reinforcement Learning Framework for Adaptive Stochastic Control in Finance

FinFlowRL: An Imitation-Reinforcement Learning Framework for Adaptive Stochastic Control in Finance ArXiv ID: 2510.15883 “View on arXiv” Authors: Yang Li, Zhi Chen Abstract Traditional stochastic control methods in finance struggle in real world markets due to their reliance on simplifying assumptions and stylized frameworks. Such methods typically perform well in specific, well defined environments but yield suboptimal results in changed, non stationary ones. We introduce FinFlowRL, a novel framework for financial optimal stochastic control. The framework pretrains an adaptive meta policy learning from multiple expert strategies, then finetunes through reinforcement learning in the noise space to optimize the generative process. By employing action chunking generating action sequences rather than single decisions, it addresses the non Markovian nature of markets. FinFlowRL consistently outperforms individually optimized experts across diverse market conditions. ...

August 30, 2025 · 2 min · Research Team

ProteuS: A Generative Approach for Simulating Concept Drift in Financial Markets

ProteuS: A Generative Approach for Simulating Concept Drift in Financial Markets ArXiv ID: 2509.11844 “View on arXiv” Authors: Andrés L. Suárez-Cetrulo, Alejandro Cervantes, David Quintana Abstract Financial markets are complex, non-stationary systems where the underlying data distributions can shift over time, a phenomenon known as regime changes, as well as concept drift in the machine learning literature. These shifts, often triggered by major economic events, pose a significant challenge for traditional statistical and machine learning models. A fundamental problem in developing and validating adaptive algorithms is the lack of a ground truth in real-world financial data, making it difficult to evaluate a model’s ability to detect and recover from these drifts. This paper addresses this challenge by introducing a novel framework, named ProteuS, for generating semi-synthetic financial time series with pre-defined structural breaks. Our methodology involves fitting ARMA-GARCH models to real-world ETF data to capture distinct market regimes, and then simulating realistic, gradual, and abrupt transitions between them. The resulting datasets, which include a comprehensive set of technical indicators, provide a controlled environment with a known ground truth of regime changes. An analysis of the generated data confirms the complexity of the task, revealing significant overlap between the different market states. We aim to provide the research community with a tool for the rigorous evaluation of concept drift detection and adaptation mechanisms, paving the way for more robust financial forecasting models. ...

August 30, 2025 · 2 min · Research Team

Robust MCVaR Portfolio Optimization with Ellipsoidal Support and Reproducing Kernel Hilbert Space-based Uncertainty

Robust MCVaR Portfolio Optimization with Ellipsoidal Support and Reproducing Kernel Hilbert Space-based Uncertainty ArXiv ID: 2509.00447 “View on arXiv” Authors: Rupendra Yadav, Aparna Mehra Abstract This study introduces a portfolio optimization framework to minimize mixed conditional value at risk (MCVaR), incorporating a chance constraint on expected returns and limiting the number of assets via cardinality constraints. A robust MCVaR model is presented, which presumes ellipsoidal support for random returns without assuming any distribution. The model utilizes an uncertainty set grounded in a reproducing kernel Hilbert space (RKHS) to manage the chance constraint, resulting in a simplified second-order cone programming (SOCP) formulation. The performance of the robust model is tested on datasets from six distinct financial markets. The outcomes of comprehensive experiments indicate that the robust model surpasses the nominal model, market portfolio, and equal-weight portfolio with higher expected returns, lower risk metrics, enhanced reward-risk ratios, and a better value of Jensen’s alpha in many cases. Furthermore, we aim to validate the robust models in different market phases (bullish, bearish, and neutral). The robust model shows a distinct advantage in bear markets, providing better risk protection against adverse conditions. In contrast, its performance in bullish and neutral phases is somewhat similar to that of the nominal model. The robust model appears effective in volatile markets, although further research is necessary to comprehend its performance across different market conditions. ...

August 30, 2025 · 2 min · Research Team

An Interval Type-2 Version of Bayes Theorem Derived from Interval Probability Range Estimates Provided by Subject Matter Experts

An Interval Type-2 Version of Bayes Theorem Derived from Interval Probability Range Estimates Provided by Subject Matter Experts ArXiv ID: 2509.08834 “View on arXiv” Authors: John T. Rickard, William A. Dembski, James Rickards Abstract Bayesian inference is widely used in many different fields to test hypotheses against observations. In most such applications, an assumption is made of precise input values to produce a precise output value. However, this is unrealistic for real-world applications. Often the best available information from subject matter experts (SMEs) in a given field is interval range estimates of the input probabilities involved in Bayes Theorem. This paper provides two key contributions to extend Bayes Theorem to an interval type-2 (IT2) version. First, we develop an IT2 version of Bayes Theorem that uses a novel and conservative method to avoid potential inconsistencies in the input IT2 MFs that otherwise might produce invalid output results. We then describe a novel and flexible algorithm for encoding SME-provided intervals into IT2 fuzzy membership functions (MFs), which we can use to specify the input probabilities in Bayes Theorem. Our algorithm generalizes and extends previous work on this problem that primarily addressed the encoding of intervals into word MFs for Computing with Words applications. ...

August 29, 2025 · 2 min · Research Team