false

A deep primal-dual BSDE method for optimal stopping problems

A deep primal-dual BSDE method for optimal stopping problems ArXiv ID: 2409.06937 “View on arXiv” Authors: Unknown Abstract We present a new deep primal-dual backward stochastic differential equation framework based on stopping time iteration to solve optimal stopping problems. A novel loss function is proposed to learn the conditional expectation, which consists of subnetwork parameterization of a continuation value and spatial gradients from present up to the stopping time. Notable features of the method include: (i) The martingale part in the loss function reduces the variance of stochastic gradients, which facilitates the training of the neural networks as well as alleviates the error propagation of value function approximation; (ii) this martingale approximates the martingale in the Doob-Meyer decomposition, and thus leads to a true upper bound for the optimal value in a non-nested Monte Carlo way. We test the proposed method in American option pricing problems, where the spatial gradient network yields the hedging ratio directly. ...

September 11, 2024 · 2 min · Research Team

Market information of the fractional stochastic regularity model

Market information of the fractional stochastic regularity model ArXiv ID: 2409.07159 “View on arXiv” Authors: Unknown Abstract The Fractional Stochastic Regularity Model (FSRM) is an extension of Black-Scholes model describing the multifractal nature of prices. It is based on a multifractional process with a random Hurst exponent $H_t$, driven by a fractional Ornstein-Uhlenbeck (fOU) process. When the regularity parameter $H_t$ is equal to $1/2$, the efficient market hypothesis holds, but when $H_t\neq 1/2$ past price returns contain some information on a future trend or mean-reversion of the log-price process. In this paper, we investigate some properties of the fOU process and, thanks to information theory and Shannon’s entropy, we determine theoretically the serial information of the regularity process $H_t$ of the FSRM, giving some insight into one’s ability to forecast future price increments and to build statistical arbitrages with this model. ...

September 11, 2024 · 2 min · Research Team

Automate Strategy Finding with LLM in Quant Investment

Automate Strategy Finding with LLM in Quant Investment ArXiv ID: 2409.06289 “View on arXiv” Authors: Unknown Abstract We present a novel three-stage framework leveraging Large Language Models (LLMs) within a risk-aware multi-agent system for automate strategy finding in quantitative finance. Our approach addresses the brittleness of traditional deep learning models in financial applications by: employing prompt-engineered LLMs to generate executable alpha factor candidates across diverse financial data, implementing multimodal agent-based evaluation that filters factors based on market status, predictive quality while maintaining category balance, and deploying dynamic weight optimization that adapts to market conditions. Experimental results demonstrate the robust performance of the strategy in Chinese & US market regimes compared to established benchmarks. Our work extends LLMs capabilities to quantitative trading, providing a scalable architecture for financial signal extraction and portfolio construction. The overall framework significantly outperforms all benchmarks with 53.17% cumulative return on SSE50 (Jan 2023 to Jan 2024), demonstrating superior risk-adjusted performance and downside protection on the market. ...

September 10, 2024 · 2 min · Research Team

Limit Order Book Simulation and Trade Evaluation with $K$-Nearest-Neighbor Resampling

Limit Order Book Simulation and Trade Evaluation with $K$-Nearest-Neighbor Resampling ArXiv ID: 2409.06514 “View on arXiv” Authors: Unknown Abstract In this paper, we show how $K$-nearest neighbor ($K$-NN) resampling, an off-policy evaluation method proposed in \cite{“giegrich2023k”}, can be applied to simulate limit order book (LOB) markets and how it can be used to evaluate and calibrate trading strategies. Using historical LOB data, we demonstrate that our simulation method is capable of recreating realistic LOB dynamics and that synthetic trading within the simulation leads to a market impact in line with the corresponding literature. Compared to other statistical LOB simulation methods, our algorithm has theoretical convergence guarantees under general conditions, does not require optimization, is easy to implement and computationally efficient. Furthermore, we show that in a benchmark comparison our method outperforms a deep learning-based algorithm for several key statistics. In the context of a LOB with pro-rata type matching, we demonstrate how our algorithm can calibrate the size of limit orders for a liquidation strategy. Finally, we describe how $K$-NN resampling can be modified for choices of higher dimensional state spaces. ...

September 10, 2024 · 2 min · Research Team

Robust financial calibration: a Bayesian approach for neural SDEs

Robust financial calibration: a Bayesian approach for neural SDEs ArXiv ID: 2409.06551 “View on arXiv” Authors: Unknown Abstract The paper presents a Bayesian framework for the calibration of financial models using neural stochastic differential equations (neural SDEs), for which we also formulate a global universal approximation theorem based on Barron-type estimates. The method is based on the specification of a prior distribution on the neural network weights and an adequately chosen likelihood function. The resulting posterior distribution can be seen as a mixture of different classical neural SDE models yielding robust bounds on the implied volatility surface. Both, historical financial time series data and option price data are taken into consideration, which necessitates a methodology to learn the change of measure between the risk-neutral and the historical measure. The key ingredient for a robust numerical optimization of the neural networks is to apply a Langevin-type algorithm, commonly used in the Bayesian approaches to draw posterior samples. ...

September 10, 2024 · 2 min · Research Team

Critical Dynamics of Random Surfaces: Time Evolution of Area and Genus

Critical Dynamics of Random Surfaces: Time Evolution of Area and Genus ArXiv ID: 2409.05547 “View on arXiv” Authors: Unknown Abstract Conformal field theories with central charge $c\le1$ on random surfaces have been extensively studied in the past. Here, this discussion is extended from their equilibrium distribution to their critical dynamics. This is motivated by the conjecture that these models describe the time evolution of certain social networks that are self-driven to a critical point. This paper focuses on the dynamics of the overall area and the genus of the surface. The time evolution of the area is shown to follow a Cox Ingersol Ross process. Planar surfaces shrink, while higher genus surfaces grow to a size of order of the inverse cosmological constant. The time evolution of the genus is argued to lead to two different phases, dominated by (i) planar surfaces, and (ii) ``foamy’’ surfaces, whose genus diverges. In phase (i), which exhibits critical phenomena, time variations of the order parameter are approximately t-distributed with 4 or more degrees of freedom. ...

September 9, 2024 · 2 min · Research Team

MANA-Net: Mitigating Aggregated Sentiment Homogenization with News Weighting for Enhanced Market Prediction

MANA-Net: Mitigating Aggregated Sentiment Homogenization with News Weighting for Enhanced Market Prediction ArXiv ID: 2409.05698 “View on arXiv” Authors: Unknown Abstract It is widely acknowledged that extracting market sentiments from news data benefits market predictions. However, existing methods of using financial sentiments remain simplistic, relying on equal-weight and static aggregation to manage sentiments from multiple news items. This leads to a critical issue termed ``Aggregated Sentiment Homogenization’’, which has been explored through our analysis of a large financial news dataset from industry practice. This phenomenon occurs when aggregating numerous sentiments, causing representations to converge towards the mean values of sentiment distributions and thereby smoothing out unique and important information. Consequently, the aggregated sentiment representations lose much predictive value of news data. To address this problem, we introduce the Market Attention-weighted News Aggregation Network (MANA-Net), a novel method that leverages a dynamic market-news attention mechanism to aggregate news sentiments for market prediction. MANA-Net learns the relevance of news sentiments to price changes and assigns varying weights to individual news items. By integrating the news aggregation step into the networks for market prediction, MANA-Net allows for trainable sentiment representations that are optimized directly for prediction. We evaluate MANA-Net using the S&P 500 and NASDAQ 100 indices, along with financial news spanning from 2003 to 2018. Experimental results demonstrate that MANA-Net outperforms various recent market prediction methods, enhancing Profit & Loss by 1.1% and the daily Sharpe ratio by 0.252. ...

September 9, 2024 · 2 min · Research Team

Bellwether Trades: Characteristics of Trades influential in Predicting Future Price Movements in Markets

Bellwether Trades: Characteristics of Trades influential in Predicting Future Price Movements in Markets ArXiv ID: 2409.05192 “View on arXiv” Authors: Unknown Abstract In this study, we leverage powerful non-linear machine learning methods to identify the characteristics of trades that contain valuable information. First, we demonstrate the effectiveness of our optimized neural network predictor in accurately predicting future market movements. Then, we utilize the information from this successful neural network predictor to pinpoint the individual trades within each data point (trading window) that had the most impact on the optimized neural network’s prediction of future price movements. This approach helps us uncover important insights about the heterogeneity in information content provided by trades of different sizes, venues, trading contexts, and over time. ...

September 8, 2024 · 2 min · Research Team

QuantFactor REINFORCE: Mining Steady Formulaic Alpha Factors with Variance-bounded REINFORCE

QuantFactor REINFORCE: Mining Steady Formulaic Alpha Factors with Variance-bounded REINFORCE ArXiv ID: 2409.05144 “View on arXiv” Authors: Unknown Abstract Alpha factor mining aims to discover investment signals from the historical financial market data, which can be used to predict asset returns and gain excess profits. Powerful deep learning methods for alpha factor mining lack interpretability, making them unacceptable in the risk-sensitive real markets. Formulaic alpha factors are preferred for their interpretability, while the search space is complex and powerful explorative methods are urged. Recently, a promising framework is proposed for generating formulaic alpha factors using deep reinforcement learning, and quickly gained research focuses from both academia and industries. This paper first argues that the originally employed policy training method, i.e., Proximal Policy Optimization (PPO), faces several important issues in the context of alpha factors mining. Herein, a novel reinforcement learning algorithm based on the well-known REINFORCE algorithm is proposed. REINFORCE employs Monte Carlo sampling to estimate the policy gradient-yielding unbiased but high variance estimates. The minimal environmental variability inherent in the underlying state transition function, which adheres to the Dirac distribution, can help alleviate this high variance issue, making REINFORCE algorithm more appropriate than PPO. A new dedicated baseline is designed to theoretically reduce the commonly suffered high variance of REINFORCE. Moreover, the information ratio is introduced as a reward shaping mechanism to encourage the generation of steady alpha factors that can better adapt to changes in market volatility. Evaluations on real assets data indicate the proposed algorithm boosts correlation with returns by 3.83%, and a stronger ability to obtain excess returns compared to the latest alpha factors mining methods, which meets the theoretical results well. ...

September 8, 2024 · 2 min · Research Team

DEPLOYERS: An agent based modeling tool for multi country real world data

DEPLOYERS: An agent based modeling tool for multi country real world data ArXiv ID: 2409.04876 “View on arXiv” Authors: Unknown Abstract We present recent progress in the design and development of DEPLOYERS, an agent-based macroeconomics modeling (ABM) framework, capable to deploy and simulate a full economic system (individual workers, goods and services firms, government, central and private banks, financial market, external sectors) whose structure and activity analysis reproduce the desired calibration data, that can be, for example a Social Accounting Matrix (SAM) or a Supply-Use Table (SUT) or an Input-Output Table (IOT).Here we extend our previous work to a multi-country version and show an example using data from a 46-countries 64-sectors FIGARO Inter-Country IOT. The simulation of each country runs on a separate thread or CPU core to simulate the activity of one step (month, week, or day) and then interacts (updates imports, exports, transfer) with that country’s foreign partners, and proceeds to the next step. This interaction can be chosen to be aggregated (a single row and column IO account) or disaggregated (64 rows and columns) with each partner. A typical run simulates thousands of individuals and firms engaged in their monthly activity and then records the results, much like a survey of the country’s economic system. This data can then be subjected to, for example, an Input-Output analysis to find out the sources of observed stylized effects as a function of time in the detailed and realistic modeling environment that can be easily implemented in an ABM framework. ...

September 7, 2024 · 2 min · Research Team