false

Minimal Shortfall Strategies for Liquidation of a Basket of Stocks using Reinforcement Learning

Minimal Shortfall Strategies for Liquidation of a Basket of Stocks using Reinforcement Learning ArXiv ID: 2502.07868 “View on arXiv” Authors: Unknown Abstract This paper studies the ubiquitous problem of liquidating large quantities of highly correlated stocks, a task frequently encountered by institutional investors and proprietary trading firms. Traditional methods in this setting suffer from the curse of dimensionality, making them impractical for high-dimensional problems. In this work, we propose a novel method based on stochastic optimal control to optimally tackle this complex multidimensional problem. The proposed method minimizes the overall execution shortfall of highly correlated stocks using a reinforcement learning approach. We rigorously establish the convergence of our optimal trading strategy and present an implementation of our algorithm using intra-day market data. ...

February 11, 2025 · 2 min · Research Team

TWICE: What Advantages Can Low-Resource Domain-Specific Embedding Model Bring? -- A Case Study on Korea Financial Texts

TWICE: What Advantages Can Low-Resource Domain-Specific Embedding Model Bring? – A Case Study on Korea Financial Texts ArXiv ID: 2502.07131 “View on arXiv” Authors: Unknown Abstract Domain specificity of embedding models is critical for effective performance. However, existing benchmarks, such as FinMTEB, are primarily designed for high-resource languages, leaving low-resource settings, such as Korean, under-explored. Directly translating established English benchmarks often fails to capture the linguistic and cultural nuances present in low-resource domains. In this paper, titled TWICE: What Advantages Can Low-Resource Domain-Specific Embedding Models Bring? A Case Study on Korea Financial Texts, we introduce KorFinMTEB, a novel benchmark for the Korean financial domain, specifically tailored to reflect its unique cultural characteristics in low-resource languages. Our experimental results reveal that while the models perform robustly on a translated version of FinMTEB, their performance on KorFinMTEB uncovers subtle yet critical discrepancies, especially in tasks requiring deeper semantic understanding, that underscore the limitations of direct translation. This discrepancy highlights the necessity of benchmarks that incorporate language-specific idiosyncrasies and cultural nuances. The insights from our study advocate for the development of domain-specific evaluation frameworks that can more accurately assess and drive the progress of embedding models in low-resource settings. ...

February 10, 2025 · 2 min · Research Team

Testing for the Minimum Mean-Variance Spanning Set

Testing for the Minimum Mean-Variance Spanning Set ArXiv ID: 2501.19213 “View on arXiv” Authors: Unknown Abstract This paper explores the estimation and inference of the minimum spanning set (MSS), the smallest subset of risky assets that spans the mean-variance efficient frontier of the full asset set. We establish identification conditions for the MSS and develop a novel procedure for its estimation and inference. Our theoretical analysis shows that the proposed MSS estimator covers the true MSS with probability approaching 1 and converges asymptotically to the true MSS at any desired confidence level, such as 0.95 or 0.99. Monte Carlo simulations confirm the strong finite-sample performance of the MSS estimator. We apply our method to evaluate the relative importance of individual stock momentum and factor momentum strategies, along with a set of well-established stock return factors. The empirical results highlight factor momentum, along with several stock momentum and return factors, as key drivers of mean-variance efficiency. Furthermore, our analysis uncovers the sources of contribution from these factors and provides a ranking of their relative importance, offering new insights into their roles in mean-variance analysis. ...

January 31, 2025 · 2 min · Research Team

Forecasting S&P 500 Using LSTM Models

Forecasting S&P 500 Using LSTM Models ArXiv ID: 2501.17366 “View on arXiv” Authors: Unknown Abstract With the volatile and complex nature of financial data influenced by external factors, forecasting the stock market is challenging. Traditional models such as ARIMA and GARCH perform well with linear data but struggle with non-linear dependencies. Machine learning and deep learning models, particularly Long Short-Term Memory (LSTM) networks, address these challenges by capturing intricate patterns and long-term dependencies. This report compares ARIMA and LSTM models in predicting the S&P 500 index, a major financial benchmark. Using historical price data and technical indicators, we evaluated these models using Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE). The ARIMA model showed reasonable performance with an MAE of 462.1, RMSE of 614, and 89.8 percent accuracy, effectively capturing short-term trends but limited by its linear assumptions. The LSTM model, leveraging sequential processing capabilities, outperformed ARIMA with an MAE of 369.32, RMSE of 412.84, and 92.46 percent accuracy, capturing both short- and long-term dependencies. Notably, the LSTM model without additional features performed best, achieving an MAE of 175.9, RMSE of 207.34, and 96.41 percent accuracy, showcasing its ability to handle market data efficiently. Accurately predicting stock movements is crucial for investment strategies, risk assessments, and market stability. Our findings confirm the potential of deep learning models in handling volatile financial data compared to traditional ones. The results highlight the effectiveness of LSTM and suggest avenues for further improvements. This study provides insights into financial forecasting, offering a comparative analysis of ARIMA and LSTM while outlining their strengths and limitations. ...

January 29, 2025 · 2 min · Research Team

Reinforcement-Learning Portfolio Allocation with Dynamic Embedding of Market Information

Reinforcement-Learning Portfolio Allocation with Dynamic Embedding of Market Information ArXiv ID: 2501.17992 “View on arXiv” Authors: Unknown Abstract We develop a portfolio allocation framework that leverages deep learning techniques to address challenges arising from high-dimensional, non-stationary, and low-signal-to-noise market information. Our approach includes a dynamic embedding method that reduces the non-stationary, high-dimensional state space into a lower-dimensional representation. We design a reinforcement learning (RL) framework that integrates generative autoencoders and online meta-learning to dynamically embed market information, enabling the RL agent to focus on the most impactful parts of the state space for portfolio allocation decisions. Empirical analysis based on the top 500 U.S. stocks demonstrates that our framework outperforms common portfolio benchmarks and the predict-then-optimize (PTO) approach using machine learning, particularly during periods of market stress. Traditional factor models do not fully explain this superior performance. The framework’s ability to time volatility reduces its market exposure during turbulent times. Ablation studies confirm the robustness of this performance across various reinforcement learning algorithms. Additionally, the embedding and meta-learning techniques effectively manage the complexities of high-dimensional, noisy, and non-stationary financial data, enhancing both portfolio performance and risk management. ...

January 29, 2025 · 2 min · Research Team

Transformer Based Time-Series Forecasting for Stock

Transformer Based Time-Series Forecasting for Stock ArXiv ID: 2502.09625 “View on arXiv” Authors: Unknown Abstract To the naked eye, stock prices are considered chaotic, dynamic, and unpredictable. Indeed, it is one of the most difficult forecasting tasks that hundreds of millions of retail traders and professional traders around the world try to do every second even before the market opens. With recent advances in the development of machine learning and the amount of data the market generated over years, applying machine learning techniques such as deep learning neural networks is unavoidable. In this work, we modeled the task as a multivariate forecasting problem, instead of a naive autoregression problem. The multivariate analysis is done using the attention mechanism via applying a mutated version of the Transformer, “Stockformer”, which we created. ...

January 29, 2025 · 2 min · Research Team

Exploratory Mean-Variance Portfolio Optimization with Regime-Switching Market Dynamics

Exploratory Mean-Variance Portfolio Optimization with Regime-Switching Market Dynamics ArXiv ID: 2501.16659 “View on arXiv” Authors: Unknown Abstract Considering the continuous-time Mean-Variance (MV) portfolio optimization problem, we study a regime-switching market setting and apply reinforcement learning (RL) techniques to assist informed exploration within the control space. We introduce and solve the Exploratory Mean Variance with Regime Switching (EMVRS) problem. We also present a Policy Improvement Theorem. Further, we recognize that the widely applied Temporal Difference (TD) learning is not adequate for the EMVRS context, hence we consider Orthogonality Condition (OC) learning, leveraging the martingale property of the induced optimal value function from the analytical solution to EMVRS. We design a RL algorithm that has more meaningful parameterization using the market parameters and propose an updating scheme for each parameter. Our empirical results demonstrate the superiority of OC learning over TD learning with a clear convergence of the market parameters towards their corresponding ``grounding true" values in a simulated market scenario. In a real market data study, EMVRS with OC learning outperforms its counterparts with the highest mean and reasonably low volatility of the annualized portfolio returns. ...

January 28, 2025 · 2 min · Research Team

Why is the estimation of metaorder impact with public market data so challenging?

Why is the estimation of metaorder impact with public market data so challenging? ArXiv ID: 2501.17096 “View on arXiv” Authors: Unknown Abstract Estimating market impact and transaction costs of large trades (metaorders) is a very important topic in finance. However, using models of price and trade based on public market data provide average price trajectories which are qualitatively different from what is observed during real metaorder executions: the price increases linearly, rather than in a concave way, during the execution and the amount of reversion after its end is very limited. We claim that this is a generic phenomenon due to the fact that even sophisticated statistical models are unable to correctly describe the origin of the autocorrelation of the order flow. We propose a modified Transient Impact Model which provides more realistic trajectories by assuming that only a fraction of the metaorder trading triggers market order flow. Interestingly, in our model there is a critical condition on the kernels of the price and order flow equations in which market impact becomes permanent. ...

January 28, 2025 · 2 min · Research Team

Marketron games: Self-propelling stocks vs dumb money and metastable dynamics of the Good, Bad and Ugly markets

Marketron games: Self-propelling stocks vs dumb money and metastable dynamics of the Good, Bad and Ugly markets ArXiv ID: 2501.12676 “View on arXiv” Authors: Unknown Abstract We present a model of price formation in an inelastic market whose dynamics are partially driven by both money flows and their impact on asset prices. The money flow to the market is viewed as an investment policy of outside investors. For the price impact effect, we use an impact function that incorporates the phenomena of market inelasticity and saturation from new money (the $dumb ; money$ effect). Due to the dependence of market investors’ flows on market performance, the model implies a feedback mechanism that gives rise to nonlinear dynamics. Consequently, the market price dynamics are seen as a nonlinear diffusion of a particle (the $marketron$) in a two-dimensional space formed by the log-price $x$ and a memory variable $y$. The latter stores information about past money flows, so that the dynamics are non-Markovian in the log price $x$ alone, but Markovian in the pair $(x,y)$, bearing a strong resemblance to spiking neuron models in neuroscience. In addition to market flows, the model dynamics are partially driven by return predictors, modeled as unobservable Ornstein-Uhlenbeck processes. By using a new interpretation of predictive signals as $self$-$propulsion$ components of the price dynamics, we treat the marketron as an active particle, amenable to methods developed in the physics of active matter. We show that, depending on the choice of parameters, our model can produce a rich variety of interesting dynamic scenarios. In particular, it predicts three distinct regimes of the market, which we call the $Good$, the $Bad$, and the $Ugly$ markets. The latter regime describes a scenario of a total market collapse or, alternatively, a corporate default event, depending on whether our model is applied to the whole market or an individual stock. ...

January 22, 2025 · 3 min · Research Team

Optimal Rebate Design: Incentives, Competition and Efficiency in Auction Markets

Optimal Rebate Design: Incentives, Competition and Efficiency in Auction Markets ArXiv ID: 2501.12591 “View on arXiv” Authors: Unknown Abstract This study explores the design of an efficient rebate policy in auction markets, focusing on a continuous-time setting with competition among market participants. In this model, a stock exchange collects transaction fees from auction investors executing block trades to buy or sell a risky asset, then redistributes these fees as rebates to competing market makers submitting limit orders. Market makers influence both the price at which the asset trades and their arrival intensity in the auction. We frame this problem as a principal-multi-agent problem and provide necessary and sufficient conditions to characterize the Nash equilibrium among market makers. The exchange’s optimization problem is formulated as a high-dimensional Hamilton-Jacobi-Bellman equation with Poisson jump processes, which is solved using a verification result. To numerically compute the optimal rebate and transaction fee policies, we apply the Deep BSDE method. Our results show that optimal transaction fees and rebate structures improve market efficiency by narrowing the spread between the auction clearing price and the asset’s fundamental value, while ensuring a minimal gain for both market makers indexed on the price of the asset on a coexisting limit order book. ...

January 22, 2025 · 2 min · Research Team