false

INVESTORBENCH: A Benchmark for Financial Decision-Making Tasks with LLM-based Agent

INVESTORBENCH: A Benchmark for Financial Decision-Making Tasks with LLM-based Agent ArXiv ID: 2412.18174 “View on arXiv” Authors: Unknown Abstract Recent advancements have underscored the potential of large language model (LLM)-based agents in financial decision-making. Despite this progress, the field currently encounters two main challenges: (1) the lack of a comprehensive LLM agent framework adaptable to a variety of financial tasks, and (2) the absence of standardized benchmarks and consistent datasets for assessing agent performance. To tackle these issues, we introduce \textsc{“InvestorBench”}, the first benchmark specifically designed for evaluating LLM-based agents in diverse financial decision-making contexts. InvestorBench enhances the versatility of LLM-enabled agents by providing a comprehensive suite of tasks applicable to different financial products, including single equities like stocks, cryptocurrencies and exchange-traded funds (ETFs). Additionally, we assess the reasoning and decision-making capabilities of our agent framework using thirteen different LLMs as backbone models, across various market environments and tasks. Furthermore, we have curated a diverse collection of open-source, multi-modal datasets and developed a comprehensive suite of environments for financial decision-making. This establishes a highly accessible platform for evaluating financial agents’ performance across various scenarios. ...

December 24, 2024 · 2 min · Research Team

Broker-Trader Partial Information Nash-Equilibria

Broker-Trader Partial Information Nash-Equilibria ArXiv ID: 2412.17712 “View on arXiv” Authors: Unknown Abstract We study partial information Nash equilibrium between a broker and an informed trader. In this setting, the informed trader, who possesses knowledge of a trading signal, trades multiple assets with the broker in a dealer market. Simultaneously, the broker offloads these assets in a lit exchange where their actions impact the asset prices. The broker, however, only observes aggregate prices and cannot distinguish between underlying trends and volatility. Both the broker and the informed trader aim to maximize their penalized expected wealth. Using convex analysis, we characterize the Nash equilibrium and demonstrate its existence and uniqueness. Furthermore, we establish that this equilibrium corresponds to the solution of a nonstandard system of forward-backward stochastic differential equations (FBSDEs) that involves the two differing filtrations. For short enough time horizons, we prove that a unique solution of this system exists. Finally, under quite general assumptions, we show that the solution to the FBSDE system admits a polynomial approximation in the strength of the transient impact to arbitrary order, and prove that the error is controlled. ...

December 23, 2024 · 2 min · Research Team

Collaborative Optimization in Financial Data Mining Through Deep Learning and ResNeXt

Collaborative Optimization in Financial Data Mining Through Deep Learning and ResNeXt ArXiv ID: 2412.17314 “View on arXiv” Authors: Unknown Abstract This study proposes a multi-task learning framework based on ResNeXt, aiming to solve the problem of feature extraction and task collaborative optimization in financial data mining. Financial data usually has the complex characteristics of high dimensionality, nonlinearity, and time series, and is accompanied by potential correlations between multiple tasks, making it difficult for traditional methods to meet the needs of data mining. This study introduces the ResNeXt model into the multi-task learning framework and makes full use of its group convolution mechanism to achieve efficient extraction of local patterns and global features of financial data. At the same time, through the design of task sharing layers and dedicated layers, it is established between multiple related tasks. Deep collaborative optimization relationships. Through flexible multi-task loss weight design, the model can effectively balance the learning needs of different tasks and improve overall performance. Experiments are conducted on a real S&P 500 financial data set, verifying the significant advantages of the proposed framework in classification and regression tasks. The results indicate that, when compared to other conventional deep learning models, the proposed method delivers superior performance in terms of accuracy, F1 score, root mean square error, and other metrics, highlighting its outstanding effectiveness and robustness in handling complex financial data. This research provides an efficient and adaptable solution for financial data mining, and at the same time opens up a new research direction for the combination of multi-task learning and deep learning, which has important theoretical significance and practical application value. ...

December 23, 2024 · 2 min · Research Team

Multimodal Deep Reinforcement Learning for Portfolio Optimization

Multimodal Deep Reinforcement Learning for Portfolio Optimization ArXiv ID: 2412.17293 “View on arXiv” Authors: Unknown Abstract We propose a reinforcement learning (RL) framework that leverages multimodal data including historical stock prices, sentiment analysis, and topic embeddings from news articles, to optimize trading strategies for SP100 stocks. Building upon recent advancements in financial reinforcement learning, we aim to enhance the state space representation by integrating financial sentiment data from SEC filings and news headlines and refining the reward function to better align with portfolio performance metrics. Our methodology includes deep reinforcement learning with state tensors comprising price data, sentiment scores, and news embeddings, processed through advanced feature extraction models like CNNs and RNNs. By benchmarking against traditional portfolio optimization techniques and advanced strategies, we demonstrate the efficacy of our approach in delivering superior portfolio performance. Empirical results showcase the potential of our agent to outperform standard benchmarks, especially when utilizing combined data sources under profit-based reward functions. ...

December 23, 2024 · 2 min · Research Team

Time Series Feature Redundancy Paradox: An Empirical Study Based on Mortgage Default Prediction

Time Series Feature Redundancy Paradox: An Empirical Study Based on Mortgage Default Prediction ArXiv ID: 2501.00034 “View on arXiv” Authors: Unknown Abstract With the widespread application of machine learning in financial risk management, conventional wisdom suggests that longer training periods and more feature variables contribute to improved model performance. This paper, focusing on mortgage default prediction, empirically discovers a phenomenon that contradicts traditional knowledge: in time series prediction, increased training data timespan and additional non-critical features actually lead to significant deterioration in prediction effectiveness. Using Fannie Mae’s mortgage data, the study compares predictive performance across different time window lengths (2012-2022) and feature combinations, revealing that shorter time windows (such as single-year periods) paired with carefully selected key features yield superior prediction results. The experimental results indicate that extended time spans may introduce noise from historical data and outdated market patterns, while excessive non-critical features interfere with the model’s learning of core default factors. This research not only challenges the traditional “more is better” approach in data modeling but also provides new insights and practical guidance for feature selection and time window optimization in financial risk prediction. ...

December 23, 2024 · 2 min · Research Team

Direct Inversion for the Squared Bessel Process and Applications

Direct Inversion for the Squared Bessel Process and Applications ArXiv ID: 2412.16655 “View on arXiv” Authors: Unknown Abstract In this paper we derive a new direct inversion method to simulate squared Bessel processes. Since the transition probability of these processes can be represented by a non-central chi-square distribution, we construct an efficient and accurate algorithm to simulate non-central chi-square variables. In this method, the dimension of the squared Bessel process, equivalently the degrees of freedom of the chi-square distribution, is treated as a variable. We therefore use a two-dimensional Chebyshev expansion to approximate the inverse function of the central chi-square distribution with one variable being the degrees of freedom. The method is accurate and efficient for any value of degrees of freedom including the computationally challenging case of small values. One advantage of the method is that noncentral chi-square samples can be generated for a whole range of values of degrees of freedom using the same Chebyshev coefficients. The squared Bessel process is a building block for the well-known Cox-Ingersoll-Ross (CIR) processes, which can be generated from squared Bessel processes through time change and linear transformation. Our direct inversion method thus allows the efficient and accurate simulation of these processes, which are used as models in a wide variety of applications. ...

December 21, 2024 · 2 min · Research Team

Battery valuation on electricity intraday markets with liquidity costs

Battery valuation on electricity intraday markets with liquidity costs ArXiv ID: 2412.15959 “View on arXiv” Authors: Unknown Abstract In this paper, we propose a complete modelling framework to value several batteries in the electricity intraday market at the trading session scale. The model consists of a stochastic model for the 24 mid-prices (one price per delivery hour) combined with a deterministic model for the liquidity costs (representing the cost of going deeper in the order book). A stochastic optimisation framework based on dynamic programming is used to calculate the value of the batteries. We carry out a back test for the years 2021, 2022 and 2023 for the German market and for the French market. We show that it is essential to take liquidity into account, especially when the number of batteries is large: it allows much higher profits and avoids high losses using our liquidity model. The use of our stochastic model for the mid-price also significantly improves the results (compared to a deterministic framework where the mid-price forecast is the spot price). ...

December 20, 2024 · 2 min · Research Team

Correct implied volatility shapes and reliable pricing in the rough Heston model

Correct implied volatility shapes and reliable pricing in the rough Heston model ArXiv ID: 2412.16067 “View on arXiv” Authors: Unknown Abstract We use modifications of the Adams method and very fast and accurate sinh-acceleration method of the Fourier inversion (iFT) (S.Boyarchenko and Levendorskiĭ, IJTAF 2019, v.22) to evaluate prices of vanilla options; for options of moderate and long maturities and strikes not very far from the spot, thousands of prices can be calculated in several msec. with relative errors of the order of 0.5% and smaller running Matlab on a Mac with moderate characteristics. We demonstrate that for the calibrated set of parameters in Euch and Rosenbaum, Math. Finance 2019, v. 29, the correct implied volatility surface is significantly flatter and fits the data very poorly, hence, the calibration results in op.cit. is an example of the {"\em ghost calibration"} (M.Boyarchenko and Levendorkiĭ, Quantitative Finance 2015, v. 15): the errors of the model and numerical method almost cancel one another. We explain how calibration errors of this sort are generated by each of popular versions of numerical realizations of iFT (Carr-Madan, Lipton-Lewis and COS methods) with prefixed parameters of a numerical method, resulting in spurious volatility smiles and skews. We suggest a general {"\em Conformal Bootstrap principle"} which allows one to avoid ghost calibration errors. We outline schemes of application of Conformal Bootstrap principle and the method of the paper to the design of accurate and fast calibration procedures. ...

December 20, 2024 · 2 min · Research Team

Federated Diffusion Modeling with Differential Privacy for Tabular Data Synthesis

Federated Diffusion Modeling with Differential Privacy for Tabular Data Synthesis ArXiv ID: 2412.16083 “View on arXiv” Authors: Unknown Abstract The increasing demand for privacy-preserving data analytics in various domains necessitates solutions for synthetic data generation that rigorously uphold privacy standards. We introduce the DP-FedTabDiff framework, a novel integration of Differential Privacy, Federated Learning and Denoising Diffusion Probabilistic Models designed to generate high-fidelity synthetic tabular data. This framework ensures compliance with privacy regulations while maintaining data utility. We demonstrate the effectiveness of DP-FedTabDiff on multiple real-world mixed-type tabular datasets, achieving significant improvements in privacy guarantees without compromising data quality. Our empirical evaluations reveal the optimal trade-offs between privacy budgets, client configurations, and federated optimization strategies. The results affirm the potential of DP-FedTabDiff to enable secure data sharing and analytics in highly regulated domains, paving the way for further advances in federated learning and privacy-preserving data synthesis. ...

December 20, 2024 · 2 min · Research Team

Optimizing Fintech Marketing: A Comparative Study of Logistic Regression and XGBoost

Optimizing Fintech Marketing: A Comparative Study of Logistic Regression and XGBoost ArXiv ID: 2412.16333 “View on arXiv” Authors: Unknown Abstract As several studies have shown, predicting credit risk is still a major concern for the financial services industry and is receiving a lot of scholarly interest. This area of study is crucial because it aids financial organizations in determining the probability that borrowers would default, which has a direct bearing on lending choices and risk management tactics. Despite the progress made in this domain, there is still a substantial knowledge gap concerning consumer actions that take place prior to the filing of credit card applications. The objective of this study is to predict customer responses to mail campaigns and assess the likelihood of default among those who engage. This research employs advanced machine learning techniques, specifically logistic regression and XGBoost, to analyze consumer behavior and predict responses to direct mail campaigns. By integrating different data preprocessing strategies, including imputation and binning, we enhance the robustness and accuracy of our predictive models. The results indicate that XGBoost consistently outperforms logistic regression across various metrics, particularly in scenarios using categorical binning and custom imputation. These findings suggest that XGBoost is particularly effective in handling complex data structures and provides a strong predictive capability in assessing credit risk. ...

December 20, 2024 · 2 min · Research Team