false

GraphCNNpred: A stock market indices prediction using a Graph based deep learning system

GraphCNNpred: A stock market indices prediction using a Graph based deep learning system ArXiv ID: 2407.03760 “View on arXiv” Authors: Unknown Abstract The application of deep learning techniques for predicting stock market prices is a prominent and widely researched topic in the field of data science. To effectively predict market trends, it is essential to utilize a diversified dataset. In this paper, we give a graph neural network based convolutional neural network (CNN) model, that can be applied on diverse source of data, in the attempt to extract features to predict the trends of indices of \text{“S”}&\text{“P”} 500, NASDAQ, DJI, NYSE, and RUSSEL. The experiments show that the associated models improve the performance of prediction in all indices over the baseline algorithms by about $4% \text{" to “} 15%$, in terms of F-measure. A trading simulation is generated from predictions and gained a Sharpe ratio of over 3. ...

July 4, 2024 · 2 min · Research Team

The Structure of Financial Equity Research Reports -- Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4

The Structure of Financial Equity Research Reports – Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4 ArXiv ID: 2407.18327 “View on arXiv” Authors: Unknown Abstract This research dissects financial equity research reports (ERRs) by mapping their content into categories. There is insufficient empirical analysis of the questions answered in ERRs. In particular, it is not understood how frequently certain information appears, what information is considered essential, and what information requires human judgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940 sentences into 169 unique question archetypes. We did not predefine the questions but derived them solely from the statements in the ERRs. This approach provides an unbiased view of the content of the observed ERRs. Subsequently, we used public corporate reports to classify the questions’ potential for automation. Answers were labeled “text-extractable” if the answers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question consist of 48.2% text-extractable (suited to processing by large language models, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions require human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that recent advances in language generation and information extraction enable the automation of approximately 80% of the statements in ERRs. Surprisingly, the models complement each other’s strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely benefit from additional automation, improving quality and efficiency. The research thus allows us to quantify the potential impacts of introducing large language models in the ERR writing process. The full question list, including the archetypes and their frequency, will be made available online after peer review. ...

July 4, 2024 · 3 min · Research Team

When can weak latent factors be statistically inferred?

When can weak latent factors be statistically inferred? ArXiv ID: 2407.03616 “View on arXiv” Authors: Unknown Abstract This article establishes a new and comprehensive estimation and inference theory for principal component analysis (PCA) under the weak factor model that allow for cross-sectional dependent idiosyncratic components under the nearly minimal factor strength relative to the noise level or signal-to-noise ratio. Our theory is applicable regardless of the relative growth rate between the cross-sectional dimension $N$ and temporal dimension $T$. This more realistic assumption and noticeable result require completely new technical device, as the commonly-used leave-one-out trick is no longer applicable to the case with cross-sectional dependence. Another notable advancement of our theory is on PCA inference $ - $ for example, under the regime where $N\asymp T$, we show that the asymptotic normality for the PCA-based estimator holds as long as the signal-to-noise ratio (SNR) grows faster than a polynomial rate of $\log N$. This finding significantly surpasses prior work that required a polynomial rate of $N$. Our theory is entirely non-asymptotic, offering finite-sample characterizations for both the estimation error and the uncertainty level of statistical inference. A notable technical innovation is our closed-form first-order approximation of PCA-based estimator, which paves the way for various statistical tests. Furthermore, we apply our theories to design easy-to-implement statistics for validating whether given factors fall in the linear spans of unknown latent factors, testing structural breaks in the factor loadings for an individual unit, checking whether two units have the same risk exposures, and constructing confidence intervals for systematic risks. Our empirical studies uncover insightful correlations between our test results and economic cycles. ...

July 4, 2024 · 2 min · Research Team

AMA-LSTM: Pioneering Robust and Fair Financial Audio Analysis for Stock Volatility Prediction

AMA-LSTM: Pioneering Robust and Fair Financial Audio Analysis for Stock Volatility Prediction ArXiv ID: 2407.18324 “View on arXiv” Authors: Unknown Abstract Stock volatility prediction is an important task in the financial industry. Recent advancements in multimodal methodologies, which integrate both textual and auditory data, have demonstrated significant improvements in this domain, such as earnings calls (Earnings calls are public available and often involve the management team of a public company and interested parties to discuss the company’s earnings). However, these multimodal methods have faced two drawbacks. First, they often fail to yield reliable models and overfit the data due to their absorption of stochastic information from the stock market. Moreover, using multimodal models to predict stock volatility suffers from gender bias and lacks an efficient way to eliminate such bias. To address these aforementioned problems, we use adversarial training to generate perturbations that simulate the inherent stochasticity and bias, by creating areas resistant to random information around the input space to improve model robustness and fairness. Our comprehensive experiments on two real-world financial audio datasets reveal that this method exceeds the performance of current state-of-the-art solution. This confirms the value of adversarial training in reducing stochasticity and bias for stock volatility prediction tasks. ...

July 3, 2024 · 2 min · Research Team

Basket Options with Volatility Skew: Calibrating a Local Volatility Model by Sample Rearrangement

Basket Options with Volatility Skew: Calibrating a Local Volatility Model by Sample Rearrangement ArXiv ID: 2407.02901 “View on arXiv” Authors: Unknown Abstract The pricing of derivatives tied to baskets of assets demands a sophisticated framework that aligns with the available market information to capture the intricate non-linear dependency structure among the assets. We describe the dynamics of the multivariate process of constituents with a copula model and propose an efficient method to extract the dependency structure from the market. The proposed method generates coherent sets of samples of the constituents process through systematic sampling rearrangement. These samples are then utilized to calibrate a local volatility model (LVM) of the basket process, which is used to price basket derivatives. We show that the method is capable of efficiently pricing basket options based on a large number of basket constituents, accomplishing the calibration process within a matter of seconds, and achieving near-perfect calibration to the index options of the market. ...

July 3, 2024 · 2 min · Research Team

Examples and Counterexamples of Cost-efficiency in Incomplete Markets

Examples and Counterexamples of Cost-efficiency in Incomplete Markets ArXiv ID: 2407.08756 “View on arXiv” Authors: Unknown Abstract We present a number of examples and counterexamples to illustrate the results on cost-efficiency in an incomplete market obtained in [“BS24”]. These examples and counterexamples do not only illustrate the results obtained in [“BS24”], but show the limitations of the results and the sharpness of the key assumptions. In particular, we make use of a simple 3-state model in which we are able to recover and illustrate all key results of the paper. This example also shows how our characterization of perfectly cost-efficient claims allows to solve an expected utility maximization problem in a simple incomplete market (trinomial model) and recover results from [“DS06, Chapter 3”], there obtained using duality. ...

July 3, 2024 · 2 min · Research Team

Robust optimal investment and consumption strategies with portfolio constraints and stochastic environment

Robust optimal investment and consumption strategies with portfolio constraints and stochastic environment ArXiv ID: 2407.02831 “View on arXiv” Authors: Unknown Abstract We investigate a continuous-time investment-consumption problem with model uncertainty in a general diffusion-based market with random model coefficients. We assume that a power utility investor is ambiguity-averse, with the preference to robustness captured by the homothetic multiplier robust specification, and the investor’s investment and consumption strategies are constrained to closed convex sets. To solve this constrained robust control problem, we employ the stochastic Hamilton-Jacobi-Bellman-Isaacs equations, backward stochastic differential equations, and bounded mean oscillation martingale theory. Furthermore, we show the investor incurs (non-negative) utility loss, i.e. the loss in welfare, if model uncertainty is ignored. When the model coefficients are deterministic, we establish formally the relationship between the investor’s robustness preference and the robust optimal investment-consumption strategy and the value function, and the impact of investment and consumption constraints on the investor’s robust optimal investment-consumption strategy and value function. Extensive numerical experiments highlight the significant impact of ambiguity aversion, consumption and investment constraints, on the investor’s robust optimal investment-consumption strategy, utility loss, and value function. Key findings include: 1) short-selling restriction always reduces the investor’s utility loss when model uncertainty is ignored; 2) the effect of consumption constraints on utility loss is more delicate and relies on the investor’s risk aversion level. ...

July 3, 2024 · 2 min · Research Team

CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications

CatMemo at the FinLLM Challenge Task: Fine-Tuning Large Language Models using Data Fusion in Financial Applications ArXiv ID: 2407.01953 “View on arXiv” Authors: Unknown Abstract The integration of Large Language Models (LLMs) into financial analysis has garnered significant attention in the NLP community. This paper presents our solution to IJCAI-2024 FinLLM challenge, investigating the capabilities of LLMs within three critical areas of financial tasks: financial classification, financial text summarization, and single stock trading. We adopted Llama3-8B and Mistral-7B as base models, fine-tuning them through Parameter Efficient Fine-Tuning (PEFT) and Low-Rank Adaptation (LoRA) approaches. To enhance model performance, we combine datasets from task 1 and task 2 for data fusion. Our approach aims to tackle these diverse tasks in a comprehensive and integrated manner, showcasing LLMs’ capacity to address diverse and complex financial tasks with improved accuracy and decision-making capabilities. ...

July 2, 2024 · 2 min · Research Team

Indian Stock Market Prediction using Augmented Financial Intelligence ML

Indian Stock Market Prediction using Augmented Financial Intelligence ML ArXiv ID: 2407.02236 “View on arXiv” Authors: Unknown Abstract This paper presents price prediction models using Machine Learning algorithms augmented with Superforecasters predictions, aimed at enhancing investment decisions. Five Machine Learning models are built, including Bidirectional LSTM, ARIMA, a combination of CNN and LSTM, GRU, and a model built using LSTM and GRU algorithms. The models are evaluated using the Mean Absolute Error to determine their predictive accuracy. Additionally, the paper suggests incorporating human intelligence by identifying Superforecasters and tracking their predictions to anticipate unpredictable shifts or changes in stock prices . The predictions made by these users can further enhance the accuracy of stock price predictions when combined with Machine Learning and Natural Language Processing techniques. Predicting the price of any commodity can be a significant task but predicting the price of a stock in the stock market deals with much more uncertainty. Recognising the limited knowledge and exposure to stocks among certain investors, this paper proposes price prediction models using Machine Learning algorithms. In this work, five Machine learning models are built using Bidirectional LSTM, ARIMA, a combination of CNN and LSTM, GRU and the last one is built using LSTM and GRU algorithms. Later these models are assessed using MAE scores to find which model is predicting with the highest accuracy. In addition to this, this paper also suggests the use of human intelligence to closely predict the shift in price patterns in the stock market The main goal is to identify Superforecasters and track their predictions to anticipate unpredictable shifts or changes in stock prices. By leveraging the combined power of Machine Learning and the Human Intelligence, predictive accuracy can be significantly increased. ...

July 2, 2024 · 2 min · Research Team

The Credit Markets Go Dark

The Credit Markets Go Dark ArXiv ID: ssrn-4879742 “View on arXiv” Authors: Unknown Abstract Keywords: Capital Structure, Corporate Debt, Equity Ownership, Fixed Income, Fixed Income Complexity vs Empirical Score Math Complexity: 1.0/10 Empirical Rigor: 2.0/10 Quadrant: Philosophers Why: The paper is a legal and economic analysis discussing trends in corporate debt ownership and the rise of private credit, relying on narrative data and industry observations without complex mathematical modeling or backtested implementations. flowchart TD A["Research Goal: Analyze diverging trends in<br>equity vs. corporate debt ownership"] --> B["Key Methodology: Empirical & Theoretical<br>Analysis of Institutional Holdings"] B --> C["Data Input: Decades of<br>Equity & Debt Ownership Data"] C --> D["Computational Process:<br>Quantitative Comparison & Trend Analysis"] D --> E["Key Finding: Equity ownership<br>is widely dispersed (institutional rise)"] D --> F["Key Finding: Corporate debt ownership<br>concentrated in opaque 'shadow banking'"] E --> G["Outcome: Credit markets 'go dark'<br>with transparency and liquidity"] F --> G

July 2, 2024 · 1 min · Research Team