financial statement
FinStat2SQL: A Text2SQL Pipeline for Financial Statement Analysis
Nguyen, Quang Hung, Trinh, Phuong Anh, Mai, Phan Quoc Hung, Trinh, Tuan Phong
Despite the advancements of large language models, text2sql still faces many challenges, particularly with complex and domain-specific queries. In finance, database designs and financial reporting layouts vary widely between financial entities and countries, making text2sql even more challenging. We present FinStat2SQL, a lightweight text2sql pipeline enabling natural language queries over financial statements. Tailored to local standards like VAS, it combines large and small language models in a multi-agent setup for entity extraction, SQL generation, and self-correction. We build a domain-specific database and evaluate models on a synthetic QA dataset. A fine-tuned 7B model achieves 61.33\% accuracy with sub-4-second response times on consumer hardware, outperforming GPT-4o-mini. FinStat2SQL offers a scalable, cost-efficient solution for financial analysis, making AI-powered querying accessible to Vietnamese enterprises.
- Asia > Vietnam (0.05)
- Europe > United Kingdom > England > Greater London > London (0.04)
- North America > United States (0.04)
- (3 more...)
- Research Report (0.82)
- Overview (0.68)
Credit Network Modeling and Analysis via Large Language Models
Sun, Enbo, Wang, Yongzhao, Zhou, Hao
We investigate the application of large language models (LLMs) to construct credit networks from firms' textual financial statements and to analyze the resulting network structures. We start with using LLMs to translate each firm's financial statement into a credit network that pertains solely to that firm. These networks are then aggregated to form a comprehensive credit network representing the whole financial system. During this process, the inconsistencies in financial statements are automatically detected and human intervention is involved. We demonstrate that this translation process is effective across financial statements corresponding to credit networks with diverse topological structures. We further investigate the reasoning capabilities of LLMs in analyzing credit networks and determining optimal strategies for executing financial operations to maximize network performance measured by the total assets of firms, which is an inherently combinatorial optimization challenge. To demonstrate this capability, we focus on two financial operations: portfolio compression and debt removal, applying them to both synthetic and real-world datasets. Our findings show that LLMs can generate coherent reasoning and recommend effective executions of these operations to enhance overall network performance.
- Europe > United Kingdom > England > Merseyside > Liverpool (0.40)
- Asia > China > Jiangxi Province > Nanchang (0.04)
- North America > United States > Florida > Duval County > Jacksonville (0.04)
MultiFinBen: Benchmarking Large Language Models for Multilingual and Multimodal Financial Application
Peng, Xueqing, Qian, Lingfei, Wang, Yan, Xiang, Ruoyu, He, Yueru, Ren, Yang, Jiang, Mingyang, Zhang, Vincent Jim, Guo, Yuqing, Zhao, Jeff, He, Huan, Han, Yi, Feng, Yun, Jiang, Yuechen, Cao, Yupeng, Li, Haohang, Yu, Yangyang, Wang, Xiaoyu, Gao, Penglei, Lin, Shengyuan, Wang, Keyi, Yang, Shanshan, Zhao, Yilun, Liu, Zhiwei, Lu, Peng, Huang, Jerry, Wang, Suyuchen, Papadopoulos, Triantafillos, Giannouris, Polydoros, Soufleri, Efstathia, Chen, Nuo, Deng, Zhiyang, Fu, Heming, Zhao, Yijia, Lin, Mingquan, Qiu, Meikang, Smith, Kaleb E, Cohan, Arman, Liu, Xiao-Yang, Huang, Jimin, Xiong, Guojun, Lopez-Lira, Alejandro, Chen, Xi, Tsujii, Junichi, Nie, Jian-Yun, Ananiadou, Sophia, Xie, Qianqian
Real-world financial analysis involves information across multiple languages and modalities, from reports and news to scanned filings and meeting recordings. Yet most existing evaluations of LLMs in finance remain text-only, monolingual, and largely saturated by current models. To bridge these gaps, we present MultiFinBen, the first expert-annotated multilingual (five languages) and multimodal (text, vision, audio) benchmark for evaluating LLMs in realistic financial contexts. MultiFinBen introduces two new task families: multilingual financial reasoning, which tests cross-lingual evidence integration from filings and news, and financial OCR, which extracts structured text from scanned documents containing tables and charts. Rather than aggregating all available datasets, we apply a structured, difficulty-aware selection based on advanced model performance, ensuring balanced challenge and removing redundant tasks. Evaluating 21 leading LLMs shows that even frontier multimodal models like GPT-4o achieve only 46.01% overall, stronger on vision and audio but dropping sharply in multilingual settings. These findings expose persistent limitations in multilingual, multimodal, and expert-level financial reasoning. All datasets, evaluation scripts, and leaderboards are publicly released.
- Europe > Austria > Vienna (0.14)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Asia > Middle East > Jordan (0.04)
- (12 more...)
- Financial News (1.00)
- Research Report > New Finding (0.45)
- Government (1.00)
- Banking & Finance > Trading (1.00)
- Information Technology > Security & Privacy (0.92)
- (2 more...)
AuditAgent: Expert-Guided Multi-Agent Reasoning for Cross-Document Fraudulent Evidence Discovery
Bai, Songran, Wu, Bingzhe, Zhang, Yiwei, Wu, Chengke, Zheng, Xiaolong, Yuan, Yaze, Wu, Ke, Li, Jianqiang
Financial fraud detection in real-world scenarios presents significant challenges due to the subtlety and dispersion of evidence across complex, multi-year financial disclosures. In this work, we introduce a novel multi-agent reasoning framework AuditAgent, enhanced with auditing domain expertise, for fine-grained evidence chain localization in financial fraud cases. Leveraging an expert-annotated dataset constructed from enforcement documents and financial reports released by the China Securities Regulatory Commission, our approach integrates subject-level risk priors, a hybrid retrieval strategy, and specialized agent modules to efficiently identify and aggregate cross-report evidence. Extensive experiments demonstrate that our method substantially outperforms General-Purpose Agent paradigm in both recall and interpretability, establishing a new benchmark for automated, transparent financial forensics. Our results highlight the value of domain-specific reasoning and dataset construction for advancing robust financial fraud detection in practical, real-world regulatory applications.
- North America > United States > New York > New York County > New York City (0.14)
- Asia > China > Guangdong Province > Shenzhen (0.05)
- Asia > Singapore (0.05)
- (4 more...)
- Banking & Finance (1.00)
- Law Enforcement & Public Safety > Fraud (0.89)
AlphaX: An AI-Based Value Investing Strategy for the Brazilian Stock Market
Autonomous trading strategies have been a subject of research within the field of artificial intelligence (AI) for aconsiderable period. Various AI techniques have been explored to develop autonomous agents capable of trading financial assets. These approaches encompass traditional methods such as neural networks, fuzzy logic, and reinforcement learning, as well as more recent advancements, including deep neural networks and deep reinforcement learning. Many developers report success in creating strategies that exhibit strong performance during simulations using historical price data, a process commonly referred to as backtesting. However, when these strategies are deployed in real markets, their performance often deteriorates, particularly in terms of risk-adjusted returns. In this study, we propose an AI-based strategy inspired by a classical investment paradigm: Value Investing. Financial AI models are highly susceptible to lookahead bias and other forms of bias that can significantly inflate performance in backtesting compared to live trading conditions. To address this issue, we conducted a series of computational simulations while controlling for these biases, thereby reducing the risk of overfitting. Our results indicate that the proposed approach outperforms major Brazilian market benchmarks. Moreover, the strategy, named AlphaX, demonstrated superior performance relative to widely used technical indicators such as the Relative Strength Index (RSI) and Money Flow Index (MFI), with statistically significant results. Finally, we discuss several open challenges and highlight emerging technologies in qualitative analysis that may contribute to the development of a comprehensive AI-based Value Investing framework in the future
- North America > United States > New Jersey > Hudson County > Hoboken (0.04)
- South America > Brazil > Rio de Janeiro > Rio de Janeiro (0.04)
- North America > United States > New York > Kings County > New York City (0.04)
- (3 more...)
FinMaster: A Holistic Benchmark for Mastering Full-Pipeline Financial Workflows with LLMs
Jiang, Junzhe, Yang, Chang, Cui, Aixin, Jin, Sihan, Wang, Ruiyu, Li, Bo, Huang, Xiao, Sun, Dongning, Wang, Xinrun
Financial tasks are pivotal to global economic stability; however, their execution faces challenges including labor intensive processes, low error tolerance, data fragmentation, and tool limitations. Although large language models (LLMs) have succeeded in various natural language processing tasks and have shown potential in automating workflows through reasoning and contextual understanding, current benchmarks for evaluating LLMs in finance lack sufficient domain-specific data, have simplistic task design, and incomplete evaluation frameworks. To address these gaps, this article presents FinMaster, a comprehensive financial benchmark designed to systematically assess the capabilities of LLM in financial literacy, accounting, auditing, and consulting. Specifically, FinMaster comprises three main modules: i) FinSim, which builds simulators that generate synthetic, privacy-compliant financial data for companies to replicate market dynamics; ii) FinSuite, which provides tasks in core financial domains, spanning 183 tasks of various types and difficulty levels; and iii) FinEval, which develops a unified interface for evaluation. Extensive experiments over state-of-the-art LLMs reveal critical capability gaps in financial reasoning, with accuracy dropping from over 90% on basic tasks to merely 40% on complex scenarios requiring multi-step reasoning. This degradation exhibits the propagation of computational errors, where single-metric calculations initially demonstrating 58% accuracy decreased to 37% in multimetric scenarios. To the best of our knowledge, FinMaster is the first benchmark that covers full-pipeline financial workflows with challenging tasks. We hope that FinMaster can bridge the gap between research and industry practitioners, driving the adoption of LLMs in real-world financial practices to enhance efficiency and accuracy.
- Asia > China > Hong Kong (0.04)
- North America > Mexico (0.04)
- Asia > Singapore (0.04)
- (2 more...)
- Financial News (1.00)
- Workflow (0.90)
DeepGreen: Effective LLM-Driven Green-washing Monitoring System Designed for Empirical Testing -- Evidence from China
Xu, Congluo, Miao, Yu, Xiao, Yiling, Lin, Chengmengjia
D EEPG REEN: E FFECTIVE LLM-D RIVEN G REEN-WASHING M ONITORING S YSTEM D ESIGNED FOR E MPIRICAL T ESTING --E VIDENCE FROM C HINA Congluo Xu Business School Sichuan University Chengdu, 610065 Y u Miao School of Economics Sichuan University Chengdu, 610065 Yiling Xiao Business School Sichuan University Chengdu, 610065 Chengmengjia Lin Business School Sichuan University Chengdu, 610065 April 11, 2025 A BSTRACT This paper proposes DeepGreen, an Large Language Model Driven (LLM-Driven) system for detecting corporate green-washing behaviour. Utilizing dual-layer LLM analysis, DeepGreen preliminar-ily identifies potential green keywords in financial statements and then assesses their implementation degree via iterative semantic analysis of LLM. A core variable GreenImplement is derived from the ratio from the two layers' output. We extract 204 financial statements of 68 companies from A-share market over three years, comprising 89,893 words, and analyse them through DeepGreen. Our analysis, supported by violin plots and K-means clustering, reveals insights and validates the variable against the Huazheng ESG rating. It offers a novel perspective for regulatory agencies and investors, serving as a proactive monitoring tool that complements traditional methods.Empirical tests show that green implementation can significantly boost the asset return rate of companies, but there is heterogeneity in scale. Small and medium-sized companies have limited contribution to asset return via green implementation, so there is a stronger motivation for green-washing. K eywords Green-washing Monitoring Large Language Models Financial Statement Analysis Unstructured Data Analysis 1 Introduction Amid intensifying global focus on sustainable development and environmental protection, the phenomenon of corporate "green-washing" has emerged as a contentious issue. "Green-washing" typically refers to those companies exaggerating or misrepresenting their environmental protection efforts in promotional materials, while their actual practices fail to meet sustainable development standards [1]. However, a more elusive challenge lies in "general green-washing", which involves subtler tactics that distort perceptions by repeatedly invoking terms such as "carbon peak" or "green development" without substantive evidence [2]. The elusiveness of general green-washing stems from its exploitation of human psychology and information processing mechanisms.
- Asia > China > Sichuan Province > Chengdu (0.84)
- Europe > Southeast Europe (0.04)
- Europe > Germany (0.04)
- (3 more...)
- Research Report > New Finding (1.00)
- Overview (0.93)
- Law (1.00)
- Energy (0.93)
- Social Sector (0.86)
- Banking & Finance > Trading (0.46)
LaRA: Benchmarking Retrieval-Augmented Generation and Long-Context LLMs - No Silver Bullet for LC or RAG Routing
Li, Kuan, Zhang, Liwen, Jiang, Yong, Xie, Pengjun, Huang, Fei, Wang, Shuai, Cheng, Minhao
Effectively incorporating external knowledge into Large Language Models (LLMs) is crucial for enhancing their capabilities and addressing real-world needs. Retrieval-Augmented Generation (RAG) offers an effective method for achieving this by retrieving the most relevant fragments into LLMs. However, the advancements in context window size for LLMs offer an alternative approach, raising the question of whether RAG remains necessary for effectively handling external knowledge. Several existing studies provide inconclusive comparisons between RAG and long-context (LC) LLMs, largely due to limitations in the benchmark designs. In this paper, we present LaRA, a novel benchmark specifically designed to rigorously compare RAG and LC LLMs. LaRA encompasses 2,326 test cases across four practical QA task categories and three types of naturally occurring long texts. Through systematic evaluation of seven open-source and four proprietary LLMs, we find that the optimal choice between RAG and LC depends on a complex interplay of factors, including the model's parameter size, long-text capabilities, context length, task type, and the characteristics of the retrieved chunks. Our findings provide actionable guidelines for practitioners to effectively leverage both RAG and LC approaches in developing and deploying LLM applications. Our code and dataset is provided at: \href{https://github.com/likuanppd/LaRA}{\textbf{https://github.com/likuanppd/LaRA}}.
- North America > United States > Washington > King County > Seattle (0.14)
- North America > United States > Florida > Miami-Dade County > Miami (0.14)
- Europe > Austria > Vienna (0.14)
- (11 more...)
Explainable Artificial Intelligence for identifying profitability predictors in Financial Statements
Piazza, Marco, Passacantando, Mauro, Magli, Francesca, Doni, Federica, Amaduzzi, Andrea, Messina, Enza
The interconnected nature of the economic variables influencing a firm's performance makes the prediction of a company's earning trend a challenging task. Existing methodologies often rely on simplistic models and financial ratios failing to capture the complexity of interacting influences. In this paper, we apply Machine Learning techniques to raw financial statements data taken from AIDA, a Database comprising Italian listed companies' data from 2013 to 2022. We present a comparative study of different models and following the European AI regulations, we complement our analysis by applying explainability techniques to the proposed models. In particular, we propose adopting an eXplainable Artificial Intelligence method based on Game Theory to identify the most sensitive features and make the result more interpretable.
- North America > United States (0.14)
- Europe > Italy (0.05)
TopoLedgerBERT: Topological Learning of Ledger Description Embeddings using Siamese BERT-Networks
Noels, Sander, Viaene, Sébastien, De Bie, Tijl
This paper addresses a long-standing problem in the field of accounting: mapping company-specific ledger accounts to a standardized chart of accounts. We propose a novel solution, TopoLedgerBERT, a unique sentence embedding method devised specifically for ledger account mapping. This model integrates hierarchical information from the charts of accounts into the sentence embedding process, aiming to accurately capture both the semantic similarity and the hierarchical structure of the ledger accounts. In addition, we introduce a data augmentation strategy that enriches the training data and, as a result, increases the performance of our proposed model. Compared to benchmark methods, TopoLedgerBERT demonstrates superior performance in terms of accuracy and mean reciprocal rank.
- North America > United States (0.14)
- Europe > Belgium > Flanders > East Flanders > Ghent (0.04)
- Europe > United Kingdom (0.04)
- Europe > Switzerland (0.04)