FLAG-Trader: Fusion LLM-Agent with Gradient-based Reinforcement Learning for Financial Trading
Xiong, Guojun, Deng, Zhiyang, Wang, Keyi, Cao, Yupeng, Li, Haohang, Yu, Yangyang, Peng, Xueqing, Lin, Mingquan, Smith, Kaleb E, Liu, Xiao-Yang, Huang, Jimin, Ananiadou, Sophia, Xie, Qianqian
–arXiv.org Artificial Intelligence
Large language models (LLMs) fine-tuned on multimodal financial data have demonstrated impressive reasoning capabilities in various financial tasks. However, they often struggle with multi-step, goal-oriented scenarios in interactive financial markets, such as trading, where complex agentic approaches are required to improve decision-making. To address this, we propose \textsc{FLAG-Trader}, a unified architecture integrating linguistic processing (via LLMs) with gradient-driven reinforcement learning (RL) policy optimization, in which a partially fine-tuned LLM acts as the policy network, leveraging pre-trained knowledge while adapting to the financial domain through parameter-efficient fine-tuning. Through policy gradient optimization driven by trading rewards, our framework not only enhances LLM performance in trading but also improves results on other financial-domain tasks. We present extensive empirical evidence to validate these enhancements.
arXiv.org Artificial Intelligence
Feb-17-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > United States
- Minnesota (0.04)
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Industry:
- Banking & Finance > Trading (1.00)
- Technology: