finllama
Open-FinLLMs: Open Multimodal Large Language Models for Financial Applications
Xie, Qianqian, Li, Dong, Xiao, Mengxi, Jiang, Zihao, Xiang, Ruoyu, Zhang, Xiao, Chen, Zhengyu, He, Yueru, Han, Weiguang, Yang, Yuzhe, Chen, Shunian, Zhang, Yifei, Shen, Lihang, Kim, Daniel, Liu, Zhiwei, Luo, Zheheng, Yu, Yangyang, Cao, Yupeng, Deng, Zhiyang, Yao, Zhiyuan, Li, Haohang, Feng, Duanyu, Dai, Yongfu, Somasundaram, VijayaSai, Lu, Peng, Zhao, Yilun, Long, Yitao, Xiong, Guojun, Smith, Kaleb, Yu, Honghai, Lai, Yanzhao, Peng, Min, Nie, Jianyun, Suchow, Jordan W., Liu, Xiao-Yang, Wang, Benyou, Lopez-Lira, Alejandro, Huang, Jimin, Ananiadou, Sophia
Large language models (LLMs) have advanced financial applications, yet they often lack sufficient financial knowledge and struggle with tasks involving multi-modal inputs like tables and time series data. To address these limitations, we introduce \textit{Open-FinLLMs}, a series of Financial LLMs. We begin with FinLLaMA, pre-trained on a 52 billion token financial corpus, incorporating text, tables, and time-series data to embed comprehensive financial knowledge. FinLLaMA is then instruction fine-tuned with 573K financial instructions, resulting in FinLLaMA-instruct, which enhances task performance. Finally, we present FinLLaVA, a multimodal LLM trained with 1.43M image-text instructions to handle complex financial data types. Extensive evaluations demonstrate FinLLaMA's superior performance over LLaMA3-8B, LLaMA3.1-8B, and BloombergGPT in both zero-shot and few-shot settings across 19 and 4 datasets, respectively. FinLLaMA-instruct outperforms GPT-4 and other Financial LLMs on 15 datasets. FinLLaVA excels in understanding tables and charts across 4 multimodal tasks. Additionally, FinLLaMA achieves impressive Sharpe Ratios in trading simulations, highlighting its robust financial application capabilities. We will continually maintain and improve our models and benchmarks to support ongoing innovation in academia and industry.
- Financial News (1.00)
- Research Report (0.64)
- Transportation > Ground > Road (1.00)
- Banking & Finance > Trading (1.00)
- Banking & Finance > Insurance (1.00)
- (4 more...)
FinLlama: Financial Sentiment Classification for Algorithmic Trading Applications
Konstantinidis, Thanos, Iacovides, Giorgos, Xu, Mingxue, Constantinides, Tony G., Mandic, Danilo
There are multiple sources of financial news online which influence market movements and trader's decisions. This highlights the need for accurate sentiment analysis, in addition to having appropriate algorithmic trading techniques, to arrive at better informed trading decisions. Standard lexicon based sentiment approaches have demonstrated their power in aiding financial decisions. However, they are known to suffer from issues related to context sensitivity and word ordering. Large Language Models (LLMs) can also be used in this context, but they are not finance-specific and tend to require significant computational resources. To facilitate a finance specific LLM framework, we introduce a novel approach based on the Llama 2 7B foundational model, in order to benefit from its generative nature and comprehensive language manipulation. This is achieved by fine-tuning the Llama2 7B model on a small portion of supervised financial sentiment analysis data, so as to jointly handle the complexities of financial lexicon and context, and further equipping it with a neural network based decision mechanism. Such a generator-classifier scheme, referred to as FinLlama, is trained not only to classify the sentiment valence but also quantify its strength, thus offering traders a nuanced insight into financial news articles. Complementing this, the implementation of parameter-efficient fine-tuning through LoRA optimises trainable parameters, thus minimising computational and memory requirements, without sacrificing accuracy. Simulation results demonstrate the ability of the proposed FinLlama to provide a framework for enhanced portfolio management decisions and increased market returns. These results underpin the ability of FinLlama to construct high-return portfolios which exhibit enhanced resilience, even during volatile periods and unpredictable market events.
- North America > United States > New York (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Research Report > New Finding (0.34)
- Research Report > Promising Solution (0.34)
- Overview > Innovation (0.34)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Information Extraction (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Discourse & Dialogue (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)