bloomberggpt
- Asia > China > Hubei Province > Wuhan (0.05)
- North America > United States > Florida > Brevard County > Cape Canaveral (0.04)
- Oceania > Australia (0.04)
- (5 more...)
- Banking & Finance > Credit (0.69)
- Banking & Finance > Trading (0.68)
- Government (0.68)
- Banking & Finance > Economy (0.47)
- Asia > China > Hubei Province > Wuhan (0.05)
- North America > United States > Florida > Brevard County > Cape Canaveral (0.04)
- Oceania > Australia (0.04)
- (5 more...)
- Banking & Finance > Credit (0.69)
- Banking & Finance > Trading (0.68)
- Government (0.68)
- Banking & Finance > Economy (0.47)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.95)
Large Language Model Adaptation for Financial Sentiment Analysis
Inserte, Pau Rodriguez, Nakhlé, Mariam, Qader, Raheel, Caillaut, Gaetan, Liu, Jingshu
Natural language processing (NLP) has recently gained relevance within financial institutions by providing highly valuable insights into companies and markets' financial documents. However, the landscape of the financial domain presents extra challenges for NLP, due to the complexity of the texts and the use of specific terminology. Generalist language models tend to fall short in tasks specifically tailored for finance, even when using large language models (LLMs) with great natural language understanding and generative capabilities. This paper presents a study on LLM adaptation methods targeted at the financial domain and with high emphasis on financial sentiment analysis. To this purpose, two foundation models with less than 1.5B parameters have been adapted using a wide range of strategies. We show that through careful fine-tuning on both financial documents and instructions, these foundation models can be adapted to the target domain. Moreover, we observe that small LLMs have comparable performance to larger scale models, while being more efficient in terms of parameters and data. In addition to the models, we show how to generate artificial instructions through LLMs to augment the number of samples of the instruction dataset.
- North America > United States > New York (0.05)
- Europe > France > Auvergne-Rhône-Alpes > Isère > Grenoble (0.04)
- Oceania > Australia (0.04)
- (5 more...)
Fine-tuning and Utilization Methods of Domain-specific LLMs
Recent releases of pre-trained Large Language Models (LLMs) have gained considerable traction, yet research on fine-tuning and employing domain-specific LLMs remains scarce. This study investigates approaches for fine-tuning and leveraging domain-specific LLMs, highlighting trends in LLMs, foundational models, and methods for domain-specific pre-training. Focusing on the financial sector, it details dataset selection, preprocessing, model choice, and considerations crucial for LLM fine-tuning in finance. Addressing the unique characteristics of financial data, the study explores the construction of domain-specific vocabularies and considerations for security and regulatory compliance. In the practical application of LLM fine-tuning, the study outlines the procedure and implementation for generating domain-specific LLMs in finance. Various financial cases, including stock price prediction, sentiment analysis of financial news, automated document processing, research, information extraction, and customer service enhancement, are exemplified. The study explores the potential of LLMs in the financial domain, identifies limitations, and proposes directions for improvement, contributing valuable insights for future research. Ultimately, it advances natural language processing technology in business, suggesting proactive LLM utilization in financial services across industries.
- Asia > Middle East > UAE (0.14)
- North America > United States (0.04)
- Asia > South Korea (0.04)
- Asia > China (0.04)
- Information Technology > Security & Privacy (1.00)
- Banking & Finance > Trading (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.68)
Efficient Continual Pre-training for Building Domain Specific Large Language Models
Xie, Yong, Aggarwal, Karan, Ahmad, Aitzaz
Large language models (LLMs) have demonstrated remarkable open-domain capabilities. Traditionally, LLMs tailored for a domain are trained from scratch to excel at handling domain-specific tasks. In this work, we explore an alternative strategy of continual pre-training as a means to develop domain-specific LLMs. We introduce FinPythia-6.9B, developed through domain-adaptive continual pre-training on the financial domain. Continual pre-trained FinPythia showcases consistent improvements on financial tasks over the original foundational model. We further explore simple but effective data selection strategies for continual pre-training. Our data selection strategies outperforms vanilla continual pre-training's performance with just 10% of corpus size and cost, without any degradation on open-domain standard tasks. Our work proposes an alternative solution to building domain-specific LLMs from scratch in a cost-effective manner.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > Washington > King County > Seattle (0.04)
- North America > United States > Illinois (0.04)
- (5 more...)
- Government > Regional Government > North America Government > United States Government (0.68)
- Banking & Finance > Trading (0.68)
- Energy (0.67)
Large Language Models in Finance: A Survey
Li, Yinheng, Wang, Shaofei, Ding, Han, Chen, Hang
Recent advances in large language models (LLMs) have opened new possibilities for artificial intelligence applications in finance. In this paper, we provide a practical survey focused on two key aspects of utilizing LLMs for financial tasks: existing solutions and guidance for adoption. First, we review current approaches employing LLMs in finance, including leveraging pretrained models via zero-shot or few-shot learning, fine-tuning on domain-specific data, and training custom LLMs from scratch. We summarize key models and evaluate their performance improvements on financial natural language processing tasks. Second, we propose a decision framework to guide financial professionals in selecting the appropriate LLM solution based on their use case constraints around data, compute, and performance needs. The framework provides a pathway from lightweight experimentation to heavy investment in customized LLMs. Lastly, we discuss limitations and challenges around leveraging LLMs in financial applications. Overall, this survey aims to synthesize the state-of-the-art and provide a roadmap for responsibly applying LLMs to advance financial AI.
- North America > United States > New York > New York County > New York City (0.15)
- North America > United States > New York > Richmond County > New York City (0.05)
- North America > United States > New York > Queens County > New York City (0.05)
- (4 more...)
- Research Report (1.00)
- Overview (1.00)
- Banking & Finance > Trading (1.00)
- Information Technology > Security & Privacy (0.93)
Meta Has A.I. Google Has A.I. Microsoft Has A.I. Amazon Has a Plan.
This article is from Big Technology, a newsletter by Alex Kantrowitz. Amazon's absence from this year's generative–A.I. bonanza has been a bit puzzling. The company invented Alexa, intuiting people's interest in speaking with computers, yet when OpenAI released ChatGPT it seemed to cede the territory. But rather than sitting out the game, Amazon is waiting to play on its terms. Instead of building one A.I. product, it wants a piece of all of them.
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning > Generative AI (0.60)
FinGPT: Open-Source Financial Large Language Models
Yang, Hongyang, Liu, Xiao-Yang, Wang, Christina Dan
Large language models (LLMs) have shown the potential of revolutionizing natural language processing tasks in diverse domains, sparking great interest in finance. Accessing high-quality financial data is the first challenge for financial LLMs (FinLLMs). While proprietary models like BloombergGPT have taken advantage of their unique data accumulation, such privileged access calls for an open-source alternative to democratize Internet-scale financial data. In this paper, we present an open-source large language model, FinGPT, for the finance sector. Unlike proprietary models, FinGPT takes a data-centric approach, providing researchers and practitioners with accessible and transparent resources to develop their FinLLMs. We highlight the importance of an automatic data curation pipeline and the lightweight low-rank adaptation technique in building FinGPT. Furthermore, we showcase several potential applications as stepping stones for users, such as robo-advising, algorithmic trading, and low-code development. Through collaborative efforts within the open-source AI4Finance community, FinGPT aims to stimulate innovation, democratize FinLLMs, and unlock new opportunities in open finance. Two associated code repos are \url{https://github.com/AI4Finance-Foundation/FinGPT} and \url{https://github.com/AI4Finance-Foundation/FinNLP}
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > New York (0.04)
- Overview (0.46)
- Research Report (0.40)
PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark for Finance
Xie, Qianqian, Han, Weiguang, Zhang, Xiao, Lai, Yanzhao, Peng, Min, Lopez-Lira, Alejandro, Huang, Jimin
Although large language models (LLMs) has shown great performance on natural language processing (NLP) in the financial domain, there are no publicly available financial tailtored LLMs, instruction tuning datasets, and evaluation benchmarks, which is critical for continually pushing forward the open-source development of financial artificial intelligence (AI). This paper introduces PIXIU, a comprehensive framework including the first financial LLM based on fine-tuning LLaMA with instruction data, the first instruction data with 136K data samples to support the fine-tuning, and an evaluation benchmark with 5 tasks and 9 datasets. We first construct the large-scale multi-task instruction data considering a variety of financial tasks, financial document types, and financial data modalities. We then propose a financial LLM called FinMA by fine-tuning LLaMA with the constructed dataset to be able to follow instructions for various financial tasks. To support the evaluation of financial LLMs, we propose a standardized benchmark that covers a set of critical financial tasks, including five financial NLP tasks and one financial prediction task. With this benchmark, we conduct a detailed analysis of FinMA and several existing LLMs, uncovering their strengths and weaknesses in handling critical financial tasks. The model, datasets, benchmark, and experimental results are open-sourced to facilitate future research in financial AI.
- Asia > China > Hubei Province > Wuhan (0.05)
- North America > United States > Florida > Brevard County > Cape Canaveral (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- (5 more...)
The ChatGPT Of Finance Is Here, Bloomberg Is Combining AI And Fintech
A Bloomberg terminal keyboard is seen in central London on April 17, 2015. Bloomberg terminals used ... [ ] by subscribers to make trades using real-time developments in business and finance were struck by a "global network problem" for several hours today, the company said. After users in financial centres around the world flocked to Twitter to complain of the unexpected outage of terminals, Bloomberg technicians began repair operations that started bringing some blanked terminals back on line at around 0945 GMT. Bloomberg is bringing to finance what GPT and ChatGPT brought to everyday general purpose chatbots. The paper that Bloomberg released reveals the great technical depth of its BloombergGPT machine learning model, applying the type of AI techniques that GPT uses to financial datasets.
- Banking & Finance (1.00)
- Information Technology > Services (0.51)