Large Language Models (LLMs) have revolutionized natural language processing, demonstrating exceptional capabilities across diverse applications—from financial analysis to risk assessment and automated reporting.
As financial institutions recognize the value of LLMs, there is a growing trend toward customizing models for financial services, ensuring they capture sector-specific expertise, regulations, and terminology.
This repository serves as a centralized collection of finance-focused LLMs, documenting how companies and research groups develop specialized models tailored for banking, asset management, trading, risk analysis, and regulatory compliance.
By bridging the gap between LLMs and financial applications, this collection showcases real-world implementations, helping track advancements and trends in AI-driven financial services.
If you know of an finance industry-specific LLM that should be added to this repository, feel free to submit a pull request or open an issue! More info in link below:
- Banking & Payments
- Encompasses retail and commercial banking, central banking, digital payments, lending, and credit services.
- Investments & Capital Markets
- Includes asset management, wealth management, stock and bond markets, trading, brokerage, private equity, and fintech innovations.
- Insurance & Risk Management
- Covers life, health, property, casualty insurance, and broader risk management solutions.
Name | Type | Date | Description | Website | Paper |
---|---|---|---|---|---|
CommBiz Gen AI | Undisclosed | Jan 2025 | Together with AWS, the Commonwealth Bank of Australia (CBA) rolled out an AI tool to assist tens of thousands of business customers with inquiries, facilitating quicker payments and efficient transactions. | 🔗 | - |
North for Banking | Pre-trained | Jan 2025 | RBC and Cohere co-developed and securely deployed an enterprise generative AI (genAI) solution optimized for financial services, building upon Cohere's proprietary foundation models | 🔗 | - |
BBVA & OpenAI | Pre-trained | Nov 2024 | BBVA signed an agreement with OpenAI for 3,000 ChatGPT Enterprises licenses, leading to increased productivity and creativity. Staff across various departments have developed over 2,900 specialized GPTs for tasks such as translating risk-specific terminology and drafting responses to client inquiries. | 🔗 | - |
IDEA-FinQA | Agentic RAG | Jun 2024 | Financial question-answering system based on Qwen1.5-14B-Chat, utilizing real-time knowledge injection and supporting various data collection and querying methodologies, and comprises three main modules: the data collector, the data querying module, and LLM-based agents tasked with specific functions. | 🔗 | 🔗 |
Ask FT | Undisclosed | Mar 2024 | LLM tool by Financial Times (FT) that enables subscribers to query and receive responses derived from two decades of published FT content. | 🔗 | - |
RAVEN | Fine-tuned | Jan 2024 | Fine-tuned LLaMA-2 13B Chat model designed to enhance financial data analysis by integrating external tools. Used supervised fine-tuning with parameter-efficient techniques, utilizing a diverse set of financial question-answering datasets, including TAT-QA, Financial PhraseBank, WikiSQL, and OTT-QA | - | 🔗 |
FinMA | Fine-tuned | Jun 2023 | Comprehensive framework that introduces FinMA (Financial Multi-task Assistant), an open-source financial LLM fine-tuned (7B and 30B versions) from LLaMA using a diverse, multi-task instruction dataset of 136,000 samples. The dataset encompasses various financial tasks, document types, and data modalities. | 🔗 | 🔗 |
XuanYuan 2.0 | Pre-trained & Fine-tuned | May 2023 | Chat model (built upon the BLOOM-176B architecture) trained by combining general-domain with domain-specific knowledge and integrating the stages of pre-training and fine-tuning, It is capable of providing accurate and contextually appropriate responses in the Chinese financial domain. | - | 🔗 |
BBT-FinT5 | Pre-trained | Feb 2023 | Chinese financial pre-training language model (1B parameters) based on the T5 model, and pre-trained on the 300Gb financial corpus called FinCorpus | - | 🔗 |
Name | Type | Date | Description | Website | Paper |
---|---|---|---|---|---|
TigerGPT | Pre-trained | Feb 2025 | Tiger Brokers integrated DeepSeek's AI model, DeepSeek-R1, into its AI-powered chatbot, TigerGPT. This adoption aims to enhance market analysis and trading capabilities for its customers through the improved logical reasoning capabilities. | 🔗 | - |
FinTral | Pre-trained | Aug 2024 | Suite of multimodal LLMs built upon the Mistral-7b model and tailored for financial analysis. FinTral integrates textual, numerical, tabular, and image data, and is pretrained on a 20 billion token, high quality dataset | - | 🔗 |
JPMorgan Chase IndexGPT | Undisclosed | Jul 2024 | JPMorgan Chase launched a generative AI-based tool (via AWS Bedrock) called IndexGPT, designed to serve as a 'research analyst' for over 50,000 employees, aiding in various tasks that enhance productivity and decision-making within the firm. It is able to generate and refine written documents, provide creative solutions and summarize extensive documents. | 🔗 | - |
InvestLM | Fine-tuned | Sep 2023 | Financial domain LLM tuned on LLaMA-65B, using a carefully curated instruction dataset related to financial investment. The small yet diverse instruction dataset covers a wide range of financial related topics, from Chartered Financial Analyst (CFA) exam questions to SEC filings to Stackexchange quantitative finance discussions. | 🔗 | 🔗 |
CFGPT | Pre-trained & Fine-tuned | Sep 2023 | Financial LLM based on InternLM-7B that is designed to handle financial texts effectively. It was pre-trained on 584 million documents (141 billion tokens) from Chinese financial sources like announcements, research reports, social media content, and financial news, and then fine-tuned on 1.5 million instruction pairs (1.5 billion tokens) tailored for specific tasks of financial analysis and decision-making. | 🔗 | 🔗 |
FinGPT | Fine-tuned | Jun 2023 | Open-source financial LLM (FinLLM) using a data-centric approach (based on Llama 2) for automated data curation and efficient adaptation, aiming to democratize AI in finance with applications in robo-advising, algorithmic trading, and low-code development. | 🔗 | 🔗 |
Fin-Llama | Fine-tuned | June 2023 | Specialized version of LLaMA 33B, fine-tuned (with QLoRA and 4-bit quantization) for financial applications using a 16.9k instruction dataset. | 🔗 | - |
Cornucopia-LLaMA-Fin-Chinese | Pre-trained | Apr 2023 | Open-source LLaMA-based model fine-tuned for Chinese financial applications. It uses instruction tuning with Chinese financial Q&A datasets to enhance domain-specific performance. | 🔗 | - |
BloombergGPT | Pre-trained | Mar 2023 | 50-billion-parameter LLM specifically designed for financial applications and the industry's unique terminology, trained on a 363-billion-token dataset sourced from Bloomberg’s proprietary data, complemented with 345 billion tokens from general-purpose datasets | 🔗 | 🔗 |
Morgan Stanley & OpenAI | Pre-trained | Mar 2023 | Morgan Stanley Wealth Management announced a partnership with OpenAI to develop an internal-facing GPT-4-powered assistant, allowing their financial advisors to query the bank’s vast research repository and internal knowledge base in natural language | 🔗 | - |
FLANG-ELECTRA | Pre-trained | Oct 2022 | Domain specific Financial LANGuage model (FLANG) which uses financial keywords and phrases for better masking, and built on the ELECTRA-base architecture. Note: Considered a smaller LM as it has fewer than 1B params | 🔗 | 🔗 |
FinBERT-21 | Pre-trained | Jul 2020 | FinBERT (BERT for Financial Text Mining) is a domain specific language model pre-trained on large-scale financial corpora, allowing it to capture language knowledge and semantic information from the finance domain. Note: Considered a smaller LM as it has fewer than 1B params | - | 🔗 |
Name | Type | Date | Description | Website | Paper |
---|---|---|---|---|---|
EXL Insurance LLM | Fine-tuned | Sep 2024 | Industry-specific LLM that supports critical claims and underwriting-related tasks, such as claims reconciliation, data extraction and interpretation, question-answering, anomaly detection and chronology summarization. EXL utilized NVIDIA NeMo end-to-end platform for the fine-tuning process on 2 billion tokens of private insurance data. | 🔗 | - |