This study develops a Small Language Model (SLM) to process financial data at LEARN, enhancing decision-making with efficiency and accuracy. Trained on five years of financial statements,audit reports and spreadsheets, the model leverages quantization, pruning, knowledge distillation, and retrieval-augmented generation (RAG). QLoRA optimizes performance, while hallucination reduction ensures reliability. The system, hosted on LEARN’s servers for security, processes structured and unstructured financial data to provide accurate insights. This research highlights SLMs as a secure, regulatory-compliant alternative to LLMs for financial analysis.
Large Language Models (LLMs) have revolutionized Natural Language Processing (NLP) with their ability to generate human-like text, enabling applications in code writing, math problem-solving, dialogue, and reasoning. In finance, models like BloombergGPT support automated decision-making, risk assessment, fraud prevention, and forecasting. However, their high computational cost and risk of generating incorrect outputs pose challenges.
Small Language Models (SLMs) offer a resource-efficient alternative with domain-specific accuracy. Techniques like quantization, pruning, and knowledge distillation allow SLMs to match LLM performance with lower computational demands. Fine-tuned on financial documents using QLoRA and enhanced with RAG for hallucination control, SLMs ensure high precision—critical for financial applications. As AI adoption in finance grows, regulatory frameworks and ethical standards continue to evolve.
Feature | SLM | LLM |
---|---|---|
Size & Complexity | Smaller in size, with fewer parameters | Massive in size, billions of parameters |
Computational Requirements | Low computational power; optimized for efficiency | Requires high-end GPUs/TPUs and large-scale infrastructure |
Training Cost | Lower cost due to smaller datasets and fewer resources | Extremely expensive due to vast datasets and high training complexity |
Performance in General Tasks | Specialized performance, optimized for specific domains | Strong general-purpose capabilities, excelling in diverse NLP tasks |
Accuracy & Precision | High accuracy within a specific domain (e.g., finance) | High accuracy in general NLP but prone to hallucinations in specialized fields |
Data Efficiency | Requires fewer data for training and fine-tuning | Needs massive datasets for training and generalization |
The first part of the review covers preliminary knowledge in several approaches and techniques in previous research:
Recent advancements in language models (LMs) have significantly improved financial tasks such as sentiment analysis, financial forecasting, and risk management. Specialized models like FINBERT (2022), BloombergGPT, and FinGPT cater specifically to financial NLP challenges, handling complex terminology and numerical data.
While larger models like BloombergGPT offer superior accuracy, smaller models (e.g., Google-Gemma-2B, OpenELM-270M) are more resource-efficient but struggle with complex tasks.
Finance-Specific Language Models | Model Parameters | Model Capabilities |
---|---|---|
BloombergGPT 50B Dataset: FinPile |
50B |
|
FinBERT (open) Dataset: Financial PhraseBank |
110M |
|
FLANG (open) | 110M |
|
InvestLM (fine-tuned LLaMA-open) Dataset: CFA, SEC |
65B |
|
FinMA (fine-tuned LLaMA-open) Dataset: PIxIU |
7B and 13B |
|
FinGPT (open) Dataset: FinQA, FinRed |
7B and 13B |
|
Google-gemma 2B | 2B |
|
TinyLlama (fine-tuned LLaMA) | 1.1B |
|
Apple-OpenELM Dataset: RefinedWeb, Pile |
270M - 3B |
|
Microsoft-phi | 1B - 3B |
|