e19-4yp-Developing-a-Small-Scale-Financial-Language-Model

Developing a Small-Scale Financial Language Model

Team

Supervisors

Table of content

  1. Abstract
  2. Related works
  3. Methodology
  4. Experiment Setup and Implementation
  5. Results and Analysis
  6. Conclusion
  7. Publications
  8. Links

Abstract

This study develops a Small Language Model (SLM) to process financial data at LEARN, enhancing decision-making with efficiency and accuracy. Trained on five years of financial statements,audit reports and spreadsheets, the model leverages quantization, pruning, knowledge distillation, and retrieval-augmented generation (RAG). QLoRA optimizes performance, while hallucination reduction ensures reliability. The system, hosted on LEARN’s servers for security, processes structured and unstructured financial data to provide accurate insights. This research highlights SLMs as a secure, regulatory-compliant alternative to LLMs for financial analysis.

Large Language Models (LLMs) have revolutionized Natural Language Processing (NLP) with their ability to generate human-like text, enabling applications in code writing, math problem-solving, dialogue, and reasoning. In finance, models like BloombergGPT support automated decision-making, risk assessment, fraud prevention, and forecasting. However, their high computational cost and risk of generating incorrect outputs pose challenges.

Small Language Models (SLMs) offer a resource-efficient alternative with domain-specific accuracy. Techniques like quantization, pruning, and knowledge distillation allow SLMs to match LLM performance with lower computational demands. Fine-tuned on financial documents using QLoRA and enhanced with RAG for hallucination control, SLMs ensure high precision—critical for financial applications. As AI adoption in finance grows, regulatory frameworks and ethical standards continue to evolve.

Feature SLM LLM
Size & Complexity Smaller in size, with fewer parameters Massive in size, billions of parameters
Computational Requirements Low computational power; optimized for efficiency Requires high-end GPUs/TPUs and large-scale infrastructure
Training Cost Lower cost due to smaller datasets and fewer resources Extremely expensive due to vast datasets and high training complexity
Performance in General Tasks Specialized performance, optimized for specific domains Strong general-purpose capabilities, excelling in diverse NLP tasks
Accuracy & Precision High accuracy within a specific domain (e.g., finance) High accuracy in general NLP but prone to hallucinations in specialized fields
Data Efficiency Requires fewer data for training and fine-tuning Needs massive datasets for training and generalization

Approaches and Techniques in Previous Research

The first part of the review covers preliminary knowledge in several approaches and techniques in previous research:

Advancements in Financial Language Models: Performance, Applications, and Trade-offs

Recent advancements in language models (LMs) have significantly improved financial tasks such as sentiment analysis, financial forecasting, and risk management. Specialized models like FINBERT (2022), BloombergGPT, and FinGPT cater specifically to financial NLP challenges, handling complex terminology and numerical data.

Key Models & Capabilities:

While larger models like BloombergGPT offer superior accuracy, smaller models (e.g., Google-Gemma-2B, OpenELM-270M) are more resource-efficient but struggle with complex tasks.

Model parameters count and capabilities of language models

Finance-Specific Language Models Model Parameters Model Capabilities
BloombergGPT 50B
Dataset: FinPile
50B
  • Sentiment analysis
  • Named entity recognition
  • Question answering
FinBERT (open)
Dataset: Financial PhraseBank
110M
  • Sentiment analysis
  • Financial entity recognition
  • Financial classification tasks
FLANG (open) 110M
  • Sentiment analysis
  • Named entity recognition
  • Document classification
InvestLM (fine-tuned LLaMA-open)
Dataset: CFA, SEC
65B
  • Sentiment analysis
  • Financial text classification
FinMA (fine-tuned LLaMA-open)
Dataset: PIxIU
7B and 13B
  • Sentiment analysis
  • Financial document summarization
  • Question answering
FinGPT (open)
Dataset: FinQA, FinRed
7B and 13B
  • Financial document summarization
  • Question answering
Google-gemma 2B 2B
  • Financial text classification
  • Financial document summarization
  • Question answering
TinyLlama (fine-tuned LLaMA) 1.1B
  • Financial text classification
  • Financial document summarization
  • Question answering
Apple-OpenELM
Dataset: RefinedWeb, Pile
270M - 3B
  • Sentiment analysis
  • Named entity recognition (NER)
  • Document classification
Microsoft-phi 1B - 3B
  • Sentiment analysis
  • Named entity recognition (NER)
  • Document classification
  • Question answering

Methodology

Desc

Experiment Setup and Implementation

Results and Analysis

Conclusion

Publications