SPEAR-MM: Selective Parameter Evaluation and Restoration via Model Merging for Efficient Financial LLM Adaptation
Kapusuzoglu, Berkcan, Chakraborty, Supriyo, Ni, Renkun, Rawls, Stephen, Sahu, Sambit
–arXiv.org Artificial Intelligence
Abstract--Large language models (LLMs) adapted to financial domains often suffer from catastrophic forgetting of general reasoning capabilities essential for customer interactions and complex financial analysis. Our method approximates layer-wise impact on external benchmarks through post-hoc analysis, then selectively freezes or restores transformer layers via spherical interpolation merging. Applied to LLaMA-3.1-8B for financial tasks, SPEAR-MM achieves 91.2% retention of general capabilities versus 69.7% for standard continual pretraining, while maintaining 94% of domain adaptation gains. The approach provides interpretable trade-off control and reduces computational costs by 90% crucial for resource-constrained financial institutions. Financial institutions increasingly require domain-specific language models that can understand regulatory documents, analyze market data, and provide accurate customer support while maintaining broad reasoning capabilities for complex financial scenarios.
arXiv.org Artificial Intelligence
Nov-12-2025
- Country:
- North America > United States (0.14)
- Genre:
- Research Report (0.85)
- Industry:
- Banking & Finance (0.89)
- Technology: