FinMTEB: Finance Massive Text Embedding Benchmark
Abstract
Embedding models play a crucial role in representing and retrieving information across various NLP applications. Recent advances in large language models (LLMs) have further enhanced the performance of embedding models. While these models are often benchmarked on general-purpose datasets, real-world applications demand domain-specific evaluation. In this work, we introduce the Finance Massive Text Embedding Benchmark (FinMTEB), a specialized counterpart to MTEB designed for the financial domain. FinMTEB comprises 64 financial domain-specific embedding datasets across 7 tasks that cover diverse textual types in both Chinese and English, such as financial news articles, corporate annual reports, ESG reports, regulatory filings, and earnings call transcripts. We also develop a finance-adapted model, FinPersona-E5, using a persona-based data synthetic method to cover diverse financial embedding tasks for training. Through extensive evaluation of 15 embedding models, including FinPersona-E5, we show three key findings: (1) performance on general-purpose benchmarks shows limited correlation with financial domain tasks; (2) domain-adapted models consistently outperform their general-purpose counterparts; and (3) surprisingly, a simple Bag-of-Words (BoW) approach outperforms sophisticated dense embeddings in financial Semantic Textual Similarity (STS) tasks, underscoring current limitations in dense embedding techniques. Our work establishes a robust evaluation framework for financial NLP applications and provides crucial insights for developing domain-specific embedding models.
Community
We are excited to share our recent work on comprehensive finance text embedding. We also develop a SoTA LLM-based embedding model for the finance domain. 🤗
Title: FinMTEB: Finance Massive Text Embedding Benchmark
Link: https://arxiv.org/abs/2502.10990
Github: https://github.com/yixuantt/FinMTEB
Leaderboard: https://huggingface.co/spaces/FinanceMTEB/FinMTEB
Model: yixuantt/Fin-e5
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LUSIFER: Language Universal Space Integration for Enhanced Multilingual Embeddings with Large Language Models (2025)
- Training Sparse Mixture Of Experts Text Embedding Models (2025)
- Following the Autoregressive Nature of LLM Embeddings via Compression and Alignment (2025)
- Injecting Domain-Specific Knowledge into Large Language Models: A Comprehensive Survey (2025)
- Optimizing Sentence Embedding with Pseudo-Labeling and Model Ensembles: A Hierarchical Framework for Enhanced NLP Tasks (2025)
- DeepThink: Aligning Language Models with Domain-Specific User Intents (2025)
- Semantic Captioning: Benchmark Dataset and Graph-Aware Few-Shot In-Context Learning for SQL2Text (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 1
Collections including this paper 0
No Collection including this paper