nadiinchi commited on
Commit
dd70779
·
verified ·
1 Parent(s): fa708b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -110,12 +110,12 @@ Interface of the `process` function:
110
 
111
  ## Model Details
112
 
113
- * Input: user question (e.g., a sentence) + retrieved context passage (e.g., a paragraph)
114
  * Output: pruned context passage, i.e., irrelevant sentences are removed + relevance score (can be used for reranking)
115
  * Model Architecture: The model was initialized from [bge-reranker-v2-m3](https://huggingface.co/BAAI/bge-reranker-v2-m3) and finetuned with two objectives: (1) output a binary mask which can be used to prune irrelevant sentences; and (2) preserve initial reranking capabilities.
116
  * Training data: [MS Marco](https://microsoft.github.io/msmarco/Datasets.html) + [MIRACL](https://huggingface.co/datasets/miracl/miracl), with synthetic silver labelling of which sentences to keep, produced using [aya-expanse-8b](https://huggingface.co/CohereLabs/aya-expanse-8b).
117
  * Languages in the training data: Arabic, Bengali, English, Spanish, Persian, Finnish, France, Hindi, Indonesian, Japanese, Korean, Russian, Swahili, Telugu, Thai, Chinese
118
- * Context length: 8192 tokens (similar to the pretrained BGE-m3 model)
119
  * Evaluation: we evaluate XProvence on 26 languages from 6 different datasets. We find that XProvence is able to prune irrelevant sentences with little-to-no drop in performance, on all languages, and outperforms existing baselines on the Pareto front.
120
 
121
 
 
110
 
111
  ## Model Details
112
 
113
+ * Input: user question (e.g., a sentence) + retrieved context passage (e.g., a paragraph). Training data consisted of monolingual examples (query and context in the same language), but we expect the model to perform well on cross-lingual pairs too, due to cross-lingual transfer.
114
  * Output: pruned context passage, i.e., irrelevant sentences are removed + relevance score (can be used for reranking)
115
  * Model Architecture: The model was initialized from [bge-reranker-v2-m3](https://huggingface.co/BAAI/bge-reranker-v2-m3) and finetuned with two objectives: (1) output a binary mask which can be used to prune irrelevant sentences; and (2) preserve initial reranking capabilities.
116
  * Training data: [MS Marco](https://microsoft.github.io/msmarco/Datasets.html) + [MIRACL](https://huggingface.co/datasets/miracl/miracl), with synthetic silver labelling of which sentences to keep, produced using [aya-expanse-8b](https://huggingface.co/CohereLabs/aya-expanse-8b).
117
  * Languages in the training data: Arabic, Bengali, English, Spanish, Persian, Finnish, France, Hindi, Indonesian, Japanese, Korean, Russian, Swahili, Telugu, Thai, Chinese
118
+ * Context length: 8192 tokens (similar to the pretrained BGE-m3 model). However, training data only included paragraph-sized examples.
119
  * Evaluation: we evaluate XProvence on 26 languages from 6 different datasets. We find that XProvence is able to prune irrelevant sentences with little-to-no drop in performance, on all languages, and outperforms existing baselines on the Pareto front.
120
 
121