akhtet commited on
Commit
b48e950
·
1 Parent(s): 7b2613f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -45,13 +45,13 @@ Thus it is useful for Natural Language Inference and related tasks such as Zero-
45
  - **Model type:** Transformer Encoder
46
  - **Language(s) (NLP):** Fine-tuned for Myanmar (Burmese) and English
47
  - **License:** MIT
48
- - **Finetuned from model:** mDeBERTa v3 base [https://huggingface.co/microsoft/mdeberta-v3-base]
49
- - **Paper :** For the foundation model mDeBERTa v3, please refer to the paper [https://arxiv.org/abs/2111.09543]
50
  - **Demo :** A demo of Zero-shot Text Classification in Myanmar can be found on this page.
51
 
52
  ## Bias, Risks, and Limitations
53
 
54
- Please refer to the papers for original foundation model: DeBERTa [https://arxiv.org/abs/2006.03654] and DeBERTaV3 [https://arxiv.org/abs/2111.09543].
55
  <!-- Any limitations with myXNLI ? -->
56
 
57
  ## How to Get Started with the Model
@@ -78,12 +78,12 @@ Fore more details on zero-shot classification, please refer to HuggingFace docum
78
 
79
  ## Training Details
80
 
81
- The model is fine-tuned on myXNLI dataset [https://huggingface.co/datasets/akhtet/myXNLI]. The English portion of myXNLI is from XNLI dataset.
82
 
83
  From this dataset, 4 different copies training data from myXNLI were concatenated, each with sentence pairs in en-en, en-my, my-en and my-my combinations.
84
 
85
  Training on cross-matched language data as above improved the NLI accuracy over training separately in each language.
86
- This approach was inspired by another model [https://huggingface.co/joeddav/xlm-roberta-large-xnli]
87
 
88
  The model was fine-tuned using this combined dataset for a single epoch.
89
 
 
45
  - **Model type:** Transformer Encoder
46
  - **Language(s) (NLP):** Fine-tuned for Myanmar (Burmese) and English
47
  - **License:** MIT
48
+ - **Finetuned from model:** mDeBERTa v3 base https://huggingface.co/microsoft/mdeberta-v3-base
49
+ - **Paper :** For the foundation model mDeBERTa v3, please refer to the paper https://arxiv.org/abs/2111.09543
50
  - **Demo :** A demo of Zero-shot Text Classification in Myanmar can be found on this page.
51
 
52
  ## Bias, Risks, and Limitations
53
 
54
+ Please refer to the papers for original foundation model: DeBERTa https://arxiv.org/abs/2006.03654 and DeBERTaV3 https://arxiv.org/abs/2111.09543.
55
  <!-- Any limitations with myXNLI ? -->
56
 
57
  ## How to Get Started with the Model
 
78
 
79
  ## Training Details
80
 
81
+ The model is fine-tuned on myXNLI dataset https://huggingface.co/datasets/akhtet/myXNLI. The English portion of myXNLI is from XNLI dataset.
82
 
83
  From this dataset, 4 different copies training data from myXNLI were concatenated, each with sentence pairs in en-en, en-my, my-en and my-my combinations.
84
 
85
  Training on cross-matched language data as above improved the NLI accuracy over training separately in each language.
86
+ This approach was inspired by another model https://huggingface.co/joeddav/xlm-roberta-large-xnli
87
 
88
  The model was fine-tuned using this combined dataset for a single epoch.
89