Update README.md
Browse files
README.md
CHANGED
@@ -45,13 +45,13 @@ Thus it is useful for Natural Language Inference and related tasks such as Zero-
|
|
45 |
- **Model type:** Transformer Encoder
|
46 |
- **Language(s) (NLP):** Fine-tuned for Myanmar (Burmese) and English
|
47 |
- **License:** MIT
|
48 |
-
- **Finetuned from model:** mDeBERTa v3 base
|
49 |
-
- **Paper :** For the foundation model mDeBERTa v3, please refer to the paper
|
50 |
- **Demo :** A demo of Zero-shot Text Classification in Myanmar can be found on this page.
|
51 |
|
52 |
## Bias, Risks, and Limitations
|
53 |
|
54 |
-
Please refer to the papers for original foundation model: DeBERTa
|
55 |
<!-- Any limitations with myXNLI ? -->
|
56 |
|
57 |
## How to Get Started with the Model
|
@@ -78,12 +78,12 @@ Fore more details on zero-shot classification, please refer to HuggingFace docum
|
|
78 |
|
79 |
## Training Details
|
80 |
|
81 |
-
The model is fine-tuned on myXNLI dataset
|
82 |
|
83 |
From this dataset, 4 different copies training data from myXNLI were concatenated, each with sentence pairs in en-en, en-my, my-en and my-my combinations.
|
84 |
|
85 |
Training on cross-matched language data as above improved the NLI accuracy over training separately in each language.
|
86 |
-
This approach was inspired by another model
|
87 |
|
88 |
The model was fine-tuned using this combined dataset for a single epoch.
|
89 |
|
|
|
45 |
- **Model type:** Transformer Encoder
|
46 |
- **Language(s) (NLP):** Fine-tuned for Myanmar (Burmese) and English
|
47 |
- **License:** MIT
|
48 |
+
- **Finetuned from model:** mDeBERTa v3 base https://huggingface.co/microsoft/mdeberta-v3-base
|
49 |
+
- **Paper :** For the foundation model mDeBERTa v3, please refer to the paper https://arxiv.org/abs/2111.09543
|
50 |
- **Demo :** A demo of Zero-shot Text Classification in Myanmar can be found on this page.
|
51 |
|
52 |
## Bias, Risks, and Limitations
|
53 |
|
54 |
+
Please refer to the papers for original foundation model: DeBERTa https://arxiv.org/abs/2006.03654 and DeBERTaV3 https://arxiv.org/abs/2111.09543.
|
55 |
<!-- Any limitations with myXNLI ? -->
|
56 |
|
57 |
## How to Get Started with the Model
|
|
|
78 |
|
79 |
## Training Details
|
80 |
|
81 |
+
The model is fine-tuned on myXNLI dataset https://huggingface.co/datasets/akhtet/myXNLI. The English portion of myXNLI is from XNLI dataset.
|
82 |
|
83 |
From this dataset, 4 different copies training data from myXNLI were concatenated, each with sentence pairs in en-en, en-my, my-en and my-my combinations.
|
84 |
|
85 |
Training on cross-matched language data as above improved the NLI accuracy over training separately in each language.
|
86 |
+
This approach was inspired by another model https://huggingface.co/joeddav/xlm-roberta-large-xnli
|
87 |
|
88 |
The model was fine-tuned using this combined dataset for a single epoch.
|
89 |
|