Update README.md
Browse files
README.md
CHANGED
|
@@ -28,9 +28,7 @@ This model is a fine-tuned version of **Microsoft's UniXcoder**, optimized for d
|
|
| 28 |
- **Architecture:** Transformer-based sequence classification
|
| 29 |
|
| 30 |
## Model Sources
|
| 31 |
-
- **Repository:** [Add Hugging Face Model Link Here]
|
| 32 |
- **Paper (UniXcoder):** [https://arxiv.org/abs/2203.03850](https://arxiv.org/abs/2203.03850)
|
| 33 |
-
- **Demo (Optional):** [Add Gradio/Streamlit Link Here]
|
| 34 |
|
| 35 |
## Uses
|
| 36 |
|
|
@@ -131,20 +129,4 @@ The model was evaluated using **20% of the dataset**, with the following results
|
|
| 131 |
| Factor | Value |
|
| 132 |
|---------|--------|
|
| 133 |
| **GPU Used** | 2x T4 GPU |
|
| 134 |
-
| **Training Time** | ~1 hour |
|
| 135 |
-
|
| 136 |
-
## Citation
|
| 137 |
-
If you use this model in your research or applications, please cite:
|
| 138 |
-
|
| 139 |
-
```
|
| 140 |
-
@article{unixcoder,
|
| 141 |
-
title={UniXcoder: Unified Cross-Modal Pretraining for Code Representation},
|
| 142 |
-
author={Guo, Daya and Wang, Shuo and Wan, Yao and others},
|
| 143 |
-
year={2022},
|
| 144 |
-
journal={arXiv preprint arXiv:2203.03850}
|
| 145 |
-
}
|
| 146 |
-
```
|
| 147 |
-
|
| 148 |
-
## Model Card Authors
|
| 149 |
-
- **Mukit Mahdin**
|
| 150 |
-
- Contact: [[email protected]]
|
|
|
|
| 28 |
- **Architecture:** Transformer-based sequence classification
|
| 29 |
|
| 30 |
## Model Sources
|
|
|
|
| 31 |
- **Paper (UniXcoder):** [https://arxiv.org/abs/2203.03850](https://arxiv.org/abs/2203.03850)
|
|
|
|
| 32 |
|
| 33 |
## Uses
|
| 34 |
|
|
|
|
| 129 |
| Factor | Value |
|
| 130 |
|---------|--------|
|
| 131 |
| **GPU Used** | 2x T4 GPU |
|
| 132 |
+
| **Training Time** | ~1 hour |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|