Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ language:
|
|
5 |
---
|
6 |
|
7 |
# **Introduction**
|
8 |
-
We introduce LUXIA-21.4B-
|
9 |
|
10 |
It's demonstrates unparalleled state-of-the-art performance in models with parameters under 35B, and it also outperformed the 72B model and the 34Bx2 MoE (Mixture of Experts) model. Please refer to the evaluation results table for details.
|
11 |
|
|
|
5 |
---
|
6 |
|
7 |
# **Introduction**
|
8 |
+
We introduce LUXIA-21.4B-Alignment, a large language model (LLM) with 21.4 billion parameters, demonstrating superior performance in various natural language processing (NLP) tasks.
|
9 |
|
10 |
It's demonstrates unparalleled state-of-the-art performance in models with parameters under 35B, and it also outperformed the 72B model and the 34Bx2 MoE (Mixture of Experts) model. Please refer to the evaluation results table for details.
|
11 |
|