Update README.md
#4
by
cyente
- opened
README.md
CHANGED
|
@@ -27,7 +27,7 @@ Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (
|
|
| 27 |
|
| 28 |
**This repo contains the 7B Qwen2.5-Coder model**, which has the following features:
|
| 29 |
- Type: Causal Language Models
|
| 30 |
-
- Training Stage: Pretraining
|
| 31 |
- Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
|
| 32 |
- Number of Parameters: 7.61B
|
| 33 |
- Number of Paramaters (Non-Embedding): 6.53B
|
|
|
|
| 27 |
|
| 28 |
**This repo contains the 7B Qwen2.5-Coder model**, which has the following features:
|
| 29 |
- Type: Causal Language Models
|
| 30 |
+
- Training Stage: Pretraining
|
| 31 |
- Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
|
| 32 |
- Number of Parameters: 7.61B
|
| 33 |
- Number of Paramaters (Non-Embedding): 6.53B
|