lbourdois commited on
Commit
2309b86
·
verified ·
1 Parent(s): c659469

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +43 -29
README.md CHANGED
@@ -1,30 +1,44 @@
1
- ---
2
- license: mit
3
- datasets:
4
- - normster/RealGuardrails
5
- base_model:
6
- - Qwen/Qwen2.5-7B
7
- - normster/RealGuardrails-Qwen2.5-7B-SFT
8
- library_name: transformers
9
- ---
10
-
11
- # RealGuardrails Models
12
-
13
- This model was trained on the [RealGuardrails](https://huggingface.co/datasets/normster/RealGuardrails) dataset, an instruction-tuning dataset focused on improving system prompt adherence and precedence. In particular, it was trained via SFT on the `systemmix` split (150K examples) using our custom training library [torchllms](https://github.com/normster/torchllms) (yielding [normster/RealGuardrails-Qwen2.5-7B-SFT](https://huggingface.co/normster/RealGuardrails-Qwen2.5-7B-SFT)), and then trained via DPO on the `preferencemix` split (30K examples), and converted back to a `transformers` compatible checkpoint.
14
-
15
- ## Training Hyperparameters
16
-
17
- | Name | Value |
18
- | :--- | :--- |
19
- | DPO beta | 0.01 |
20
- | optimizer | AdamW |
21
- | batch size | 128 |
22
- | learning rate | 1e-5 |
23
- | lr scheduler | cosine with 50 warmup steps |
24
- | betas | (0.9, 0.999) |
25
- | eps | 1e-8 |
26
- | weight decay | 0 |
27
- | epochs | 1 |
28
- | max grad norm | 1.0 |
29
- | precision | bf16 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  | max length | 4096 |
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - normster/RealGuardrails
5
+ base_model:
6
+ - Qwen/Qwen2.5-7B
7
+ - normster/RealGuardrails-Qwen2.5-7B-SFT
8
+ library_name: transformers
9
+ language:
10
+ - zho
11
+ - eng
12
+ - fra
13
+ - spa
14
+ - por
15
+ - deu
16
+ - ita
17
+ - rus
18
+ - jpn
19
+ - kor
20
+ - vie
21
+ - tha
22
+ - ara
23
+ ---
24
+
25
+ # RealGuardrails Models
26
+
27
+ This model was trained on the [RealGuardrails](https://huggingface.co/datasets/normster/RealGuardrails) dataset, an instruction-tuning dataset focused on improving system prompt adherence and precedence. In particular, it was trained via SFT on the `systemmix` split (150K examples) using our custom training library [torchllms](https://github.com/normster/torchllms) (yielding [normster/RealGuardrails-Qwen2.5-7B-SFT](https://huggingface.co/normster/RealGuardrails-Qwen2.5-7B-SFT)), and then trained via DPO on the `preferencemix` split (30K examples), and converted back to a `transformers` compatible checkpoint.
28
+
29
+ ## Training Hyperparameters
30
+
31
+ | Name | Value |
32
+ | :--- | :--- |
33
+ | DPO beta | 0.01 |
34
+ | optimizer | AdamW |
35
+ | batch size | 128 |
36
+ | learning rate | 1e-5 |
37
+ | lr scheduler | cosine with 50 warmup steps |
38
+ | betas | (0.9, 0.999) |
39
+ | eps | 1e-8 |
40
+ | weight decay | 0 |
41
+ | epochs | 1 |
42
+ | max grad norm | 1.0 |
43
+ | precision | bf16 |
44
  | max length | 4096 |