lapp0 commited on
Commit
4a71b01
·
verified ·
1 Parent(s): e7ebca3

End of training

Browse files
Files changed (1) hide show
  1. README.md +29 -16
README.md CHANGED
@@ -1,6 +1,7 @@
1
  ---
2
- license: mit
3
  base_model: gpt2
 
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -8,14 +9,23 @@ model-index:
8
  results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
  # gpt2_model_card_distily_test
15
 
16
- This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
 
 
 
17
  It achieves the following results on the evaluation set:
18
- - Loss: 2480.0
 
 
 
 
 
 
 
 
 
19
 
20
  ## Model description
21
 
@@ -28,12 +38,16 @@ More information needed
28
  ## Training and evaluation data
29
 
30
  More information needed
 
31
 
32
  ## Training procedure
33
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
 
 
 
37
  - learning_rate: 0.0001
38
  - train_batch_size: 1
39
  - eval_batch_size: 2
@@ -42,19 +56,18 @@ The following hyperparameters were used during training:
42
  - lr_scheduler_type: cosine
43
  - num_epochs: 1.0
44
 
45
- ### Training results
46
-
47
- | Training Loss | Epoch | Step | Validation Loss |
48
- |:-------------:|:------:|:----:|:---------------:|
49
- | No log | 0 | 0 | 6944.0 |
50
- | 2707.0 | 0.2513 | 50 | 2672.0 |
51
- | 2531.5 | 0.5025 | 100 | 2528.0 |
52
- | 2324.5 | 0.7538 | 150 | 2480.0 |
53
 
 
 
 
 
 
 
54
 
55
  ### Framework versions
56
-
57
  - Transformers 4.43.3
58
  - Pytorch 2.3.0
59
  - Datasets 2.20.0
60
- - Tokenizers 0.19.1
 
1
  ---
 
2
  base_model: gpt2
3
+ library_name: distily
4
+ license: mit
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
9
  results: []
10
  ---
11
 
 
 
 
12
  # gpt2_model_card_distily_test
13
 
14
+ This student model is distilled from the teacher model [gpt2](https://huggingface.co/gpt2) using the dataset (unspecified).
15
+
16
+ The [Distily](https://github.com/lapp0/distily) library was used for this distillation.
17
+
18
  It achieves the following results on the evaluation set:
19
+ - eval_enwikippl: 16455.1230
20
+ - eval_frwikippl: 38444.9648
21
+ - eval_zhwikippl: 56717.4922
22
+ - eval_loss: 0.0004
23
+ - eval_runtime: 0.0554
24
+ - eval_samples_per_second: 18.066
25
+ - eval_steps_per_second: 18.066
26
+
27
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
+ should probably proofread and complete it, then remove this comment.
29
 
30
  ## Model description
31
 
 
38
  ## Training and evaluation data
39
 
40
  More information needed
41
+ -->
42
 
43
  ## Training procedure
44
 
45
  ### Training hyperparameters
46
 
47
  The following hyperparameters were used during training:
48
+ - distillation_strategy: logits_activations
49
+ - loss_fn: reverse_kl
50
+ - train_embeddings: True
51
  - learning_rate: 0.0001
52
  - train_batch_size: 1
53
  - eval_batch_size: 2
 
56
  - lr_scheduler_type: cosine
57
  - num_epochs: 1.0
58
 
59
+ ### Resource Usage
60
+ Peak GPU Memory: 1.2452 GB
 
 
 
 
 
 
61
 
62
+ ### Model Results
63
+ | epoch | step | eval_enwikippl | eval_frwikippl | eval_loss | eval_runtime | eval_samples_per_second | eval_steps_per_second | eval_zhwikippl |
64
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- |
65
+ | 0 | 0 | 63012.375 | 58568.7617 | 0.0042 | 0.076 | 13.155 | 13.155 | 62696.3008 |
66
+ | 0.4040 | 40 | 20128.3281 | 41006.9219 | 0.0004 | 0.0553 | 18.079 | 18.079 | 58574.4609 |
67
+ | 0.8081 | 80 | 16455.1230 | 38444.9648 | 0.0004 | 0.0554 | 18.066 | 18.066 | 56717.4922 |
68
 
69
  ### Framework versions
70
+ - Distily 0.1.0
71
  - Transformers 4.43.3
72
  - Pytorch 2.3.0
73
  - Datasets 2.20.0