AndrewACN commited on
Commit
ce93ba1
·
verified ·
1 Parent(s): f67636e

End of training

Browse files
README.md CHANGED
@@ -3,8 +3,6 @@ license: apache-2.0
3
  base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
- metrics:
7
- - accuracy
8
  model-index:
9
  - name: distilbert-base-uncased-finetuned-voxi-mnli
10
  results: []
@@ -17,8 +15,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.7452
21
- - Accuracy: 0.8208
 
 
 
 
 
22
 
23
  ## Model description
24
 
@@ -38,8 +41,8 @@ More information needed
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 5e-05
41
- - train_batch_size: 64
42
- - eval_batch_size: 64
43
  - seed: 42
44
  - distributed_type: multi-GPU
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
@@ -47,10 +50,6 @@ The following hyperparameters were used during training:
47
  - num_epochs: 5
48
  - mixed_precision_training: Native AMP
49
 
50
- ### Training results
51
-
52
-
53
-
54
  ### Framework versions
55
 
56
  - Transformers 4.41.2
 
3
  base_model: distilbert-base-uncased
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: distilbert-base-uncased-finetuned-voxi-mnli
8
  results: []
 
15
 
16
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - eval_loss: 0.7060
19
+ - eval_accuracy: 0.7036
20
+ - eval_runtime: 7.6448
21
+ - eval_samples_per_second: 1296.968
22
+ - eval_steps_per_second: 81.101
23
+ - epoch: 0.1041
24
+ - step: 2561
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 5e-05
44
+ - train_batch_size: 16
45
+ - eval_batch_size: 16
46
  - seed: 42
47
  - distributed_type: multi-GPU
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
50
  - num_epochs: 5
51
  - mixed_precision_training: Native AMP
52
 
 
 
 
 
53
  ### Framework versions
54
 
55
  - Transformers 4.41.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55a2fb8be0fe7e328e73d71f4e3d8d30519151fb36c021f7cf06faee2e04133a
3
  size 267835644
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:270dd9197d7ac291f023725c07575e2f73f77bd7b2fdcd410a4dfd83ac8889d0
3
  size 267835644
runs/Oct10_17-57-37_1007-155517-cc8extdd-10-139-64-4/events.out.tfevents.1728583078.1007-155517-cc8extdd-10-139-64-4.1653.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4207adc421e0ab648a70122af15c2910d53fd81d1d145ed58ce8795f615876d6
3
+ size 4955
runs/Oct10_17-58-30_1007-155517-cc8extdd-10-139-64-4/events.out.tfevents.1728583119.1007-155517-cc8extdd-10-139-64-4.1653.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3c6728a865d4c1c493eadb302c8e91c048c1d1a1d5978e50c12102a9eb4e349
3
+ size 5004
runs/Oct10_18-18-44_1007-155517-cc8extdd-10-139-64-4/events.out.tfevents.1728584355.1007-155517-cc8extdd-10-139-64-4.1321.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:347f8b4a18e125a1bffbd263b94aeec81c7c28c9c658187da3221900f01578a6
3
+ size 6333
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8fd8def01c83c31a3e3a648ffefd3f97f16a1e8e89e8af9349e6312a5b338879
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d9cb82748e567b2107cb38d356320ca3d54aaf1a27d5cc75d2fdc7ca0ee446
3
  size 5240