Training in progress, step 500
Browse files- all_results.json +13 -0
- config.json +2 -17
- eval_results.json +8 -0
- preprocessor_config.json +7 -9
- pytorch_model.bin +1 -1
- runs/Apr25_03-19-25_7626d6fa6bff/events.out.tfevents.1682394094.7626d6fa6bff.149.5 +3 -0
- runs/Apr25_03-47-00_7626d6fa6bff/1682394456.302469/events.out.tfevents.1682394456.7626d6fa6bff.149.7 +3 -0
- runs/Apr25_03-47-00_7626d6fa6bff/events.out.tfevents.1682394456.7626d6fa6bff.149.6 +3 -0
- train_results.json +8 -0
- training_args.bin +1 -1
all_results.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"eval_accuracy": 0.0014388489208633094,
|
4 |
+
"eval_loss": 9.965200424194336,
|
5 |
+
"eval_runtime": 19.8708,
|
6 |
+
"eval_samples_per_second": 69.952,
|
7 |
+
"eval_steps_per_second": 8.757,
|
8 |
+
"total_flos": 1.8890449427391898e+18,
|
9 |
+
"train_loss": 3.100843106976088,
|
10 |
+
"train_runtime": 1215.6267,
|
11 |
+
"train_samples_per_second": 20.027,
|
12 |
+
"train_steps_per_second": 2.505
|
13 |
+
}
|
config.json
CHANGED
@@ -1,27 +1,13 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"ViTForImageClassification"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.0,
|
7 |
-
"depths": [
|
8 |
-
3,
|
9 |
-
4,
|
10 |
-
6,
|
11 |
-
3
|
12 |
-
],
|
13 |
-
"downsample_in_first_stage": false,
|
14 |
-
"embedding_size": 64,
|
15 |
"encoder_stride": 16,
|
16 |
-
"hidden_act": "
|
17 |
"hidden_dropout_prob": 0.0,
|
18 |
"hidden_size": 768,
|
19 |
-
"hidden_sizes": [
|
20 |
-
256,
|
21 |
-
512,
|
22 |
-
1024,
|
23 |
-
2048
|
24 |
-
],
|
25 |
"id2label": {
|
26 |
"0": "Porygon",
|
27 |
"1": "Goldeen",
|
@@ -330,7 +316,6 @@
|
|
330 |
"Zubat": "58"
|
331 |
},
|
332 |
"layer_norm_eps": 1e-12,
|
333 |
-
"layer_type": "bottleneck",
|
334 |
"model_type": "vit",
|
335 |
"num_attention_heads": 12,
|
336 |
"num_channels": 3,
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "google/vit-base-patch16-224-in21k",
|
3 |
"architectures": [
|
4 |
"ViTForImageClassification"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.0,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
"encoder_stride": 16,
|
8 |
+
"hidden_act": "gelu",
|
9 |
"hidden_dropout_prob": 0.0,
|
10 |
"hidden_size": 768,
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
"id2label": {
|
12 |
"0": "Porygon",
|
13 |
"1": "Goldeen",
|
|
|
316 |
"Zubat": "58"
|
317 |
},
|
318 |
"layer_norm_eps": 1e-12,
|
|
|
319 |
"model_type": "vit",
|
320 |
"num_attention_heads": 12,
|
321 |
"num_channels": 3,
|
eval_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"eval_accuracy": 0.0014388489208633094,
|
4 |
+
"eval_loss": 9.965200424194336,
|
5 |
+
"eval_runtime": 19.8708,
|
6 |
+
"eval_samples_per_second": 69.952,
|
7 |
+
"eval_steps_per_second": 8.757
|
8 |
+
}
|
preprocessor_config.json
CHANGED
@@ -1,21 +1,19 @@
|
|
1 |
{
|
2 |
-
"crop_pct": 0.875,
|
3 |
"do_normalize": true,
|
4 |
"do_rescale": true,
|
5 |
"do_resize": true,
|
6 |
-
"feature_extractor_type": "ConvNextFeatureExtractor",
|
7 |
"image_mean": [
|
8 |
-
0.
|
9 |
-
0.
|
10 |
-
0.
|
11 |
],
|
12 |
"image_processor_type": "ViTFeatureExtractor",
|
13 |
"image_std": [
|
14 |
-
0.
|
15 |
-
0.
|
16 |
-
0.
|
17 |
],
|
18 |
-
"resample":
|
19 |
"rescale_factor": 0.00392156862745098,
|
20 |
"size": {
|
21 |
"height": 224,
|
|
|
1 |
{
|
|
|
2 |
"do_normalize": true,
|
3 |
"do_rescale": true,
|
4 |
"do_resize": true,
|
|
|
5 |
"image_mean": [
|
6 |
+
0.5,
|
7 |
+
0.5,
|
8 |
+
0.5
|
9 |
],
|
10 |
"image_processor_type": "ViTFeatureExtractor",
|
11 |
"image_std": [
|
12 |
+
0.5,
|
13 |
+
0.5,
|
14 |
+
0.5
|
15 |
],
|
16 |
+
"resample": 2,
|
17 |
"rescale_factor": 0.00392156862745098,
|
18 |
"size": {
|
19 |
"height": 224,
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 343723949
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ff6b4d036ae85dc3856cb9a06ed9bea056d3916563caa966d4b9cc24d8db950
|
3 |
size 343723949
|
runs/Apr25_03-19-25_7626d6fa6bff/events.out.tfevents.1682394094.7626d6fa6bff.149.5
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:293717aea2ee4d8e4b784677f00974b40fecf2ed07539844b93e5d052ab30ea6
|
3 |
+
size 411
|
runs/Apr25_03-47-00_7626d6fa6bff/1682394456.302469/events.out.tfevents.1682394456.7626d6fa6bff.149.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e216263065f0e8355953102ef7b50fa721f179fa9d99b23ad3599ce648cd6aae
|
3 |
+
size 5951
|
runs/Apr25_03-47-00_7626d6fa6bff/events.out.tfevents.1682394456.7626d6fa6bff.149.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f98926c4909215c5e62eccd6307d51d747ba6b4c1475654f0eda142010f7057
|
3 |
+
size 10946
|
train_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"total_flos": 1.8890449427391898e+18,
|
4 |
+
"train_loss": 3.100843106976088,
|
5 |
+
"train_runtime": 1215.6267,
|
6 |
+
"train_samples_per_second": 20.027,
|
7 |
+
"train_steps_per_second": 2.505
|
8 |
+
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3643
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3bf6a82183b97db3b1fb2a47966a3e237008df2b8458632237e02eaa60399d3b
|
3 |
size 3643
|