parlange commited on
Commit
f72f835
·
verified ·
1 Parent(s): 40f92a7

Upload ViT model from experiment a3

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. README.md +161 -0
  3. config.json +76 -0
  4. confusion_matrices/ViT_Confusion_Matrix_a.png +0 -0
  5. confusion_matrices/ViT_Confusion_Matrix_b.png +0 -0
  6. confusion_matrices/ViT_Confusion_Matrix_c.png +0 -0
  7. confusion_matrices/ViT_Confusion_Matrix_d.png +0 -0
  8. confusion_matrices/ViT_Confusion_Matrix_e.png +0 -0
  9. confusion_matrices/ViT_Confusion_Matrix_f.png +0 -0
  10. confusion_matrices/ViT_Confusion_Matrix_g.png +0 -0
  11. confusion_matrices/ViT_Confusion_Matrix_h.png +0 -0
  12. confusion_matrices/ViT_Confusion_Matrix_i.png +0 -0
  13. confusion_matrices/ViT_Confusion_Matrix_j.png +0 -0
  14. confusion_matrices/ViT_Confusion_Matrix_k.png +0 -0
  15. confusion_matrices/ViT_Confusion_Matrix_l.png +0 -0
  16. evaluation_results.csv +145 -0
  17. model.safetensors +3 -0
  18. pytorch_model.bin +3 -0
  19. roc_confusion_matrix/ViT_roc_confusion_matrix_a.png +0 -0
  20. roc_confusion_matrix/ViT_roc_confusion_matrix_b.png +0 -0
  21. roc_confusion_matrix/ViT_roc_confusion_matrix_c.png +0 -0
  22. roc_confusion_matrix/ViT_roc_confusion_matrix_d.png +0 -0
  23. roc_confusion_matrix/ViT_roc_confusion_matrix_e.png +0 -0
  24. roc_confusion_matrix/ViT_roc_confusion_matrix_f.png +0 -0
  25. roc_confusion_matrix/ViT_roc_confusion_matrix_g.png +0 -0
  26. roc_confusion_matrix/ViT_roc_confusion_matrix_h.png +0 -0
  27. roc_confusion_matrix/ViT_roc_confusion_matrix_i.png +0 -0
  28. roc_confusion_matrix/ViT_roc_confusion_matrix_j.png +0 -0
  29. roc_confusion_matrix/ViT_roc_confusion_matrix_k.png +0 -0
  30. roc_confusion_matrix/ViT_roc_confusion_matrix_l.png +0 -0
  31. roc_curves/ViT_ROC_a.png +0 -0
  32. roc_curves/ViT_ROC_b.png +0 -0
  33. roc_curves/ViT_ROC_c.png +0 -0
  34. roc_curves/ViT_ROC_d.png +0 -0
  35. roc_curves/ViT_ROC_e.png +0 -0
  36. roc_curves/ViT_ROC_f.png +0 -0
  37. roc_curves/ViT_ROC_g.png +0 -0
  38. roc_curves/ViT_ROC_h.png +0 -0
  39. roc_curves/ViT_ROC_i.png +0 -0
  40. roc_curves/ViT_ROC_j.png +0 -0
  41. roc_curves/ViT_ROC_k.png +0 -0
  42. roc_curves/ViT_ROC_l.png +0 -0
  43. training_curves/ViT_accuracy.png +0 -0
  44. training_curves/ViT_auc.png +0 -0
  45. training_curves/ViT_combined_metrics.png +3 -0
  46. training_curves/ViT_f1.png +0 -0
  47. training_curves/ViT_loss.png +0 -0
  48. training_curves/ViT_metrics.csv +41 -0
  49. training_metrics.csv +41 -0
  50. training_notebook_a3.ipynb +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ training_curves/ViT_combined_metrics.png filter=lfs diff=lfs merge=lfs -text
37
+ training_notebook_a3.ipynb filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,161 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - vision-transformer
5
+ - image-classification
6
+ - pytorch
7
+ - timm
8
+ - vit
9
+ - gravitational-lensing
10
+ - strong-lensing
11
+ - astronomy
12
+ - astrophysics
13
+ datasets:
14
+ - C21
15
+ metrics:
16
+ - accuracy
17
+ - auc
18
+ - f1
19
+ model-index:
20
+ - name: ViT-a3
21
+ results:
22
+ - task:
23
+ type: image-classification
24
+ name: Strong Gravitational Lens Discovery
25
+ dataset:
26
+ type: common-test-sample
27
+ name: Common Test Sample (More et al. 2024)
28
+ metrics:
29
+ - type: accuracy
30
+ value: 0.8340
31
+ name: Average Accuracy
32
+ - type: auc
33
+ value: 0.8374
34
+ name: Average AUC-ROC
35
+ - type: f1
36
+ value: 0.5537
37
+ name: Average F1-Score
38
+ ---
39
+
40
+ # 🌌 vit-gravit-a3
41
+
42
+ 🔭 This model is part of **GraViT**: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery
43
+
44
+ 🔗 **GitHub Repository**: [https://github.com/parlange/gravit](https://github.com/parlange/gravit)
45
+
46
+ ## 🛰️ Model Details
47
+
48
+ - **🤖 Model Type**: ViT
49
+ - **🧪 Experiment**: A3 - C21-all-blocks-ResNet18
50
+ - **🌌 Dataset**: C21
51
+ - **🪐 Fine-tuning Strategy**: all-blocks
52
+
53
+
54
+
55
+ ## 💻 Quick Start
56
+
57
+ ```python
58
+ import torch
59
+ import timm
60
+
61
+ # Load the model directly from the Hub
62
+ model = timm.create_model(
63
+ 'hf-hub:parlange/vit-gravit-a3',
64
+ pretrained=True
65
+ )
66
+ model.eval()
67
+
68
+ # Example inference
69
+ dummy_input = torch.randn(1, 3, 224, 224)
70
+ with torch.no_grad():
71
+ output = model(dummy_input)
72
+ predictions = torch.softmax(output, dim=1)
73
+ print(f"Lens probability: {predictions[0][1]:.4f}")
74
+ ```
75
+
76
+ ## ⚡️ Training Configuration
77
+
78
+ **Training Dataset:** C21 (Cañameras et al. 2021)
79
+ **Fine-tuning Strategy:** all-blocks
80
+
81
+
82
+ | 🔧 Parameter | 📝 Value |
83
+ |--------------|----------|
84
+ | Batch Size | 192 |
85
+ | Learning Rate | AdamW with ReduceLROnPlateau |
86
+ | Epochs | 100 |
87
+ | Patience | 10 |
88
+ | Optimizer | AdamW |
89
+ | Scheduler | ReduceLROnPlateau |
90
+ | Image Size | 224x224 |
91
+ | Fine Tune Mode | all_blocks |
92
+ | Stochastic Depth Probability | 0.1 |
93
+
94
+
95
+ ## 📈 Training Curves
96
+
97
+ ![Combined Training Metrics](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/training_curves/ViT_combined_metrics.png)
98
+
99
+
100
+ ## 🏁 Final Epoch Training Metrics
101
+
102
+ | Metric | Training | Validation |
103
+ |:---------:|:-----------:|:-------------:|
104
+ | 📉 Loss | 0.0060 | 0.0320 |
105
+ | 🎯 Accuracy | 0.9978 | 0.9950 |
106
+ | 📊 AUC-ROC | 1.0000 | 0.9997 |
107
+ | ⚖️ F1 Score | 0.9978 | 0.9950 |
108
+
109
+
110
+ ## ☑️ Evaluation Results
111
+
112
+ ### ROC Curves and Confusion Matrices
113
+
114
+ Performance across all test datasets (a through l) in the Common Test Sample (More et al. 2024):
115
+
116
+ ![ROC + Confusion Matrix - Dataset A](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_a.png)
117
+ ![ROC + Confusion Matrix - Dataset B](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_b.png)
118
+ ![ROC + Confusion Matrix - Dataset C](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_c.png)
119
+ ![ROC + Confusion Matrix - Dataset D](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_d.png)
120
+ ![ROC + Confusion Matrix - Dataset E](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_e.png)
121
+ ![ROC + Confusion Matrix - Dataset F](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_f.png)
122
+ ![ROC + Confusion Matrix - Dataset G](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_g.png)
123
+ ![ROC + Confusion Matrix - Dataset H](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_h.png)
124
+ ![ROC + Confusion Matrix - Dataset I](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_i.png)
125
+ ![ROC + Confusion Matrix - Dataset J](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_j.png)
126
+ ![ROC + Confusion Matrix - Dataset K](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_k.png)
127
+ ![ROC + Confusion Matrix - Dataset L](https://huggingface.co/parlange/vit-gravit-a3/resolve/main/roc_confusion_matrix/ViT_roc_confusion_matrix_l.png)
128
+
129
+ ### 📋 Performance Summary
130
+
131
+ Average performance across 12 test datasets from the Common Test Sample (More et al. 2024):
132
+
133
+ | Metric | Value |
134
+ |-----------|----------|
135
+ | 🎯 Average Accuracy | 0.8340 |
136
+ | 📈 Average AUC-ROC | 0.8374 |
137
+ | ⚖️ Average F1-Score | 0.5537 |
138
+
139
+
140
+ ## 📘 Citation
141
+
142
+ If you use this model in your research, please cite:
143
+
144
+ ```bibtex
145
+ @misc{parlange2025gravit,
146
+ title={GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery},
147
+ author={René Parlange and Juan C. Cuevas-Tello and Octavio Valenzuela and Omar de J. Cabrera-Rosas and Tomás Verdugo and Anupreeta More and Anton T. Jaelani},
148
+ year={2025},
149
+ eprint={2509.00226},
150
+ archivePrefix={arXiv},
151
+ primaryClass={cs.CV},
152
+ url={https://arxiv.org/abs/2509.00226},
153
+ }
154
+ ```
155
+
156
+ ---
157
+
158
+
159
+ ## Model Card Contact
160
+
161
+ For questions about this model, please contact the author through: https://github.com/parlange/
config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architecture": "vit_base_patch16_224",
3
+ "num_classes": 2,
4
+ "num_features": 768,
5
+ "global_pool": "token",
6
+ "crop_pct": 0.875,
7
+ "interpolation": "bicubic",
8
+ "mean": [
9
+ 0.485,
10
+ 0.456,
11
+ 0.406
12
+ ],
13
+ "std": [
14
+ 0.229,
15
+ 0.224,
16
+ 0.225
17
+ ],
18
+ "first_conv": "patch_embed.proj",
19
+ "classifier": "head",
20
+ "input_size": [
21
+ 3,
22
+ 224,
23
+ 224
24
+ ],
25
+ "pool_size": [
26
+ 7,
27
+ 7
28
+ ],
29
+ "pretrained_cfg": {
30
+ "tag": "gravit_a3",
31
+ "custom_load": false,
32
+ "input_size": [
33
+ 3,
34
+ 224,
35
+ 224
36
+ ],
37
+ "fixed_input_size": true,
38
+ "interpolation": "bicubic",
39
+ "crop_pct": 0.875,
40
+ "crop_mode": "center",
41
+ "mean": [
42
+ 0.485,
43
+ 0.456,
44
+ 0.406
45
+ ],
46
+ "std": [
47
+ 0.229,
48
+ 0.224,
49
+ 0.225
50
+ ],
51
+ "num_classes": 2,
52
+ "pool_size": [
53
+ 7,
54
+ 7
55
+ ],
56
+ "first_conv": "patch_embed.proj",
57
+ "classifier": "head"
58
+ },
59
+ "model_name": "vit_gravit_a3",
60
+ "experiment": "a3",
61
+ "training_strategy": "all-blocks",
62
+ "dataset": "C21",
63
+ "hyperparameters": {
64
+ "batch_size": "192",
65
+ "learning_rate": "AdamW with ReduceLROnPlateau",
66
+ "epochs": "100",
67
+ "patience": "10",
68
+ "optimizer": "AdamW",
69
+ "scheduler": "ReduceLROnPlateau",
70
+ "image_size": "224x224",
71
+ "fine_tune_mode": "all_blocks",
72
+ "stochastic_depth_probability": "0.1"
73
+ },
74
+ "hf_hub_id": "parlange/vit-gravit-a3",
75
+ "license": "apache-2.0"
76
+ }
confusion_matrices/ViT_Confusion_Matrix_a.png ADDED
confusion_matrices/ViT_Confusion_Matrix_b.png ADDED
confusion_matrices/ViT_Confusion_Matrix_c.png ADDED
confusion_matrices/ViT_Confusion_Matrix_d.png ADDED
confusion_matrices/ViT_Confusion_Matrix_e.png ADDED
confusion_matrices/ViT_Confusion_Matrix_f.png ADDED
confusion_matrices/ViT_Confusion_Matrix_g.png ADDED
confusion_matrices/ViT_Confusion_Matrix_h.png ADDED
confusion_matrices/ViT_Confusion_Matrix_i.png ADDED
confusion_matrices/ViT_Confusion_Matrix_j.png ADDED
confusion_matrices/ViT_Confusion_Matrix_k.png ADDED
confusion_matrices/ViT_Confusion_Matrix_l.png ADDED
evaluation_results.csv ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model,Dataset,Loss,Accuracy,AUCROC,F1
2
+ ViT,a,0.3652414279945689,0.8912291732159698,0.8917799263351749,0.4383116883116883
3
+ ViT,b,0.21351260967325467,0.9459289531593839,0.9444456721915285,0.6108597285067874
4
+ ViT,c,0.850089883324636,0.758252121974222,0.8170580110497236,0.2598652550529355
5
+ ViT,d,0.125797027155771,0.9723357434768941,0.9632872928176796,0.7541899441340782
6
+ ViT,e,0.5662868288882347,0.8803512623490669,0.9096268826156058,0.712401055408971
7
+ ViT,f,0.3309973750718474,0.8993106653241422,0.9044572137856804,0.17197452229299362
8
+ ViT,g,0.06683265567384661,0.9775,0.9995110555555556,0.9779303580186366
9
+ ViT,h,0.4043246931489557,0.878,0.9965371111111111,0.8909740840035746
10
+ ViT,i,0.020328776298090816,0.9915,0.999877,0.991546494281452
11
+ ViT,j,7.168746640741825,0.509,0.4538551111111111,0.1088929219600726
12
+ ViT,k,7.1222426953688265,0.523,0.48649888888888887,0.11173184357541899
13
+ ViT,l,2.481225689433318,0.7816614668711332,0.6814674636422873,0.6155851410483195
14
+ MLP-Mixer,a,0.16029433381890995,0.9534737503929582,0.8991058931860036,0.6084656084656085
15
+ MLP-Mixer,b,0.1373700322756056,0.9629047469349261,0.9401123388581952,0.6609195402298851
16
+ MLP-Mixer,c,0.27720538959380253,0.9141779314680918,0.8577974217311233,0.4572564612326044
17
+ MLP-Mixer,d,0.11598697743656425,0.9682489783087079,0.955084714548803,0.6948640483383686
18
+ MLP-Mixer,e,0.40732662754315313,0.8957189901207464,0.9144403239234089,0.7076923076923077
19
+ MLP-Mixer,f,0.10986769875007599,0.9635194795135931,0.9131062483453624,0.3281027104136947
20
+ MLP-Mixer,g,0.028587095644325017,0.9901666666666666,0.9997029999999999,0.9902398676592225
21
+ MLP-Mixer,h,0.10272313961014151,0.9643333333333334,0.998619888888889,0.9654838709677419
22
+ MLP-Mixer,i,0.01725051350519061,0.993,0.9998728888888889,0.9930325149303252
23
+ MLP-Mixer,j,3.9579579369425772,0.5163333333333333,0.6010726111111112,0.09369144284821987
24
+ MLP-Mixer,k,3.9466213275045154,0.5191666666666667,0.6683313333333333,0.09419152276295134
25
+ MLP-Mixer,l,1.3240398510736613,0.8240177674369414,0.744620509879184,0.6619260463226331
26
+ CvT,a,0.2081917969545678,0.9352404904118202,0.9240874769797423,0.5672268907563025
27
+ CvT,b,0.15224700975965233,0.9575605155611443,0.9504567219152855,0.6666666666666666
28
+ CvT,c,0.49155200616401235,0.8597925180760767,0.8735598526703499,0.3770949720670391
29
+ CvT,d,0.09181369168063785,0.9739075762338887,0.9780386740331491,0.7648725212464589
30
+ CvT,e,0.32513720903680565,0.9143798024149287,0.9261106486036479,0.7758620689655172
31
+ CvT,f,0.18988747454602042,0.940825652544342,0.9312245836823443,0.2611218568665377
32
+ CvT,g,0.04789180162362754,0.9845,0.9996506666666666,0.9847165160230074
33
+ CvT,h,0.22777999278716743,0.9326666666666666,0.9980902777777778,0.9368355222013759
34
+ CvT,i,0.01585208964161575,0.9931666666666666,0.9999184444444444,0.9932040444223438
35
+ CvT,j,4.557504334926605,0.5136666666666667,0.4219552222222222,0.1049079754601227
36
+ CvT,k,4.525464609175921,0.5223333333333333,0.7008621111111111,0.10660847880299251
37
+ CvT,l,1.5619914421430425,0.809793242028449,0.7106891758223672,0.6473875110283306
38
+ Swin,a,0.13412107322078123,0.9607041810751336,0.921658379373849,0.6753246753246753
39
+ Swin,b,0.12724596003046465,0.9556743162527507,0.9398747697974218,0.6483790523690773
40
+ Swin,c,0.17965890587556665,0.9443571204023892,0.9029281767955801,0.5949656750572082
41
+ Swin,d,0.06030154695547883,0.9833385727758567,0.9931565377532229,0.8306709265175719
42
+ Swin,e,0.4499740816497384,0.8562019758507134,0.8587376069022933,0.6649616368286445
43
+ Swin,f,0.09911939381673643,0.9672372395631632,0.9347786366220656,0.3806734992679356
44
+ Swin,g,0.038731103701516986,0.985,0.9998738888888888,0.9852216748768473
45
+ Swin,h,0.06651869903318584,0.979,0.999785,0.9794319294809011
46
+ Swin,i,0.0032394034853205087,0.9996666666666667,0.9999967777777777,0.9996667777407531
47
+ Swin,j,5.625512751281262,0.496,0.10115866666666667,0.04182509505703422
48
+ Swin,k,5.590021039650775,0.5106666666666667,0.37088833333333326,0.04302477183833116
49
+ Swin,l,1.8411567582560506,0.822484268415208,0.5854755526940438,0.65565699046056
50
+ CaiT,a,0.20098744187104706,0.9478151524677775,0.9047292817679559,0.6047619047619047
51
+ CaiT,b,0.12978406120776775,0.9685633448601069,0.9467771639042358,0.7175141242937854
52
+ CaiT,c,0.2854890818960897,0.9248663942156554,0.8821031307550645,0.5152129817444219
53
+ CaiT,d,0.10324450797222483,0.9789374410562716,0.9716519337016575,0.7912772585669782
54
+ CaiT,e,0.4533339374577044,0.8990120746432492,0.9009763112086581,0.7341040462427746
55
+ CaiT,f,0.11921021888981406,0.9653009062040121,0.9248623124563286,0.36182336182336183
56
+ CaiT,g,0.02093089486740064,0.9923333333333333,0.9999474444444445,0.9923916639100232
57
+ CaiT,h,0.10348050882376265,0.9691666666666666,0.9995327777777778,0.9700889248181084
58
+ CaiT,i,0.006860514354542829,0.9978333333333333,0.9999788888888889,0.9978380176284717
59
+ CaiT,j,4.86097429022938,0.5145,0.5177155,0.08367411135577225
60
+ CaiT,k,4.8469039183312566,0.52,0.7220807222222222,0.0845518118245391
61
+ CaiT,l,1.617374696626156,0.8247051980328909,0.7258400455085416,0.6629384850025419
62
+ DeiT,a,0.27381800207860135,0.8953159383841559,0.8986252302025783,0.43463497453310695
63
+ DeiT,b,0.1319006348473214,0.9534737503929582,0.9522191528545119,0.6336633663366337
64
+ DeiT,c,0.5039299008644665,0.7972335743476894,0.8374493554327808,0.28412874583795783
65
+ DeiT,d,0.07908134242881022,0.9742219427852876,0.9686187845303867,0.757396449704142
66
+ DeiT,e,0.3087258418422892,0.8869374313940724,0.9196775902520246,0.713091922005571
67
+ DeiT,f,0.2156604077872147,0.9148787855317171,0.9145406292179694,0.18892988929889298
68
+ DeiT,g,0.04722818533703685,0.9821666666666666,0.9995671111111111,0.9824099950682229
69
+ DeiT,h,0.24446571580693124,0.8993333333333333,0.9973094444444446,0.9082066869300912
70
+ DeiT,i,0.01922516017779708,0.9931666666666666,0.9998760000000001,0.9931859730762839
71
+ DeiT,j,3.201324864923954,0.5038333333333334,0.5165665555555555,0.07345160286336756
72
+ DeiT,k,3.173321856930852,0.5148333333333334,0.5144209444444445,0.074992055926279
73
+ DeiT,l,1.1505240741554759,0.7888530484902967,0.6992895835538881,0.6182965299684543
74
+ DeiT3,a,0.14936432559529178,0.9519019176359635,0.9349825046040517,0.6240786240786241
75
+ DeiT3,b,0.14933169767358925,0.9515875510845646,0.9439152854511971,0.6225490196078431
76
+ DeiT3,c,0.22061521056612046,0.9368123231688148,0.9148646408839779,0.5582417582417583
77
+ DeiT3,d,0.10169743415564346,0.9682489783087079,0.9643388581952119,0.7154929577464789
78
+ DeiT3,e,0.49460477683729975,0.8518111964873765,0.8834708241882994,0.6529562982005142
79
+ DeiT3,f,0.13154652377257423,0.9591046394547286,0.9363108852365101,0.3248081841432225
80
+ DeiT3,g,0.05130978459212929,0.9823333333333333,0.9997385555555556,0.982605841811618
81
+ DeiT3,h,0.08910193234775216,0.9745,0.9995217222222222,0.975085490962384
82
+ DeiT3,i,0.02605568784568459,0.9911666666666666,0.9998692222222222,0.9912266181095845
83
+ DeiT3,j,3.271038075208664,0.5056666666666667,0.4370591666666666,0.08286951144094001
84
+ DeiT3,k,3.245784008204937,0.5145,0.5090863888888889,0.08425023577491354
85
+ DeiT3,l,1.112910136327876,0.8202104595209138,0.7017805623573649,0.656912209889001
86
+ Twins_SVT,a,0.18666393640786363,0.9355548569632192,0.9178057090239411,0.5665961945031712
87
+ Twins_SVT,b,0.12907026239917624,0.9559886828041496,0.958646408839779,0.6568627450980392
88
+ Twins_SVT,c,0.37308393637551635,0.8491040553285131,0.8678618784530387,0.3582887700534759
89
+ Twins_SVT,d,0.07745791368212238,0.9789374410562716,0.982377532228361,0.8
90
+ Twins_SVT,e,0.46567638177510545,0.8572996706915478,0.8989404374479679,0.6733668341708543
91
+ Twins_SVT,f,0.1620182347072841,0.9354039191387189,0.9297958448525039,0.24319419237749546
92
+ Twins_SVT,g,0.03985891605913639,0.9833333333333333,0.9995775555555556,0.9835688465330266
93
+ Twins_SVT,h,0.16922680978477,0.9266666666666666,0.9980967777777777,0.9315281668222845
94
+ Twins_SVT,i,0.012495769090950489,0.9955,0.9999184444444444,0.9955097289206719
95
+ Twins_SVT,j,5.5489834444224835,0.49816666666666665,0.40547916666666667,0.051653543307086616
96
+ Twins_SVT,k,5.521620307348669,0.5103333333333333,0.5116922222222222,0.052869116698903935
97
+ Twins_SVT,l,1.8607527904559573,0.8012267992173867,0.6587701738584388,0.6306377124889456
98
+ Twins_PCPVT,a,0.5452342244934938,0.7969192077962904,0.895244014732965,0.3228511530398323
99
+ Twins_PCPVT,b,0.33666214395135274,0.8836843759823955,0.926718232044199,0.45427728613569324
100
+ Twins_PCPVT,c,0.8452249476460681,0.680289217227287,0.8652670349907919,0.23245283018867924
101
+ Twins_PCPVT,d,0.14044462036633035,0.9487582521219742,0.9665690607734807,0.6539278131634819
102
+ Twins_PCPVT,e,0.9610598787387299,0.6893523600439078,0.849345341708923,0.5211505922165821
103
+ Twins_PCPVT,f,0.4996267252911643,0.8163581442181086,0.909773537083411,0.11496827174318776
104
+ Twins_PCPVT,g,0.16858429829776286,0.9403333333333334,0.9974785,0.9434260429835651
105
+ Twins_PCPVT,h,0.4382073585242033,0.8325,0.9937448888888889,0.8559139784946237
106
+ Twins_PCPVT,i,0.06455629007518292,0.9748333333333333,0.999179,0.97533082829603
107
+ Twins_PCPVT,j,2.452636483669281,0.5051666666666667,0.3821618888888889,0.20123755716976055
108
+ Twins_PCPVT,k,2.348608487725258,0.5396666666666666,0.6640067222222222,0.21310541310541312
109
+ Twins_PCPVT,l,1.069163286557427,0.7349690656231823,0.6925868024466941,0.5836517693969098
110
+ PiT,a,1.4542810870805876,0.7161270040867652,0.9052854511970534,0.26645004061738425
111
+ PiT,b,0.48083062532262133,0.8921722728701666,0.9610782688766114,0.488822652757079
112
+ PiT,c,3.0596617273248543,0.5278214397988054,0.823461325966851,0.17923497267759564
113
+ PiT,d,0.0542862427054065,0.9864822382898459,0.9955837937384899,0.8840970350404312
114
+ PiT,e,1.2118028251844757,0.7727771679473107,0.9089305986528419,0.6130841121495327
115
+ PiT,f,1.3020594693266185,0.7730617303074897,0.9206398944503998,0.10067526089625538
116
+ PiT,g,0.2426671743527986,0.9443333333333334,0.9978501666666667,0.9471351693573916
117
+ PiT,h,1.609877428545151,0.7511666666666666,0.97719,0.8003209843520128
118
+ PiT,i,0.01652756105083972,0.9943333333333333,0.999909,0.9943502824858758
119
+ PiT,j,4.793344738483429,0.4985,0.4088527777777778,0.1740323908866319
120
+ PiT,k,4.567205076335464,0.5485,0.8206032222222223,0.18965001495662578
121
+ PiT,l,2.3343544872733246,0.7027655861667812,0.7048351046138528,0.5527174345508077
122
+ ResNet-18,a,1.1835124935604648,0.6365922665828356,0.9285580110497238,0.22830440587449932
123
+ ResNet-18,b,1.0141342383478842,0.7117258723671801,0.936340699815838,0.2716441620333598
124
+ ResNet-18,c,1.8204121201415364,0.5328513046211883,0.8883425414364641,0.18708971553610504
125
+ ResNet-18,d,0.011235734144227473,0.9949701351776171,0.9996408839779005,0.9553072625698324
126
+ ResNet-18,e,1.056551858831839,0.6125137211855104,0.9263868916975707,0.4920863309352518
127
+ ResNet-18,f,1.0591482175922742,0.6988614359848192,0.9375419355678716,0.08085106382978724
128
+ ResNet-18,g,0.5340945276358107,0.8486666666666667,0.9978233333333334,0.8685201274254272
129
+ ResNet-18,h,0.9615578431227186,0.7538333333333334,0.9955927777777778,0.8024080267558529
130
+ ResNet-18,i,0.002392592921940377,0.9988333333333334,0.9999976666666667,0.998834304746045
131
+ ResNet-18,j,7.01291295003891,0.35383333333333333,0.07525633333333334,0.01524003048006096
132
+ ResNet-18,k,6.481211027059704,0.504,0.7462510555555556,0.019762845849802372
133
+ ResNet-18,l,2.7791932088225773,0.6373010417217493,0.6165565470444929,0.4826910023380345
134
+ Ensemble,a,,0.9487582521219742,0.9267523020257826,0.6320541760722348
135
+ Ensemble,b,,0.9698208110657026,0.9539668508287292,0.7446808510638298
136
+ Ensemble,c,,0.8924866394215656,0.8899631675874772,0.45016077170418006
137
+ Ensemble,d,,0.9823954731216599,0.9802780847145488,0.8333333333333334
138
+ Ensemble,e,,0.9023051591657519,0.9130326193899946,0.7588075880758808
139
+ Ensemble,f,,0.9549221593989621,0.9363232543302679,0.3248259860788863
140
+ Ensemble,g,,0.9905,0.999917,0.9905831818932761
141
+ Ensemble,h,,0.9495,0.9994414444444444,0.9518971265280203
142
+ Ensemble,i,,0.9971666666666666,0.9999844444444443,0.997172792283386
143
+ Ensemble,j,,0.5115,0.3640104444444444,0.07801195344447939
144
+ Ensemble,k,,0.5181666666666667,0.6087856666666667,0.07900605288308378
145
+ Ensemble,l,,0.817037702924224,0.6790663441280981,0.6534455128205128
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ca3e493402ce2d2572f804e125aa63cba69ddd2a333021a823d8821ba61a47
3
+ size 343214864
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61688f57296e5e0fde9bffa9e785857982cebffd7ceae916a938bd946ac0d1d8
3
+ size 343259038
roc_confusion_matrix/ViT_roc_confusion_matrix_a.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_b.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_c.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_d.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_e.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_f.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_g.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_h.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_i.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_j.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_k.png ADDED
roc_confusion_matrix/ViT_roc_confusion_matrix_l.png ADDED
roc_curves/ViT_ROC_a.png ADDED
roc_curves/ViT_ROC_b.png ADDED
roc_curves/ViT_ROC_c.png ADDED
roc_curves/ViT_ROC_d.png ADDED
roc_curves/ViT_ROC_e.png ADDED
roc_curves/ViT_ROC_f.png ADDED
roc_curves/ViT_ROC_g.png ADDED
roc_curves/ViT_ROC_h.png ADDED
roc_curves/ViT_ROC_i.png ADDED
roc_curves/ViT_ROC_j.png ADDED
roc_curves/ViT_ROC_k.png ADDED
roc_curves/ViT_ROC_l.png ADDED
training_curves/ViT_accuracy.png ADDED
training_curves/ViT_auc.png ADDED
training_curves/ViT_combined_metrics.png ADDED

Git LFS Details

  • SHA256: 6c2c9ecacc12d393fde28ec01906a887bcc8fd027a68cfdfe7e300e2a39fca82
  • Pointer size: 131 Bytes
  • Size of remote file: 156 kB
training_curves/ViT_f1.png ADDED
training_curves/ViT_loss.png ADDED
training_curves/ViT_metrics.csv ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
2
+ 1,0.1594870525151491,0.05837649315595627,0.94505,0.974,0.985395874375,0.99834,0.9451637851458224,0.9736842105263158
3
+ 2,0.04891255444511771,0.04843974649906158,0.9819125,0.984,0.99838926375,0.998526,0.9819009618631878,0.9840637450199203
4
+ 3,0.04098877249099314,0.06516808295249939,0.985425,0.975,0.998859596875,0.997988,0.9854198969639374,0.9749247743229689
5
+ 4,0.03653456708919257,0.04256650495901704,0.9867375,0.986,0.9990655153125,0.9992679999999999,0.9867379973251003,0.9858585858585859
6
+ 5,0.03568214138075709,0.05940364444255829,0.9868,0.979,0.9991360790624999,0.9983299999999999,0.9868,0.9787234042553191
7
+ 6,0.029999172747135164,0.03990015828609467,0.9892375,0.988,0.999345638125,0.998996,0.9892325200405188,0.9879759519038076
8
+ 7,0.03169462961144745,0.04534034144878388,0.988275,0.988,0.9993018990624999,0.9993799999999999,0.9882685477012357,0.9879275653923542
9
+ 8,0.029803055073274298,0.04331541895493865,0.9892625,0.983,0.999391361875,0.99799,0.989260486341189,0.9829145728643216
10
+ 9,0.02843129145205021,0.04371415561437607,0.9896125,0.985,0.999460985,0.9982759999999999,0.9896118507406713,0.9850746268656716
11
+ 10,0.027733767334651202,0.03946993482299149,0.9898625,0.988,0.9994940365625,0.998422,0.989861612891128,0.9879759519038076
12
+ 11,0.026831346929166465,0.054907928824424745,0.9900875,0.982,0.99951548375,0.998138,0.9900891105195406,0.9820717131474104
13
+ 12,0.02800533079933375,0.050178519189357756,0.99005,0.982,0.9994389078124999,0.9985700000000001,0.990046018407363,0.9820717131474104
14
+ 13,0.025427712446730585,0.0769548636674881,0.9906875,0.977,0.9995551875000002,0.9972219999999999,0.9906887802927098,0.977205153617443
15
+ 14,0.025302810053154826,0.052582308113574984,0.9909125,0.981,0.9995339706249999,0.998996,0.9909114775412234,0.9809428284854563
16
+ 15,0.026751336156297476,0.043837911561131475,0.9904125,0.987,0.999521585625,0.998282,0.990409502969678,0.986986986986987
17
+ 16,0.024374210640136153,0.05087221178412438,0.9911125,0.983,0.999586088125,0.9994480000000001,0.9911117222756991,0.9827411167512691
18
+ 17,0.014263770198030398,0.03483956679701805,0.9949125,0.988,0.9998447559375,0.999428,0.9949105278295339,0.9880239520958084
19
+ 18,0.01196307003966067,0.032166809320508034,0.995475,0.987,0.999904258125,0.999582,0.9954736420926278,0.9870388833499502
20
+ 19,0.011337511599739083,0.027693722724914552,0.9956875,0.991,0.9999116637500001,0.999648,0.9956874460930761,0.9909729187562688
21
+ 20,0.010317483875318431,0.026324373602867126,0.9964125,0.992,0.9999230290625,0.999668,0.9964117374942175,0.992
22
+ 21,0.009391166536684613,0.031228957712650298,0.99665,0.988,0.999930400625,0.999572,0.9966499162479062,0.9879759519038076
23
+ 22,0.00938724503798876,0.03080810856842436,0.9965875,0.989,0.9999409593750002,0.9996319999999999,0.9965871160505557,0.988988988988989
24
+ 23,0.00925770991302561,0.036078432086855176,0.99675,0.989,0.9999399396875,0.9995940000000001,0.9967496749674968,0.9890329012961117
25
+ 24,0.008796912119051557,0.030019028067586988,0.9971125,0.988,0.9999192143750002,0.9997039999999999,0.9971122473216406,0.9880478087649402
26
+ 25,0.009039205305592622,0.03161667108512484,0.99685,0.992,0.9999435053125,0.999608,0.9968496062007751,0.9920159680638723
27
+ 26,0.00832465733307763,0.03130640342843253,0.9968125,0.989,0.9999427559375,0.999682,0.9968128585534127,0.988988988988989
28
+ 27,0.007553178189964092,0.02917952951043844,0.9972,0.99,0.9999617928124999,0.99969,0.9972,0.9899799599198397
29
+ 28,0.007608072868280578,0.028727686047554015,0.99705,0.991,0.9999638615625001,0.999696,0.9970501474926253,0.990990990990991
30
+ 29,0.007120153656997718,0.029447013556957245,0.9973375,0.991,0.9999664453124999,0.999688,0.9973374001525057,0.990990990990991
31
+ 30,0.007157817871007137,0.02913126605004072,0.9975625,0.991,0.9999617618750001,0.999682,0.9975616160858312,0.990990990990991
32
+ 31,0.007336907451570732,0.028489415229298175,0.9971875,0.991,0.9999640784374999,0.999708,0.9971867615249003,0.991008991008991
33
+ 32,0.006501691047073109,0.03007084161043167,0.9977375,0.992,0.999968305,0.999672,0.9977373020139262,0.992
34
+ 33,0.00693205452451366,0.02848208716508816,0.9975,0.993,0.9999662856249999,0.999718,0.9974997499749975,0.993006993006993
35
+ 34,0.007461001089593629,0.028252465724945067,0.997175,0.992,0.9999625840624999,0.99971,0.997174576186428,0.992
36
+ 35,0.00710381545725686,0.029008929789066315,0.9974125,0.993,0.9999660171875,0.9997199999999999,0.9974120147527662,0.992992992992993
37
+ 36,0.006280751934862929,0.030127739772200585,0.9976625,0.992,0.999974466875,0.9997020000000001,0.9976623538971185,0.9919839679358717
38
+ 37,0.006127115182683338,0.03197662199282786,0.9976625,0.991,0.9999751071874999,0.99969,0.9976624123404627,0.990990990990991
39
+ 38,0.00666614195129805,0.030885459803044794,0.9975,0.992,0.9999712131249999,0.999694,0.9974996249437416,0.992
40
+ 39,0.007003229365494917,0.030323315382003783,0.9974375,0.992,0.9999688965624999,0.999702,0.9974374679683496,0.992
41
+ 40,0.006024414020474069,0.03197560523450375,0.9977875,0.995,0.9999746728125001,0.999688,0.997786586967124,0.995004995004995
training_metrics.csv ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
2
+ 1,0.1594870525151491,0.05837649315595627,0.94505,0.974,0.985395874375,0.99834,0.9451637851458224,0.9736842105263158
3
+ 2,0.04891255444511771,0.04843974649906158,0.9819125,0.984,0.99838926375,0.998526,0.9819009618631878,0.9840637450199203
4
+ 3,0.04098877249099314,0.06516808295249939,0.985425,0.975,0.998859596875,0.997988,0.9854198969639374,0.9749247743229689
5
+ 4,0.03653456708919257,0.04256650495901704,0.9867375,0.986,0.9990655153125,0.9992679999999999,0.9867379973251003,0.9858585858585859
6
+ 5,0.03568214138075709,0.05940364444255829,0.9868,0.979,0.9991360790624999,0.9983299999999999,0.9868,0.9787234042553191
7
+ 6,0.029999172747135164,0.03990015828609467,0.9892375,0.988,0.999345638125,0.998996,0.9892325200405188,0.9879759519038076
8
+ 7,0.03169462961144745,0.04534034144878388,0.988275,0.988,0.9993018990624999,0.9993799999999999,0.9882685477012357,0.9879275653923542
9
+ 8,0.029803055073274298,0.04331541895493865,0.9892625,0.983,0.999391361875,0.99799,0.989260486341189,0.9829145728643216
10
+ 9,0.02843129145205021,0.04371415561437607,0.9896125,0.985,0.999460985,0.9982759999999999,0.9896118507406713,0.9850746268656716
11
+ 10,0.027733767334651202,0.03946993482299149,0.9898625,0.988,0.9994940365625,0.998422,0.989861612891128,0.9879759519038076
12
+ 11,0.026831346929166465,0.054907928824424745,0.9900875,0.982,0.99951548375,0.998138,0.9900891105195406,0.9820717131474104
13
+ 12,0.02800533079933375,0.050178519189357756,0.99005,0.982,0.9994389078124999,0.9985700000000001,0.990046018407363,0.9820717131474104
14
+ 13,0.025427712446730585,0.0769548636674881,0.9906875,0.977,0.9995551875000002,0.9972219999999999,0.9906887802927098,0.977205153617443
15
+ 14,0.025302810053154826,0.052582308113574984,0.9909125,0.981,0.9995339706249999,0.998996,0.9909114775412234,0.9809428284854563
16
+ 15,0.026751336156297476,0.043837911561131475,0.9904125,0.987,0.999521585625,0.998282,0.990409502969678,0.986986986986987
17
+ 16,0.024374210640136153,0.05087221178412438,0.9911125,0.983,0.999586088125,0.9994480000000001,0.9911117222756991,0.9827411167512691
18
+ 17,0.014263770198030398,0.03483956679701805,0.9949125,0.988,0.9998447559375,0.999428,0.9949105278295339,0.9880239520958084
19
+ 18,0.01196307003966067,0.032166809320508034,0.995475,0.987,0.999904258125,0.999582,0.9954736420926278,0.9870388833499502
20
+ 19,0.011337511599739083,0.027693722724914552,0.9956875,0.991,0.9999116637500001,0.999648,0.9956874460930761,0.9909729187562688
21
+ 20,0.010317483875318431,0.026324373602867126,0.9964125,0.992,0.9999230290625,0.999668,0.9964117374942175,0.992
22
+ 21,0.009391166536684613,0.031228957712650298,0.99665,0.988,0.999930400625,0.999572,0.9966499162479062,0.9879759519038076
23
+ 22,0.00938724503798876,0.03080810856842436,0.9965875,0.989,0.9999409593750002,0.9996319999999999,0.9965871160505557,0.988988988988989
24
+ 23,0.00925770991302561,0.036078432086855176,0.99675,0.989,0.9999399396875,0.9995940000000001,0.9967496749674968,0.9890329012961117
25
+ 24,0.008796912119051557,0.030019028067586988,0.9971125,0.988,0.9999192143750002,0.9997039999999999,0.9971122473216406,0.9880478087649402
26
+ 25,0.009039205305592622,0.03161667108512484,0.99685,0.992,0.9999435053125,0.999608,0.9968496062007751,0.9920159680638723
27
+ 26,0.00832465733307763,0.03130640342843253,0.9968125,0.989,0.9999427559375,0.999682,0.9968128585534127,0.988988988988989
28
+ 27,0.007553178189964092,0.02917952951043844,0.9972,0.99,0.9999617928124999,0.99969,0.9972,0.9899799599198397
29
+ 28,0.007608072868280578,0.028727686047554015,0.99705,0.991,0.9999638615625001,0.999696,0.9970501474926253,0.990990990990991
30
+ 29,0.007120153656997718,0.029447013556957245,0.9973375,0.991,0.9999664453124999,0.999688,0.9973374001525057,0.990990990990991
31
+ 30,0.007157817871007137,0.02913126605004072,0.9975625,0.991,0.9999617618750001,0.999682,0.9975616160858312,0.990990990990991
32
+ 31,0.007336907451570732,0.028489415229298175,0.9971875,0.991,0.9999640784374999,0.999708,0.9971867615249003,0.991008991008991
33
+ 32,0.006501691047073109,0.03007084161043167,0.9977375,0.992,0.999968305,0.999672,0.9977373020139262,0.992
34
+ 33,0.00693205452451366,0.02848208716508816,0.9975,0.993,0.9999662856249999,0.999718,0.9974997499749975,0.993006993006993
35
+ 34,0.007461001089593629,0.028252465724945067,0.997175,0.992,0.9999625840624999,0.99971,0.997174576186428,0.992
36
+ 35,0.00710381545725686,0.029008929789066315,0.9974125,0.993,0.9999660171875,0.9997199999999999,0.9974120147527662,0.992992992992993
37
+ 36,0.006280751934862929,0.030127739772200585,0.9976625,0.992,0.999974466875,0.9997020000000001,0.9976623538971185,0.9919839679358717
38
+ 37,0.006127115182683338,0.03197662199282786,0.9976625,0.991,0.9999751071874999,0.99969,0.9976624123404627,0.990990990990991
39
+ 38,0.00666614195129805,0.030885459803044794,0.9975,0.992,0.9999712131249999,0.999694,0.9974996249437416,0.992
40
+ 39,0.007003229365494917,0.030323315382003783,0.9974375,0.992,0.9999688965624999,0.999702,0.9974374679683496,0.992
41
+ 40,0.006024414020474069,0.03197560523450375,0.9977875,0.995,0.9999746728125001,0.999688,0.997786586967124,0.995004995004995
training_notebook_a3.ipynb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:114b011c655469b68717044f932313045fe1581f964f013703e81d6d48adcb00
3
+ size 21999529