Upload ViT model from experiment c2
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +2 -0
- README.md +166 -0
- config.json +76 -0
- confusion_matrices/ViT_Confusion_Matrix_a.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_b.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_c.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_d.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_e.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_f.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_g.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_h.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_i.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_j.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_k.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_l.png +0 -0
- evaluation_results.csv +133 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_a.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_b.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_c.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_d.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_e.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_f.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_g.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_h.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_i.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_j.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_k.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_l.png +0 -0
- roc_curves/ViT_ROC_a.png +0 -0
- roc_curves/ViT_ROC_b.png +0 -0
- roc_curves/ViT_ROC_c.png +0 -0
- roc_curves/ViT_ROC_d.png +0 -0
- roc_curves/ViT_ROC_e.png +0 -0
- roc_curves/ViT_ROC_f.png +0 -0
- roc_curves/ViT_ROC_g.png +0 -0
- roc_curves/ViT_ROC_h.png +0 -0
- roc_curves/ViT_ROC_i.png +0 -0
- roc_curves/ViT_ROC_j.png +0 -0
- roc_curves/ViT_ROC_k.png +0 -0
- roc_curves/ViT_ROC_l.png +0 -0
- training_curves/ViT_accuracy.png +0 -0
- training_curves/ViT_auc.png +0 -0
- training_curves/ViT_combined_metrics.png +3 -0
- training_curves/ViT_f1.png +0 -0
- training_curves/ViT_loss.png +0 -0
- training_curves/ViT_metrics.csv +36 -0
- training_metrics.csv +36 -0
- training_notebook_c2.ipynb +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
training_curves/ViT_combined_metrics.png filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
training_notebook_c2.ipynb filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,166 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- image-classification
|
| 5 |
+
- pytorch
|
| 6 |
+
- timm
|
| 7 |
+
- vit
|
| 8 |
+
- vision-transformer
|
| 9 |
+
- transformer
|
| 10 |
+
- gravitational-lensing
|
| 11 |
+
- strong-lensing
|
| 12 |
+
- astronomy
|
| 13 |
+
- astrophysics
|
| 14 |
+
datasets:
|
| 15 |
+
- parlange/gravit-c21-j24
|
| 16 |
+
metrics:
|
| 17 |
+
- accuracy
|
| 18 |
+
- auc
|
| 19 |
+
- f1
|
| 20 |
+
paper:
|
| 21 |
+
- title: "GraViT: A Gravitational Lens Discovery Toolkit with Vision Transformers"
|
| 22 |
+
url: "https://arxiv.org/abs/2509.00226"
|
| 23 |
+
authors: "Parlange et al."
|
| 24 |
+
model-index:
|
| 25 |
+
- name: ViT-c2
|
| 26 |
+
results:
|
| 27 |
+
- task:
|
| 28 |
+
type: image-classification
|
| 29 |
+
name: Strong Gravitational Lens Discovery
|
| 30 |
+
dataset:
|
| 31 |
+
type: common-test-sample
|
| 32 |
+
name: Common Test Sample (More et al. 2024)
|
| 33 |
+
metrics:
|
| 34 |
+
- type: accuracy
|
| 35 |
+
value: 0.8590
|
| 36 |
+
name: Average Accuracy
|
| 37 |
+
- type: auc
|
| 38 |
+
value: 0.8885
|
| 39 |
+
name: Average AUC-ROC
|
| 40 |
+
- type: f1
|
| 41 |
+
value: 0.6412
|
| 42 |
+
name: Average F1-Score
|
| 43 |
+
---
|
| 44 |
+
|
| 45 |
+
# 🌌 vit-gravit-c2
|
| 46 |
+
|
| 47 |
+
🔭 This model is part of **GraViT**: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery
|
| 48 |
+
|
| 49 |
+
🔗 **GitHub Repository**: [https://github.com/parlange/gravit](https://github.com/parlange/gravit)
|
| 50 |
+
|
| 51 |
+
## 🛰️ Model Details
|
| 52 |
+
|
| 53 |
+
- **🤖 Model Type**: ViT
|
| 54 |
+
- **🧪 Experiment**: C2 - C21+J24-half
|
| 55 |
+
- **🌌 Dataset**: C21+J24
|
| 56 |
+
- **🪐 Fine-tuning Strategy**: half
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
## 💻 Quick Start
|
| 61 |
+
|
| 62 |
+
```python
|
| 63 |
+
import torch
|
| 64 |
+
import timm
|
| 65 |
+
|
| 66 |
+
# Load the model directly from the Hub
|
| 67 |
+
model = timm.create_model(
|
| 68 |
+
'hf-hub:parlange/vit-gravit-c2',
|
| 69 |
+
pretrained=True
|
| 70 |
+
)
|
| 71 |
+
model.eval()
|
| 72 |
+
|
| 73 |
+
# Example inference
|
| 74 |
+
dummy_input = torch.randn(1, 3, 224, 224)
|
| 75 |
+
with torch.no_grad():
|
| 76 |
+
output = model(dummy_input)
|
| 77 |
+
predictions = torch.softmax(output, dim=1)
|
| 78 |
+
print(f"Lens probability: {predictions[0][1]:.4f}")
|
| 79 |
+
```
|
| 80 |
+
|
| 81 |
+
## ⚡️ Training Configuration
|
| 82 |
+
|
| 83 |
+
**Training Dataset:** C21+J24 (Cañameras et al. 2021 + Jaelani et al. 2024)
|
| 84 |
+
**Fine-tuning Strategy:** half
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
| 🔧 Parameter | 📝 Value |
|
| 88 |
+
|--------------|----------|
|
| 89 |
+
| Batch Size | 192 |
|
| 90 |
+
| Learning Rate | AdamW with ReduceLROnPlateau |
|
| 91 |
+
| Epochs | 100 |
|
| 92 |
+
| Patience | 10 |
|
| 93 |
+
| Optimizer | AdamW |
|
| 94 |
+
| Scheduler | ReduceLROnPlateau |
|
| 95 |
+
| Image Size | 224x224 |
|
| 96 |
+
| Fine Tune Mode | half |
|
| 97 |
+
| Stochastic Depth Probability | 0.1 |
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
## 📈 Training Curves
|
| 101 |
+
|
| 102 |
+

|
| 103 |
+
|
| 104 |
+
|
| 105 |
+
## 🏁 Final Epoch Training Metrics
|
| 106 |
+
|
| 107 |
+
| Metric | Training | Validation |
|
| 108 |
+
|:---------:|:-----------:|:-------------:|
|
| 109 |
+
| 📉 Loss | 0.0277 | 0.0615 |
|
| 110 |
+
| 🎯 Accuracy | 0.9898 | 0.9832 |
|
| 111 |
+
| 📊 AUC-ROC | 0.9995 | 0.9986 |
|
| 112 |
+
| ⚖️ F1 Score | 0.9898 | 0.9833 |
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
## ☑️ Evaluation Results
|
| 116 |
+
|
| 117 |
+
### ROC Curves and Confusion Matrices
|
| 118 |
+
|
| 119 |
+
Performance across all test datasets (a through l) in the Common Test Sample (More et al. 2024):
|
| 120 |
+
|
| 121 |
+

|
| 122 |
+

|
| 123 |
+

|
| 124 |
+

|
| 125 |
+

|
| 126 |
+

|
| 127 |
+

|
| 128 |
+

|
| 129 |
+

|
| 130 |
+

|
| 131 |
+

|
| 132 |
+

|
| 133 |
+
|
| 134 |
+
### 📋 Performance Summary
|
| 135 |
+
|
| 136 |
+
Average performance across 12 test datasets from the Common Test Sample (More et al. 2024):
|
| 137 |
+
|
| 138 |
+
| Metric | Value |
|
| 139 |
+
|-----------|----------|
|
| 140 |
+
| 🎯 Average Accuracy | 0.8590 |
|
| 141 |
+
| 📈 Average AUC-ROC | 0.8885 |
|
| 142 |
+
| ⚖️ Average F1-Score | 0.6412 |
|
| 143 |
+
|
| 144 |
+
|
| 145 |
+
## 📘 Citation
|
| 146 |
+
|
| 147 |
+
If you use this model in your research, please cite:
|
| 148 |
+
|
| 149 |
+
```bibtex
|
| 150 |
+
@misc{parlange2025gravit,
|
| 151 |
+
title={GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery},
|
| 152 |
+
author={René Parlange and Juan C. Cuevas-Tello and Octavio Valenzuela and Omar de J. Cabrera-Rosas and Tomás Verdugo and Anupreeta More and Anton T. Jaelani},
|
| 153 |
+
year={2025},
|
| 154 |
+
eprint={2509.00226},
|
| 155 |
+
archivePrefix={arXiv},
|
| 156 |
+
primaryClass={cs.CV},
|
| 157 |
+
url={https://arxiv.org/abs/2509.00226},
|
| 158 |
+
}
|
| 159 |
+
```
|
| 160 |
+
|
| 161 |
+
---
|
| 162 |
+
|
| 163 |
+
|
| 164 |
+
## Model Card Contact
|
| 165 |
+
|
| 166 |
+
For questions about this model, please contact the author through: https://github.com/parlange/
|
config.json
ADDED
|
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architecture": "vit_base_patch16_224",
|
| 3 |
+
"num_classes": 2,
|
| 4 |
+
"num_features": 768,
|
| 5 |
+
"global_pool": "token",
|
| 6 |
+
"crop_pct": 0.875,
|
| 7 |
+
"interpolation": "bicubic",
|
| 8 |
+
"mean": [
|
| 9 |
+
0.485,
|
| 10 |
+
0.456,
|
| 11 |
+
0.406
|
| 12 |
+
],
|
| 13 |
+
"std": [
|
| 14 |
+
0.229,
|
| 15 |
+
0.224,
|
| 16 |
+
0.225
|
| 17 |
+
],
|
| 18 |
+
"first_conv": "patch_embed.proj",
|
| 19 |
+
"classifier": "head",
|
| 20 |
+
"input_size": [
|
| 21 |
+
3,
|
| 22 |
+
224,
|
| 23 |
+
224
|
| 24 |
+
],
|
| 25 |
+
"pool_size": [
|
| 26 |
+
7,
|
| 27 |
+
7
|
| 28 |
+
],
|
| 29 |
+
"pretrained_cfg": {
|
| 30 |
+
"tag": "gravit_c2",
|
| 31 |
+
"custom_load": false,
|
| 32 |
+
"input_size": [
|
| 33 |
+
3,
|
| 34 |
+
224,
|
| 35 |
+
224
|
| 36 |
+
],
|
| 37 |
+
"fixed_input_size": true,
|
| 38 |
+
"interpolation": "bicubic",
|
| 39 |
+
"crop_pct": 0.875,
|
| 40 |
+
"crop_mode": "center",
|
| 41 |
+
"mean": [
|
| 42 |
+
0.485,
|
| 43 |
+
0.456,
|
| 44 |
+
0.406
|
| 45 |
+
],
|
| 46 |
+
"std": [
|
| 47 |
+
0.229,
|
| 48 |
+
0.224,
|
| 49 |
+
0.225
|
| 50 |
+
],
|
| 51 |
+
"num_classes": 2,
|
| 52 |
+
"pool_size": [
|
| 53 |
+
7,
|
| 54 |
+
7
|
| 55 |
+
],
|
| 56 |
+
"first_conv": "patch_embed.proj",
|
| 57 |
+
"classifier": "head"
|
| 58 |
+
},
|
| 59 |
+
"model_name": "vit_gravit_c2",
|
| 60 |
+
"experiment": "c2",
|
| 61 |
+
"training_strategy": "half",
|
| 62 |
+
"dataset": "C21+J24",
|
| 63 |
+
"hyperparameters": {
|
| 64 |
+
"batch_size": "192",
|
| 65 |
+
"learning_rate": "AdamW with ReduceLROnPlateau",
|
| 66 |
+
"epochs": "100",
|
| 67 |
+
"patience": "10",
|
| 68 |
+
"optimizer": "AdamW",
|
| 69 |
+
"scheduler": "ReduceLROnPlateau",
|
| 70 |
+
"image_size": "224x224",
|
| 71 |
+
"fine_tune_mode": "half",
|
| 72 |
+
"stochastic_depth_probability": "0.1"
|
| 73 |
+
},
|
| 74 |
+
"hf_hub_id": "parlange/vit-gravit-c2",
|
| 75 |
+
"license": "apache-2.0"
|
| 76 |
+
}
|
confusion_matrices/ViT_Confusion_Matrix_a.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_b.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_c.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_d.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_e.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_f.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_g.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_h.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_i.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_j.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_k.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_l.png
ADDED
|
evaluation_results.csv
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Model,Dataset,Loss,Accuracy,AUCROC,F1
|
| 2 |
+
ViT,a,0.35447569389258865,0.8949115044247787,0.9020846228498507,0.7480106100795756
|
| 3 |
+
ViT,b,0.2228443425890036,0.9264382269726501,0.9263609576427256,0.5465116279069767
|
| 4 |
+
ViT,c,0.46229862214933587,0.8349575605155611,0.8684438305709024,0.34944237918215615
|
| 5 |
+
ViT,d,0.11673173789463115,0.9537881169443572,0.9704917127071824,0.6573426573426573
|
| 6 |
+
ViT,e,0.3652098159562351,0.8825466520307355,0.920767426019829,0.7249357326478149
|
| 7 |
+
ViT,f,0.24608832064126923,0.9108519842016175,0.9203345361214339,0.22926829268292684
|
| 8 |
+
ViT,g,0.10483654439449311,0.9635,0.997473,0.9644999189495866
|
| 9 |
+
ViT,h,0.23178722894191742,0.915,0.9939590555555555,0.9210526315789473
|
| 10 |
+
ViT,i,0.04857917896906535,0.978,0.9990572222222223,0.9782966129562644
|
| 11 |
+
ViT,j,2.494326035181681,0.6106666666666667,0.5831323333333334,0.42349457058242845
|
| 12 |
+
ViT,k,2.4380686638752618,0.6251666666666666,0.7805802777777779,0.43278688524590164
|
| 13 |
+
ViT,l,1.0272723743838732,0.8127329565949261,0.7993805230717175,0.7184308053873272
|
| 14 |
+
MLP-Mixer,a,1.230455079964832,0.6227876106194691,0.8958911227772556,0.49028400597907323
|
| 15 |
+
MLP-Mixer,b,1.0728926989350893,0.7004086765168186,0.9182900552486188,0.25604996096799376
|
| 16 |
+
MLP-Mixer,c,1.374837134027586,0.5576862621817039,0.8979152854511969,0.18904899135446687
|
| 17 |
+
MLP-Mixer,d,0.09552026474693218,0.9603898145237346,0.9868913443830571,0.7224669603524229
|
| 18 |
+
MLP-Mixer,e,0.9593323631422711,0.7069154774972558,0.9188677817301143,0.5512605042016807
|
| 19 |
+
MLP-Mixer,f,0.9257462782946794,0.7154410381794245,0.9306221006103087,0.09779367918902802
|
| 20 |
+
MLP-Mixer,g,0.5643243643840155,0.8425,0.991425611111111,0.8635773061931572
|
| 21 |
+
MLP-Mixer,h,0.7244052359660467,0.7668333333333334,0.9891666111111111,0.8104592873594364
|
| 22 |
+
MLP-Mixer,i,0.04615406060218811,0.9803333333333333,0.9994367777777778,0.980655737704918
|
| 23 |
+
MLP-Mixer,j,3.0292422666549683,0.45216666666666666,0.392282,0.28309705561613957
|
| 24 |
+
MLP-Mixer,k,2.5110719747940697,0.59,0.7661271111111111,0.3453964874933475
|
| 25 |
+
MLP-Mixer,l,1.4846716919555334,0.6762053625105207,0.7295511702036557,0.5855010004617516
|
| 26 |
+
CvT,a,0.7465745627352621,0.6493362831858407,0.7317079694031161,0.4389380530973451
|
| 27 |
+
CvT,b,0.7336456650122649,0.6765168186104998,0.7552670349907918,0.1942051683633516
|
| 28 |
+
CvT,c,0.8642418710588097,0.5919522162841874,0.6964806629834255,0.16041397153945666
|
| 29 |
+
CvT,d,0.06205783033066015,0.9761081420936812,0.9876427255985267,0.7654320987654321
|
| 30 |
+
CvT,e,0.6019917449757506,0.7178924259055982,0.7936123514720351,0.4910891089108911
|
| 31 |
+
CvT,f,0.5685286294680824,0.7414895617829603,0.8061353821076506,0.08274941608274941
|
| 32 |
+
CvT,g,0.4509977758725484,0.8055,0.9201512777777776,0.8277999114652501
|
| 33 |
+
CvT,h,0.5202355206807454,0.7606666666666667,0.9072719444444444,0.7961964235026966
|
| 34 |
+
CvT,i,0.09494428576032321,0.9643333333333334,0.9977035555555557,0.9632554945054945
|
| 35 |
+
CvT,j,2.988422914981842,0.3456666666666667,0.14668444444444442,0.022896963663514187
|
| 36 |
+
CvT,k,2.6323694267769655,0.5045,0.6181494444444444,0.0300163132137031
|
| 37 |
+
CvT,l,1.337245315202257,0.645425033064807,0.6032419706344807,0.5021944632005402
|
| 38 |
+
Swin,a,0.47572549887463056,0.8407079646017699,0.905882487792577,0.6742081447963801
|
| 39 |
+
Swin,b,0.24361524523634911,0.9163784973278843,0.9362615101289135,0.5283687943262412
|
| 40 |
+
Swin,c,0.4370936370240709,0.8535051870480981,0.9087605893186003,0.3900523560209424
|
| 41 |
+
Swin,d,0.038348094671021904,0.9880540710468406,0.9911620626151013,0.8869047619047619
|
| 42 |
+
Swin,e,0.3579506372581067,0.8781558726673985,0.9260273972602739,0.7286063569682152
|
| 43 |
+
Swin,f,0.24650774364781286,0.9156479217603912,0.9413092437445593,0.24937238493723848
|
| 44 |
+
Swin,g,0.11494702147444089,0.9593333333333334,0.9989898888888888,0.9607969151670951
|
| 45 |
+
Swin,h,0.2175228010714054,0.926,0.9979807777777777,0.9308841843088418
|
| 46 |
+
Swin,i,0.006121216081082821,0.9973333333333333,0.9999798888888889,0.9973315543695798
|
| 47 |
+
Swin,j,2.5422419211069744,0.5825,0.4893003333333333,0.3679031037093111
|
| 48 |
+
Swin,k,2.433416116627554,0.6205,0.7913794999999999,0.39036144578313253
|
| 49 |
+
Swin,l,1.035569912688268,0.8089455332451605,0.7797953948083542,0.7088143668682426
|
| 50 |
+
CaiT,a,0.3509529214517205,0.9081858407079646,0.8966973093999068,0.7726027397260274
|
| 51 |
+
CaiT,b,0.1907231829655279,0.9380697893744105,0.9234548802946593,0.5887265135699373
|
| 52 |
+
CaiT,c,0.3048490960337163,0.90883370009431,0.8791160220994475,0.493006993006993
|
| 53 |
+
CaiT,d,0.06549901952829443,0.9849104055328513,0.969243093922652,0.8545454545454545
|
| 54 |
+
CaiT,e,0.31167979835318943,0.9187705817782656,0.9264058124574283,0.7921348314606742
|
| 55 |
+
CaiT,f,0.1541684599891403,0.9499717886025955,0.9222261921687871,0.3464373464373464
|
| 56 |
+
CaiT,g,0.07805611325552066,0.9708333333333333,0.9986172777777778,0.9714937286202965
|
| 57 |
+
CaiT,h,0.13856186520308256,0.9553333333333334,0.997130611111111,0.9569961489088575
|
| 58 |
+
CaiT,i,0.011666435472667217,0.9956666666666667,0.9999013333333333,0.9956594323873121
|
| 59 |
+
CaiT,j,1.8389671653707822,0.6116666666666667,0.7423962222222222,0.4151606425702811
|
| 60 |
+
CaiT,k,1.7725774958133698,0.6365,0.8888650555555555,0.4312907431551499
|
| 61 |
+
CaiT,l,0.7395369254032035,0.8362991463268006,0.8693810723675515,0.7436693965922997
|
| 62 |
+
DeiT,a,0.48058320357736234,0.8263274336283186,0.8941450218931248,0.6594360086767896
|
| 63 |
+
DeiT,b,0.23002449519573911,0.9251807607670544,0.9313581952117864,0.5608856088560885
|
| 64 |
+
DeiT,c,0.49494195908204974,0.8154668343288274,0.8907605893186004,0.34118967452300786
|
| 65 |
+
DeiT,d,0.05036040664735698,0.9849104055328513,0.9769023941068141,0.8636363636363636
|
| 66 |
+
DeiT,e,0.338863200106291,0.8792535675082327,0.9161961704382048,0.7342995169082126
|
| 67 |
+
DeiT,f,0.26403015722496653,0.9037050968591311,0.9291450866890099,0.2289156626506024
|
| 68 |
+
DeiT,g,0.10851164469867945,0.9641666666666666,0.9990410000000001,0.9653393519264872
|
| 69 |
+
DeiT,h,0.2489620513096452,0.906,0.9981344444444444,0.9139194139194139
|
| 70 |
+
DeiT,i,0.013259729760388533,0.9958333333333333,0.9998315555555556,0.9958423415932147
|
| 71 |
+
DeiT,j,1.2026229511300723,0.7143333333333334,0.7246498888888889,0.6356292517006803
|
| 72 |
+
DeiT,k,1.1073710439900557,0.746,0.8698901111111111,0.6623836951705804
|
| 73 |
+
DeiT,l,0.5658274294531473,0.8476012985451485,0.867833726587774,0.7854785478547854
|
| 74 |
+
DeiT3,a,0.39277621998196155,0.8661504424778761,0.9195532732705195,0.7125890736342043
|
| 75 |
+
DeiT3,b,0.338128161960636,0.8824269097767997,0.9331012891344382,0.44510385756676557
|
| 76 |
+
DeiT3,c,0.323060417608134,0.8883998742533794,0.922292817679558,0.4580152671755725
|
| 77 |
+
DeiT3,d,0.12409640010358478,0.9553599497013517,0.9608121546961326,0.6787330316742082
|
| 78 |
+
DeiT3,e,0.24973662732461413,0.9209659714599341,0.9483084840687203,0.8064516129032258
|
| 79 |
+
DeiT3,f,0.2540075041596123,0.9116042881324055,0.9380772021883802,0.24193548387096775
|
| 80 |
+
DeiT3,g,0.1656125110021482,0.9416666666666667,0.9990236666666666,0.944760101010101
|
| 81 |
+
DeiT3,h,0.15762409150910875,0.9448333333333333,0.9990646111111111,0.9476017096723128
|
| 82 |
+
DeiT3,i,0.05214000094247361,0.9803333333333333,0.9997376666666667,0.9806684141546527
|
| 83 |
+
DeiT3,j,1.1591287109454473,0.696,0.7744774999999999,0.6248457424928013
|
| 84 |
+
DeiT3,k,1.0456561943689981,0.7346666666666667,0.845634,0.6561555075593952
|
| 85 |
+
DeiT3,l,0.5223108836063022,0.854033906456655,0.8898184372191467,0.7933968686181075
|
| 86 |
+
Twins_SVT,a,0.4211153812640536,0.8307522123893806,0.8825833123189902,0.6433566433566433
|
| 87 |
+
Twins_SVT,b,0.3625493723054758,0.8550770198050928,0.8962191528545118,0.37449118046132973
|
| 88 |
+
Twins_SVT,c,0.47319920195681764,0.7868594781515247,0.8548139963167587,0.2893081761006289
|
| 89 |
+
Twins_SVT,d,0.1203458983801289,0.9783087079534738,0.9818324125230202,0.8
|
| 90 |
+
Twins_SVT,e,0.5213294555274637,0.7486278814489572,0.8316203738742148,0.5465346534653466
|
| 91 |
+
Twins_SVT,f,0.3335461875583885,0.8666541282678202,0.9034523383543173,0.16292798110979928
|
| 92 |
+
Twins_SVT,g,0.2639119902451833,0.9085,0.9744078888888889,0.912676952441546
|
| 93 |
+
Twins_SVT,h,0.32257486327489215,0.8723333333333333,0.9662636666666669,0.8822263222632226
|
| 94 |
+
Twins_SVT,i,0.13550377811988196,0.9738333333333333,0.9972788888888889,0.9733672603901612
|
| 95 |
+
Twins_SVT,j,1.2430085968176523,0.49,0.43771377777777776,0.1896186440677966
|
| 96 |
+
Twins_SVT,k,1.1146003757913907,0.5553333333333333,0.7234002222222222,0.2115839243498818
|
| 97 |
+
Twins_SVT,l,0.6286477774643219,0.7480461704941685,0.7275090480198628,0.6162439337057046
|
| 98 |
+
Twins_PCPVT,a,0.45601994748664115,0.7699115044247787,0.8394007473464615,0.5458515283842795
|
| 99 |
+
Twins_PCPVT,b,0.3125818614145001,0.8773970449544168,0.9010699815837937,0.390625
|
| 100 |
+
Twins_PCPVT,c,0.5049686531944119,0.7500785916378497,0.8135911602209945,0.23923444976076555
|
| 101 |
+
Twins_PCPVT,d,0.3149096430453517,0.8918579063187677,0.9015690607734806,0.4208754208754209
|
| 102 |
+
Twins_PCPVT,e,0.42039827045572575,0.8079034028540066,0.8655339438431847,0.5882352941176471
|
| 103 |
+
Twins_PCPVT,f,0.3770137148085496,0.8412638706037239,0.8693597175042401,0.12899896800825594
|
| 104 |
+
Twins_PCPVT,g,0.2785677030881246,0.9015,0.9626754444444443,0.9027480664801711
|
| 105 |
+
Twins_PCPVT,h,0.3805647597312927,0.834,0.928301,0.8463437210737427
|
| 106 |
+
Twins_PCPVT,i,0.2798018006483714,0.9091666666666667,0.9656723333333334,0.9096335599403084
|
| 107 |
+
Twins_PCPVT,j,0.614702238559723,0.6835,0.7995154444444446,0.6018033130635353
|
| 108 |
+
Twins_PCPVT,k,0.6159363424777985,0.6911666666666667,0.7903985,0.6076646199449502
|
| 109 |
+
Twins_PCPVT,l,0.45535326129802217,0.7889864133702056,0.8498913163479216,0.7103004291845494
|
| 110 |
+
PiT,a,0.3937257931823224,0.8296460176991151,0.8874127904755356,0.641860465116279
|
| 111 |
+
PiT,b,0.2796248870145521,0.8777114115058158,0.91848802946593,0.4150375939849624
|
| 112 |
+
PiT,c,0.5313189482209218,0.7613957874882112,0.8498581952117863,0.26666666666666666
|
| 113 |
+
PiT,d,0.049343678185640734,0.9798805407104684,0.9911620626151012,0.8117647058823529
|
| 114 |
+
PiT,e,0.3259278782832505,0.8518111964873765,0.9145841216983274,0.6715328467153284
|
| 115 |
+
PiT,f,0.2841162405192056,0.8750235094978371,0.9172267022129574,0.17196261682242991
|
| 116 |
+
PiT,g,0.1590204114516576,0.9338333333333333,0.9916004444444445,0.9369340746624305
|
| 117 |
+
PiT,h,0.2924602138201396,0.8721666666666666,0.981646111111111,0.8849212303075769
|
| 118 |
+
PiT,i,0.03693298858900865,0.988,0.999485,0.9879396984924623
|
| 119 |
+
PiT,j,2.9977854507366817,0.461,0.277717,0.06477732793522267
|
| 120 |
+
PiT,k,2.8756980224698783,0.5151666666666667,0.7229978888888889,0.07149696776252792
|
| 121 |
+
PiT,l,1.2244331041709067,0.7434170975111218,0.6790239785353327,0.599849990624414
|
| 122 |
+
Ensemble,a,,0.9070796460176991,0.941851401847734,0.79
|
| 123 |
+
Ensemble,b,,0.9374410562716127,0.9600349907918969,0.6135922330097088
|
| 124 |
+
Ensemble,c,,0.895001571832757,0.9307624309392265,0.48615384615384616
|
| 125 |
+
Ensemble,d,,0.9911977365608299,0.9944677716390424,0.9186046511627907
|
| 126 |
+
Ensemble,e,,0.9264544456641054,0.955384848255506,0.825065274151436
|
| 127 |
+
Ensemble,f,,0.941696445363927,0.9599335198386041,0.33760683760683763
|
| 128 |
+
Ensemble,g,,0.9701666666666666,0.9990522222222222,0.9710027539283979
|
| 129 |
+
Ensemble,h,,0.9476666666666667,0.9979163333333333,0.9502219403931516
|
| 130 |
+
Ensemble,i,,0.9986666666666667,0.9999886666666667,0.9986671109630123
|
| 131 |
+
Ensemble,j,,0.5698333333333333,0.6426453333333333,0.31556616282153277
|
| 132 |
+
Ensemble,k,,0.5983333333333334,0.8897323333333333,0.33055555555555555
|
| 133 |
+
Ensemble,l,,0.8179632078874595,0.832089495815299,0.712386018237082
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:714b63b39bcfecbd1659a435651c91e7d7414a6a46f6a03cbb2d01e575fd056b
|
| 3 |
+
size 343214864
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c27e9a389eabd73858118d5a23fed1f59fb1d8906a9e4b31983fe2cef1a1b1c9
|
| 3 |
+
size 343259038
|
roc_confusion_matrix/ViT_roc_confusion_matrix_a.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_b.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_c.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_d.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_e.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_f.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_g.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_h.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_i.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_j.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_k.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_l.png
ADDED
|
roc_curves/ViT_ROC_a.png
ADDED
|
roc_curves/ViT_ROC_b.png
ADDED
|
roc_curves/ViT_ROC_c.png
ADDED
|
roc_curves/ViT_ROC_d.png
ADDED
|
roc_curves/ViT_ROC_e.png
ADDED
|
roc_curves/ViT_ROC_f.png
ADDED
|
roc_curves/ViT_ROC_g.png
ADDED
|
roc_curves/ViT_ROC_h.png
ADDED
|
roc_curves/ViT_ROC_i.png
ADDED
|
roc_curves/ViT_ROC_j.png
ADDED
|
roc_curves/ViT_ROC_k.png
ADDED
|
roc_curves/ViT_ROC_l.png
ADDED
|
training_curves/ViT_accuracy.png
ADDED
|
training_curves/ViT_auc.png
ADDED
|
training_curves/ViT_combined_metrics.png
ADDED
|
Git LFS Details
|
training_curves/ViT_f1.png
ADDED
|
training_curves/ViT_loss.png
ADDED
|
training_curves/ViT_metrics.csv
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.2523881019792201,0.15659080327925634,0.89101994048637,0.9475218658892128,0.9614815034763909,0.9904631573579037,0.8906224522197335,0.948644793152639
|
| 3 |
+
2,0.12423779873559851,0.10793833425893033,0.9519957587987824,0.956997084548105,0.9898794034587902,0.9928324932638611,0.9516401350699469,0.9567765567765568
|
| 4 |
+
3,0.10529630348674004,0.08779277159416989,0.9600591715976331,0.9693877551020408,0.9925351758495597,0.9951189555372336,0.959792030713345,0.9690721649484536
|
| 5 |
+
4,0.09480468541721541,0.07493716165553724,0.9643602284776139,0.9715743440233237,0.9939349845344156,0.9964672457904444,0.9641653483733407,0.971677559912854
|
| 6 |
+
5,0.08620487234102249,0.07931267747617497,0.9671648937989534,0.9723032069970845,0.9949847680779575,0.9962207498576274,0.967002938800763,0.9722222222222222
|
| 7 |
+
6,0.0821567311519898,0.07912938319602791,0.969285494407771,0.9730320699708455,0.9954400735733278,0.9963546226487263,0.9691324075347174,0.9727740986019132
|
| 8 |
+
7,0.07842890813014797,0.06848325287114079,0.9706023189793754,0.9744897959183674,0.9958517618127383,0.9968879888481839,0.9704715279567122,0.9745454545454545
|
| 9 |
+
8,0.07323522930321691,0.07025298998357168,0.9722440742894278,0.9737609329446064,0.9964020427924299,0.9965926187217911,0.9721205874774542,0.9737609329446064
|
| 10 |
+
9,0.07074919585322677,0.07600417623497306,0.973654957759004,0.9723032069970845,0.9966524143202075,0.9969506753138573,0.9735570527399906,0.9724238026124818
|
| 11 |
+
10,0.06907813417501553,0.06526092207873832,0.9740397441597975,0.978134110787172,0.9967897613044078,0.9971174850614964,0.9739470703326125,0.9782293178519593
|
| 12 |
+
11,0.06508755874415982,0.060960897464446356,0.9758439648390738,0.9774052478134111,0.9970726140319375,0.9979281591853735,0.9757516973811833,0.9773557341124909
|
| 13 |
+
12,0.062282602138362175,0.06926894213993418,0.9762971577111195,0.9744897959183674,0.9973810927180742,0.9967870530136254,0.976210909340565,0.9743589743589743
|
| 14 |
+
13,0.060860307139227,0.06267463378934054,0.977434415295687,0.9795918367346939,0.9974749183749438,0.9977199126214417,0.9773540542164022,0.9797395079594791
|
| 15 |
+
14,0.05755463975885633,0.07566890224770971,0.9785118172179088,0.9766763848396501,0.9977389300470996,0.9971812340096389,0.978445650961926,0.9764359351988218
|
| 16 |
+
15,0.057541033925899064,0.04975939745154047,0.9786058761158806,0.9825072886297376,0.9977538220893306,0.9986517097467891,0.9785357651459259,0.9823788546255506
|
| 17 |
+
16,0.05634996499272484,0.0617915925518243,0.9791274754591784,0.978134110787172,0.997829316750365,0.9969241132521313,0.9790527842376706,0.9780701754385965
|
| 18 |
+
17,0.05454238486925483,0.05333676516911987,0.9799141498785785,0.9810495626822158,0.997946748161562,0.9985412115700091,0.979850919103456,0.9811866859623734
|
| 19 |
+
18,0.053665978950810676,0.08533909733871503,0.9801450217190546,0.967930029154519,0.998023593281266,0.9963652474734167,0.9800902028707149,0.9683908045977011
|
| 20 |
+
19,0.051960351455705815,0.05903876217695784,0.9809060437117352,0.9861516034985423,0.998134591332559,0.9976434138836708,0.9808541468391766,0.9860601614086574
|
| 21 |
+
20,0.052026702721111096,0.08011818320949293,0.9806922734890721,0.9752186588921283,0.9981493410904061,0.996891176295591,0.9806309938410335,0.9751461988304093
|
| 22 |
+
21,0.05115936366098929,0.0657448608634597,0.9809231453295482,0.9810495626822158,0.9981968896653561,0.9971791090447008,0.9808647322692146,0.9811046511627907
|
| 23 |
+
22,0.039293639540274,0.058326971969016046,0.9854892772856312,0.9839650145772595,0.9989317592618322,0.9985231493680354,0.9854545766227533,0.9839883551673945
|
| 24 |
+
23,0.033775414704451334,0.05694627253414721,0.987840749734925,0.9817784256559767,0.9992104630761648,0.9986453348519749,0.9878144923562076,0.9818181818181818
|
| 25 |
+
24,0.033230631860863655,0.05545937554481773,0.9876526319389815,0.9832361516034985,0.9992240350761206,0.9987367083443124,0.9876325390979633,0.9832239241429613
|
| 26 |
+
25,0.03237407778996703,0.0587208214543119,0.9882426377535315,0.9810495626822158,0.9992709570440971,0.9986835842208603,0.9882171472642358,0.9810771470160117
|
| 27 |
+
26,0.03220497033339533,0.06503236680848215,0.9882169853268119,0.9832361516034985,0.9992812725888158,0.9983733393399009,0.9881963955321045,0.9832483612527313
|
| 28 |
+
27,0.03042350235752952,0.06034932669480236,0.9890207613640251,0.9839650145772595,0.999364067769463,0.9985539613596376,0.9890068493150685,0.9839181286549707
|
| 29 |
+
28,0.029133988782805785,0.06084428801430042,0.9891319218798098,0.9832361516034985,0.9994052302176598,0.998483837516681,0.9891105989598952,0.9832727272727273
|
| 30 |
+
29,0.02860731827776991,0.06154062449063226,0.9898416390190512,0.9832361516034985,0.9994026410220773,0.9985199619206284,0.9898245854460738,0.9832970225127088
|
| 31 |
+
30,0.02884942300385336,0.06216380139134021,0.989773232547799,0.9825072886297376,0.9994035903645443,0.9985242118505044,0.9897565905548228,0.9825834542815675
|
| 32 |
+
31,0.029171249110760186,0.061966189061109594,0.9894739542360708,0.9832361516034985,0.9994051022640773,0.9985072121309998,0.9894542058957072,0.9832483612527313
|
| 33 |
+
32,0.028459394045261276,0.06270270375467012,0.9896449704142012,0.9825072886297376,0.9994358494368771,0.9985316492277877,0.9896278531968652,0.982532751091703
|
| 34 |
+
33,0.029123750241781602,0.06183103346550951,0.9890293121729316,0.9832361516034985,0.9994153125212766,0.9985465239823543,0.9890092945560457,0.9832727272727273
|
| 35 |
+
34,0.028456661750446508,0.06152658220943954,0.9893371412935664,0.9832361516034985,0.9994328933435357,0.998566711149266,0.9893239043517718,0.9832727272727273
|
| 36 |
+
35,0.02765608663334115,0.061544772590518694,0.9898245374012381,0.9832361516034985,0.9994650532701209,0.998566711149266,0.9898081534772182,0.9832727272727273
|
training_metrics.csv
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.2523881019792201,0.15659080327925634,0.89101994048637,0.9475218658892128,0.9614815034763909,0.9904631573579037,0.8906224522197335,0.948644793152639
|
| 3 |
+
2,0.12423779873559851,0.10793833425893033,0.9519957587987824,0.956997084548105,0.9898794034587902,0.9928324932638611,0.9516401350699469,0.9567765567765568
|
| 4 |
+
3,0.10529630348674004,0.08779277159416989,0.9600591715976331,0.9693877551020408,0.9925351758495597,0.9951189555372336,0.959792030713345,0.9690721649484536
|
| 5 |
+
4,0.09480468541721541,0.07493716165553724,0.9643602284776139,0.9715743440233237,0.9939349845344156,0.9964672457904444,0.9641653483733407,0.971677559912854
|
| 6 |
+
5,0.08620487234102249,0.07931267747617497,0.9671648937989534,0.9723032069970845,0.9949847680779575,0.9962207498576274,0.967002938800763,0.9722222222222222
|
| 7 |
+
6,0.0821567311519898,0.07912938319602791,0.969285494407771,0.9730320699708455,0.9954400735733278,0.9963546226487263,0.9691324075347174,0.9727740986019132
|
| 8 |
+
7,0.07842890813014797,0.06848325287114079,0.9706023189793754,0.9744897959183674,0.9958517618127383,0.9968879888481839,0.9704715279567122,0.9745454545454545
|
| 9 |
+
8,0.07323522930321691,0.07025298998357168,0.9722440742894278,0.9737609329446064,0.9964020427924299,0.9965926187217911,0.9721205874774542,0.9737609329446064
|
| 10 |
+
9,0.07074919585322677,0.07600417623497306,0.973654957759004,0.9723032069970845,0.9966524143202075,0.9969506753138573,0.9735570527399906,0.9724238026124818
|
| 11 |
+
10,0.06907813417501553,0.06526092207873832,0.9740397441597975,0.978134110787172,0.9967897613044078,0.9971174850614964,0.9739470703326125,0.9782293178519593
|
| 12 |
+
11,0.06508755874415982,0.060960897464446356,0.9758439648390738,0.9774052478134111,0.9970726140319375,0.9979281591853735,0.9757516973811833,0.9773557341124909
|
| 13 |
+
12,0.062282602138362175,0.06926894213993418,0.9762971577111195,0.9744897959183674,0.9973810927180742,0.9967870530136254,0.976210909340565,0.9743589743589743
|
| 14 |
+
13,0.060860307139227,0.06267463378934054,0.977434415295687,0.9795918367346939,0.9974749183749438,0.9977199126214417,0.9773540542164022,0.9797395079594791
|
| 15 |
+
14,0.05755463975885633,0.07566890224770971,0.9785118172179088,0.9766763848396501,0.9977389300470996,0.9971812340096389,0.978445650961926,0.9764359351988218
|
| 16 |
+
15,0.057541033925899064,0.04975939745154047,0.9786058761158806,0.9825072886297376,0.9977538220893306,0.9986517097467891,0.9785357651459259,0.9823788546255506
|
| 17 |
+
16,0.05634996499272484,0.0617915925518243,0.9791274754591784,0.978134110787172,0.997829316750365,0.9969241132521313,0.9790527842376706,0.9780701754385965
|
| 18 |
+
17,0.05454238486925483,0.05333676516911987,0.9799141498785785,0.9810495626822158,0.997946748161562,0.9985412115700091,0.979850919103456,0.9811866859623734
|
| 19 |
+
18,0.053665978950810676,0.08533909733871503,0.9801450217190546,0.967930029154519,0.998023593281266,0.9963652474734167,0.9800902028707149,0.9683908045977011
|
| 20 |
+
19,0.051960351455705815,0.05903876217695784,0.9809060437117352,0.9861516034985423,0.998134591332559,0.9976434138836708,0.9808541468391766,0.9860601614086574
|
| 21 |
+
20,0.052026702721111096,0.08011818320949293,0.9806922734890721,0.9752186588921283,0.9981493410904061,0.996891176295591,0.9806309938410335,0.9751461988304093
|
| 22 |
+
21,0.05115936366098929,0.0657448608634597,0.9809231453295482,0.9810495626822158,0.9981968896653561,0.9971791090447008,0.9808647322692146,0.9811046511627907
|
| 23 |
+
22,0.039293639540274,0.058326971969016046,0.9854892772856312,0.9839650145772595,0.9989317592618322,0.9985231493680354,0.9854545766227533,0.9839883551673945
|
| 24 |
+
23,0.033775414704451334,0.05694627253414721,0.987840749734925,0.9817784256559767,0.9992104630761648,0.9986453348519749,0.9878144923562076,0.9818181818181818
|
| 25 |
+
24,0.033230631860863655,0.05545937554481773,0.9876526319389815,0.9832361516034985,0.9992240350761206,0.9987367083443124,0.9876325390979633,0.9832239241429613
|
| 26 |
+
25,0.03237407778996703,0.0587208214543119,0.9882426377535315,0.9810495626822158,0.9992709570440971,0.9986835842208603,0.9882171472642358,0.9810771470160117
|
| 27 |
+
26,0.03220497033339533,0.06503236680848215,0.9882169853268119,0.9832361516034985,0.9992812725888158,0.9983733393399009,0.9881963955321045,0.9832483612527313
|
| 28 |
+
27,0.03042350235752952,0.06034932669480236,0.9890207613640251,0.9839650145772595,0.999364067769463,0.9985539613596376,0.9890068493150685,0.9839181286549707
|
| 29 |
+
28,0.029133988782805785,0.06084428801430042,0.9891319218798098,0.9832361516034985,0.9994052302176598,0.998483837516681,0.9891105989598952,0.9832727272727273
|
| 30 |
+
29,0.02860731827776991,0.06154062449063226,0.9898416390190512,0.9832361516034985,0.9994026410220773,0.9985199619206284,0.9898245854460738,0.9832970225127088
|
| 31 |
+
30,0.02884942300385336,0.06216380139134021,0.989773232547799,0.9825072886297376,0.9994035903645443,0.9985242118505044,0.9897565905548228,0.9825834542815675
|
| 32 |
+
31,0.029171249110760186,0.061966189061109594,0.9894739542360708,0.9832361516034985,0.9994051022640773,0.9985072121309998,0.9894542058957072,0.9832483612527313
|
| 33 |
+
32,0.028459394045261276,0.06270270375467012,0.9896449704142012,0.9825072886297376,0.9994358494368771,0.9985316492277877,0.9896278531968652,0.982532751091703
|
| 34 |
+
33,0.029123750241781602,0.06183103346550951,0.9890293121729316,0.9832361516034985,0.9994153125212766,0.9985465239823543,0.9890092945560457,0.9832727272727273
|
| 35 |
+
34,0.028456661750446508,0.06152658220943954,0.9893371412935664,0.9832361516034985,0.9994328933435357,0.998566711149266,0.9893239043517718,0.9832727272727273
|
| 36 |
+
35,0.02765608663334115,0.061544772590518694,0.9898245374012381,0.9832361516034985,0.9994650532701209,0.998566711149266,0.9898081534772182,0.9832727272727273
|
training_notebook_c2.ipynb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3840ec461cf0a549c6e39196be842adfbe3b10df2d6dbc623fd1d1a0a8689940
|
| 3 |
+
size 25453968
|