Upload ViT model from experiment a2
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +2 -0
- README.md +161 -0
- config.json +76 -0
- confusion_matrices/ViT_Confusion_Matrix_a.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_b.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_c.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_d.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_e.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_f.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_g.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_h.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_i.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_j.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_k.png +0 -0
- confusion_matrices/ViT_Confusion_Matrix_l.png +0 -0
- evaluation_results.csv +133 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_a.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_b.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_c.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_d.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_e.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_f.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_g.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_h.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_i.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_j.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_k.png +0 -0
- roc_confusion_matrix/ViT_roc_confusion_matrix_l.png +0 -0
- roc_curves/ViT_ROC_a.png +0 -0
- roc_curves/ViT_ROC_b.png +0 -0
- roc_curves/ViT_ROC_c.png +0 -0
- roc_curves/ViT_ROC_d.png +0 -0
- roc_curves/ViT_ROC_e.png +0 -0
- roc_curves/ViT_ROC_f.png +0 -0
- roc_curves/ViT_ROC_g.png +0 -0
- roc_curves/ViT_ROC_h.png +0 -0
- roc_curves/ViT_ROC_i.png +0 -0
- roc_curves/ViT_ROC_j.png +0 -0
- roc_curves/ViT_ROC_k.png +0 -0
- roc_curves/ViT_ROC_l.png +0 -0
- training_curves/ViT_accuracy.png +0 -0
- training_curves/ViT_auc.png +0 -0
- training_curves/ViT_combined_metrics.png +3 -0
- training_curves/ViT_f1.png +0 -0
- training_curves/ViT_loss.png +0 -0
- training_curves/ViT_metrics.csv +56 -0
- training_metrics.csv +56 -0
- training_notebook_a2.ipynb +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
training_curves/ViT_combined_metrics.png filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
training_notebook_a2.ipynb filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,161 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- vision-transformer
|
| 5 |
+
- image-classification
|
| 6 |
+
- pytorch
|
| 7 |
+
- timm
|
| 8 |
+
- vit
|
| 9 |
+
- gravitational-lensing
|
| 10 |
+
- strong-lensing
|
| 11 |
+
- astronomy
|
| 12 |
+
- astrophysics
|
| 13 |
+
datasets:
|
| 14 |
+
- C21
|
| 15 |
+
metrics:
|
| 16 |
+
- accuracy
|
| 17 |
+
- auc
|
| 18 |
+
- f1
|
| 19 |
+
model-index:
|
| 20 |
+
- name: ViT-a2
|
| 21 |
+
results:
|
| 22 |
+
- task:
|
| 23 |
+
type: image-classification
|
| 24 |
+
name: Strong Gravitational Lens Discovery
|
| 25 |
+
dataset:
|
| 26 |
+
type: common-test-sample
|
| 27 |
+
name: Common Test Sample (More et al. 2024)
|
| 28 |
+
metrics:
|
| 29 |
+
- type: accuracy
|
| 30 |
+
value: 0.8205
|
| 31 |
+
name: Average Accuracy
|
| 32 |
+
- type: auc
|
| 33 |
+
value: 0.8511
|
| 34 |
+
name: Average AUC-ROC
|
| 35 |
+
- type: f1
|
| 36 |
+
value: 0.5319
|
| 37 |
+
name: Average F1-Score
|
| 38 |
+
---
|
| 39 |
+
|
| 40 |
+
# 🌌 vit-gravit-a2
|
| 41 |
+
|
| 42 |
+
🔭 This model is part of **GraViT**: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery
|
| 43 |
+
|
| 44 |
+
🔗 **GitHub Repository**: [https://github.com/parlange/gravit](https://github.com/parlange/gravit)
|
| 45 |
+
|
| 46 |
+
## 🛰️ Model Details
|
| 47 |
+
|
| 48 |
+
- **🤖 Model Type**: ViT
|
| 49 |
+
- **🧪 Experiment**: A2 - C21-half
|
| 50 |
+
- **🌌 Dataset**: C21
|
| 51 |
+
- **🪐 Fine-tuning Strategy**: half
|
| 52 |
+
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
## 💻 Quick Start
|
| 56 |
+
|
| 57 |
+
```python
|
| 58 |
+
import torch
|
| 59 |
+
import timm
|
| 60 |
+
|
| 61 |
+
# Load the model directly from the Hub
|
| 62 |
+
model = timm.create_model(
|
| 63 |
+
'hf-hub:parlange/vit-gravit-a2',
|
| 64 |
+
pretrained=True
|
| 65 |
+
)
|
| 66 |
+
model.eval()
|
| 67 |
+
|
| 68 |
+
# Example inference
|
| 69 |
+
dummy_input = torch.randn(1, 3, 224, 224)
|
| 70 |
+
with torch.no_grad():
|
| 71 |
+
output = model(dummy_input)
|
| 72 |
+
predictions = torch.softmax(output, dim=1)
|
| 73 |
+
print(f"Lens probability: {predictions[0][1]:.4f}")
|
| 74 |
+
```
|
| 75 |
+
|
| 76 |
+
## ⚡️ Training Configuration
|
| 77 |
+
|
| 78 |
+
**Training Dataset:** C21 (Cañameras et al. 2021)
|
| 79 |
+
**Fine-tuning Strategy:** half
|
| 80 |
+
|
| 81 |
+
|
| 82 |
+
| 🔧 Parameter | 📝 Value |
|
| 83 |
+
|--------------|----------|
|
| 84 |
+
| Batch Size | 192 |
|
| 85 |
+
| Learning Rate | AdamW with ReduceLROnPlateau |
|
| 86 |
+
| Epochs | 100 |
|
| 87 |
+
| Patience | 10 |
|
| 88 |
+
| Optimizer | AdamW |
|
| 89 |
+
| Scheduler | ReduceLROnPlateau |
|
| 90 |
+
| Image Size | 224x224 |
|
| 91 |
+
| Fine Tune Mode | half |
|
| 92 |
+
| Stochastic Depth Probability | 0.1 |
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
## 📈 Training Curves
|
| 96 |
+
|
| 97 |
+

|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
## 🏁 Final Epoch Training Metrics
|
| 101 |
+
|
| 102 |
+
| Metric | Training | Validation |
|
| 103 |
+
|:---------:|:-----------:|:-------------:|
|
| 104 |
+
| 📉 Loss | 0.0159 | 0.0354 |
|
| 105 |
+
| 🎯 Accuracy | 0.9939 | 0.9870 |
|
| 106 |
+
| 📊 AUC-ROC | 0.9998 | 0.9986 |
|
| 107 |
+
| ⚖️ F1 Score | 0.9939 | 0.9870 |
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
## ☑️ Evaluation Results
|
| 111 |
+
|
| 112 |
+
### ROC Curves and Confusion Matrices
|
| 113 |
+
|
| 114 |
+
Performance across all test datasets (a through l) in the Common Test Sample (More et al. 2024):
|
| 115 |
+
|
| 116 |
+

|
| 117 |
+

|
| 118 |
+

|
| 119 |
+

|
| 120 |
+

|
| 121 |
+

|
| 122 |
+

|
| 123 |
+

|
| 124 |
+

|
| 125 |
+

|
| 126 |
+

|
| 127 |
+

|
| 128 |
+
|
| 129 |
+
### 📋 Performance Summary
|
| 130 |
+
|
| 131 |
+
Average performance across 12 test datasets from the Common Test Sample (More et al. 2024):
|
| 132 |
+
|
| 133 |
+
| Metric | Value |
|
| 134 |
+
|-----------|----------|
|
| 135 |
+
| 🎯 Average Accuracy | 0.8205 |
|
| 136 |
+
| 📈 Average AUC-ROC | 0.8511 |
|
| 137 |
+
| ⚖️ Average F1-Score | 0.5319 |
|
| 138 |
+
|
| 139 |
+
|
| 140 |
+
## 📘 Citation
|
| 141 |
+
|
| 142 |
+
If you use this model in your research, please cite:
|
| 143 |
+
|
| 144 |
+
```bibtex
|
| 145 |
+
@misc{parlange2025gravit,
|
| 146 |
+
title={GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery},
|
| 147 |
+
author={René Parlange and Juan C. Cuevas-Tello and Octavio Valenzuela and Omar de J. Cabrera-Rosas and Tomás Verdugo and Anupreeta More and Anton T. Jaelani},
|
| 148 |
+
year={2025},
|
| 149 |
+
eprint={2509.00226},
|
| 150 |
+
archivePrefix={arXiv},
|
| 151 |
+
primaryClass={cs.CV},
|
| 152 |
+
url={https://arxiv.org/abs/2509.00226},
|
| 153 |
+
}
|
| 154 |
+
```
|
| 155 |
+
|
| 156 |
+
---
|
| 157 |
+
|
| 158 |
+
|
| 159 |
+
## Model Card Contact
|
| 160 |
+
|
| 161 |
+
For questions about this model, please contact the author through: https://github.com/parlange/
|
config.json
ADDED
|
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architecture": "vit_base_patch16_224",
|
| 3 |
+
"num_classes": 2,
|
| 4 |
+
"num_features": 768,
|
| 5 |
+
"global_pool": "token",
|
| 6 |
+
"crop_pct": 0.875,
|
| 7 |
+
"interpolation": "bicubic",
|
| 8 |
+
"mean": [
|
| 9 |
+
0.485,
|
| 10 |
+
0.456,
|
| 11 |
+
0.406
|
| 12 |
+
],
|
| 13 |
+
"std": [
|
| 14 |
+
0.229,
|
| 15 |
+
0.224,
|
| 16 |
+
0.225
|
| 17 |
+
],
|
| 18 |
+
"first_conv": "patch_embed.proj",
|
| 19 |
+
"classifier": "head",
|
| 20 |
+
"input_size": [
|
| 21 |
+
3,
|
| 22 |
+
224,
|
| 23 |
+
224
|
| 24 |
+
],
|
| 25 |
+
"pool_size": [
|
| 26 |
+
7,
|
| 27 |
+
7
|
| 28 |
+
],
|
| 29 |
+
"pretrained_cfg": {
|
| 30 |
+
"tag": "gravit_a2",
|
| 31 |
+
"custom_load": false,
|
| 32 |
+
"input_size": [
|
| 33 |
+
3,
|
| 34 |
+
224,
|
| 35 |
+
224
|
| 36 |
+
],
|
| 37 |
+
"fixed_input_size": true,
|
| 38 |
+
"interpolation": "bicubic",
|
| 39 |
+
"crop_pct": 0.875,
|
| 40 |
+
"crop_mode": "center",
|
| 41 |
+
"mean": [
|
| 42 |
+
0.485,
|
| 43 |
+
0.456,
|
| 44 |
+
0.406
|
| 45 |
+
],
|
| 46 |
+
"std": [
|
| 47 |
+
0.229,
|
| 48 |
+
0.224,
|
| 49 |
+
0.225
|
| 50 |
+
],
|
| 51 |
+
"num_classes": 2,
|
| 52 |
+
"pool_size": [
|
| 53 |
+
7,
|
| 54 |
+
7
|
| 55 |
+
],
|
| 56 |
+
"first_conv": "patch_embed.proj",
|
| 57 |
+
"classifier": "head"
|
| 58 |
+
},
|
| 59 |
+
"model_name": "vit_gravit_a2",
|
| 60 |
+
"experiment": "a2",
|
| 61 |
+
"training_strategy": "half",
|
| 62 |
+
"dataset": "C21",
|
| 63 |
+
"hyperparameters": {
|
| 64 |
+
"batch_size": "192",
|
| 65 |
+
"learning_rate": "AdamW with ReduceLROnPlateau",
|
| 66 |
+
"epochs": "100",
|
| 67 |
+
"patience": "10",
|
| 68 |
+
"optimizer": "AdamW",
|
| 69 |
+
"scheduler": "ReduceLROnPlateau",
|
| 70 |
+
"image_size": "224x224",
|
| 71 |
+
"fine_tune_mode": "half",
|
| 72 |
+
"stochastic_depth_probability": "0.1"
|
| 73 |
+
},
|
| 74 |
+
"hf_hub_id": "parlange/vit-gravit-a2",
|
| 75 |
+
"license": "apache-2.0"
|
| 76 |
+
}
|
confusion_matrices/ViT_Confusion_Matrix_a.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_b.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_c.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_d.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_e.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_f.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_g.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_h.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_i.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_j.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_k.png
ADDED
|
confusion_matrices/ViT_Confusion_Matrix_l.png
ADDED
|
evaluation_results.csv
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Model,Dataset,Loss,Accuracy,AUCROC,F1
|
| 2 |
+
ViT,a,0.3987342550652614,0.8997170701037409,0.917220073664825,0.46921797004991683
|
| 3 |
+
ViT,b,0.37321944226074877,0.9182646966362779,0.9273655616942909,0.5202952029520295
|
| 4 |
+
ViT,c,0.821146962441052,0.7922037095253065,0.866000920810313,0.2990455991516437
|
| 5 |
+
ViT,d,0.223954888615208,0.9440427538509902,0.9518121546961327,0.6130434782608696
|
| 6 |
+
ViT,e,1.1019757730900652,0.7782656421514819,0.85117687126315,0.5826446280991735
|
| 7 |
+
ViT,f,0.43205039906387277,0.8869181318255751,0.9119053612426381,0.1618828932261768
|
| 8 |
+
ViT,g,0.1527516215024516,0.9611666666666666,0.9983996666666667,0.962461736748832
|
| 9 |
+
ViT,h,0.39022785605769605,0.8943333333333333,0.9959461111111112,0.9040556900726392
|
| 10 |
+
ViT,i,0.07361652948241681,0.9748333333333333,0.9992922222222222,0.9753469387755102
|
| 11 |
+
ViT,j,6.027309565424919,0.5033333333333333,0.49077061111111114,0.13872832369942195
|
| 12 |
+
ViT,k,5.948174474835396,0.517,0.5691439444444444,0.14209591474245115
|
| 13 |
+
ViT,l,2.1620761937842525,0.7761620221035376,0.7346207805818022,0.6140942656577628
|
| 14 |
+
MLP-Mixer,a,1.0354112619425808,0.7239861678717384,0.9444742173112339,0.2779605263157895
|
| 15 |
+
MLP-Mixer,b,0.882929647823281,0.7834014460861365,0.9520524861878453,0.3291139240506329
|
| 16 |
+
MLP-Mixer,c,1.6300886590238712,0.6205595724614901,0.9170009208103129,0.21877022653721684
|
| 17 |
+
MLP-Mixer,d,0.055560619117974934,0.9792518076076705,0.9959686924493554,0.8366336633663366
|
| 18 |
+
MLP-Mixer,e,1.1025473987755214,0.70801317233809,0.9327480511617346,0.5596026490066225
|
| 19 |
+
MLP-Mixer,f,0.9539809344458585,0.763147703508636,0.9512486274645963,0.09952885747938751
|
| 20 |
+
MLP-Mixer,g,0.4660766951590776,0.8855,0.9941798888888888,0.8969551522423879
|
| 21 |
+
MLP-Mixer,h,0.8621955468207598,0.7991666666666667,0.9892695555555555,0.8322894919972165
|
| 22 |
+
MLP-Mixer,i,0.027433219969272612,0.9893333333333333,0.9997324444444444,0.9894109861019192
|
| 23 |
+
MLP-Mixer,j,5.562473163604737,0.4046666666666667,0.2374668888888889,0.05552617662612375
|
| 24 |
+
MLP-Mixer,k,5.123829672321677,0.5085,0.47802688888888895,0.06647673314339982
|
| 25 |
+
MLP-Mixer,l,2.271000012816855,0.6846808735656497,0.6265659288601144,0.5226162837242815
|
| 26 |
+
CvT,a,0.7062503800231756,0.6988368437598239,0.8373821362799264,0.2336
|
| 27 |
+
CvT,b,0.8609082461227902,0.6444514303678088,0.8087320441988951,0.20520028109627547
|
| 28 |
+
CvT,c,0.8150388154171053,0.6516818610499843,0.8144069981583795,0.20857142857142857
|
| 29 |
+
CvT,d,0.046723183949472995,0.9833385727758567,0.9917753222836095,0.8463768115942029
|
| 30 |
+
CvT,e,1.055778265130245,0.5916575192096597,0.7447665178233557,0.4397590361445783
|
| 31 |
+
CvT,f,0.6479923738885578,0.7303074897374332,0.8562897926766284,0.07737148913619502
|
| 32 |
+
CvT,g,0.47873973870277403,0.8053333333333333,0.948009388888889,0.8337129840546698
|
| 33 |
+
CvT,h,0.45442124152183533,0.8091666666666667,0.9536063888888889,0.8364519354377946
|
| 34 |
+
CvT,i,0.0470859190672636,0.985,0.999245111111111,0.9848637739656912
|
| 35 |
+
CvT,j,3.978341913700104,0.31966666666666665,0.09071466666666667,0.006812652068126521
|
| 36 |
+
CvT,k,3.5466880963295697,0.49933333333333335,0.5394476666666667,0.009234828496042216
|
| 37 |
+
CvT,l,1.575411782030338,0.6541695309608164,0.5834616840778439,0.4856873230575653
|
| 38 |
+
Swin,a,0.7181912302146427,0.6636277900031436,0.8782845303867403,0.22351233671988388
|
| 39 |
+
Swin,b,0.47780195057523284,0.8000628733102798,0.9174990791896869,0.326271186440678
|
| 40 |
+
Swin,c,1.1387689553203542,0.5369380697893744,0.8261528545119704,0.1729365524985963
|
| 41 |
+
Swin,d,0.0309918804128743,0.9871109713926438,0.9966482504604052,0.8825214899713467
|
| 42 |
+
Swin,e,0.5848406514011806,0.7464324917672887,0.8943237720426852,0.5714285714285714
|
| 43 |
+
Swin,f,0.6053860638443279,0.7410735032143134,0.9040542417311522,0.0843604491920022
|
| 44 |
+
Swin,g,0.24402792798448353,0.8985,0.999229,0.9078529278256923
|
| 45 |
+
Swin,h,0.5944505902426317,0.759,0.9974725555555555,0.8058017727639001
|
| 46 |
+
Swin,i,0.007144115434028208,0.9976666666666667,0.9999902222222222,0.9976720984369803
|
| 47 |
+
Swin,j,3.9496381425857545,0.419,0.11956033333333334,0.06591639871382636
|
| 48 |
+
Swin,k,3.7127543271519245,0.5181666666666667,0.4697328333333333,0.07841887153331208
|
| 49 |
+
Swin,l,1.5895203038408208,0.6710908994764951,0.5955901653865907,0.5130734304055112
|
| 50 |
+
CaiT,a,0.11189156073487824,0.967934611757309,0.9444355432780847,0.7243243243243244
|
| 51 |
+
CaiT,b,0.14077727915177515,0.9566174159069475,0.9445782688766113,0.6600985221674877
|
| 52 |
+
CaiT,c,0.1453998264081071,0.9556743162527507,0.9236316758747698,0.6552567237163814
|
| 53 |
+
CaiT,d,0.0895651844381368,0.9757937755422823,0.9521049723756906,0.7768115942028986
|
| 54 |
+
CaiT,e,0.5067908598680527,0.8507135016465422,0.8893589646560206,0.6633663366336634
|
| 55 |
+
CaiT,f,0.09735689565035,0.968553946247386,0.9382155086735556,0.39762611275964393
|
| 56 |
+
CaiT,g,0.04386970533267595,0.9845,0.9998996666666666,0.984726556084743
|
| 57 |
+
CaiT,h,0.04632043000892736,0.984,0.999908,0.9842416283650689
|
| 58 |
+
CaiT,i,0.016718765972414985,0.9946666666666667,0.9999668888888888,0.9946914399469144
|
| 59 |
+
CaiT,j,4.091754978463054,0.5111666666666667,0.502447,0.09726069559864574
|
| 60 |
+
CaiT,k,4.064604013383389,0.5213333333333333,0.4420845,0.09912170639899624
|
| 61 |
+
CaiT,l,1.3512259357719865,0.8281423510126381,0.7136882113330858,0.669379450661241
|
| 62 |
+
DeiT,a,0.2868140126428901,0.9104055328513047,0.9108996316758747,0.5060658578856152
|
| 63 |
+
DeiT,b,0.21093111757876476,0.9380697893744105,0.9419907918968692,0.5971370143149284
|
| 64 |
+
DeiT,c,0.4344255732631803,0.8610499842816725,0.8816003683241251,0.3978201634877384
|
| 65 |
+
DeiT,d,0.10285686065156817,0.977051241747878,0.9568195211786372,0.8
|
| 66 |
+
DeiT,e,0.6603749339457532,0.8210757409440176,0.8680541890562324,0.6417582417582418
|
| 67 |
+
DeiT,f,0.23529704651809683,0.9209975989466347,0.9196866062244752,0.2225609756097561
|
| 68 |
+
DeiT,g,0.07966256512608379,0.9715,0.9994743333333334,0.9722086786933203
|
| 69 |
+
DeiT,h,0.1981518727680668,0.9306666666666666,0.9987041666666666,0.9349796811503595
|
| 70 |
+
DeiT,i,0.022365192129276693,0.9921666666666666,0.9998618888888888,0.9922043456626306
|
| 71 |
+
DeiT,j,4.922376271247864,0.4945,0.5104821111111111,0.07839562443026436
|
| 72 |
+
DeiT,k,4.865078900694847,0.5151666666666667,0.5016916666666666,0.08146510893590149
|
| 73 |
+
DeiT,l,1.6993422816543697,0.7937708212151657,0.7000134906492581,0.6261503067484663
|
| 74 |
+
DeiT3,a,0.3992725303153578,0.9091480666457089,0.9363876611418048,0.5126475548060708
|
| 75 |
+
DeiT3,b,0.3124380833470795,0.9305249921408362,0.952292817679558,0.579047619047619
|
| 76 |
+
DeiT3,c,0.4766911079170793,0.8915435397673688,0.9285046040515654,0.46841294298921415
|
| 77 |
+
DeiT3,d,0.11976070332006253,0.9723357434768941,0.9784143646408839,0.7755102040816326
|
| 78 |
+
DeiT3,e,0.9584465352031193,0.7969264544456641,0.8915991826231742,0.621676891615542
|
| 79 |
+
DeiT3,f,0.33781881941231834,0.921617225621563,0.9456139627538378,0.23100303951367782
|
| 80 |
+
DeiT3,g,0.1398290316515031,0.9671666666666666,0.9996258333333332,0.9681590431550025
|
| 81 |
+
DeiT3,h,0.22691049999523785,0.9465,0.9994123333333333,0.9491364284582475
|
| 82 |
+
DeiT3,i,0.03767789686251308,0.9893333333333333,0.9998997777777778,0.9894284770399736
|
| 83 |
+
DeiT3,j,6.353702080726624,0.486,0.2588749444444445,0.06545454545454546
|
| 84 |
+
DeiT3,k,6.251550879061222,0.5081666666666667,0.38331972222222227,0.06820334701610357
|
| 85 |
+
DeiT3,l,2.203556089656545,0.7932949077256624,0.6245051702293716,0.6248200403109704
|
| 86 |
+
Twins_SVT,a,0.42451784632109274,0.8126375353662371,0.8918922651933701,0.3377777777777778
|
| 87 |
+
Twins_SVT,b,0.3914457758111573,0.8142093681232316,0.8958011049723758,0.3396648044692737
|
| 88 |
+
Twins_SVT,c,0.44262768309920586,0.8047783715812638,0.8853812154696132,0.3286486486486486
|
| 89 |
+
Twins_SVT,d,0.07794447650992994,0.9820811065702609,0.9897808471454881,0.8421052631578947
|
| 90 |
+
Twins_SVT,e,0.5991019170841715,0.712403951701427,0.818761825474911,0.5371024734982333
|
| 91 |
+
Twins_SVT,f,0.3470329082674194,0.8442413445898846,0.9101541579685174,0.13131749460043196
|
| 92 |
+
Twins_SVT,g,0.2309408655166626,0.9008333333333334,0.9874507777777777,0.9088681268188084
|
| 93 |
+
Twins_SVT,h,0.2580758459568024,0.8958333333333334,0.9890552222222222,0.9047110840067083
|
| 94 |
+
Twins_SVT,i,0.06473293882608414,0.9898333333333333,0.9991408888888889,0.9898248540450375
|
| 95 |
+
Twins_SVT,j,2.941682351350784,0.41483333333333333,0.16385111111111111,0.02823138665928591
|
| 96 |
+
Twins_SVT,k,2.775474429190159,0.5038333333333334,0.41604216666666666,0.03312763884378045
|
| 97 |
+
Twins_SVT,l,1.1202095961319853,0.7359737718788008,0.5944821592359222,0.5594282184770141
|
| 98 |
+
Twins_PCPVT,a,0.4347801183106501,0.7944042753850991,0.8711482504604051,0.2952586206896552
|
| 99 |
+
Twins_PCPVT,b,0.339161620420017,0.8575919522162841,0.9065580110497238,0.3768913342503439
|
| 100 |
+
Twins_PCPVT,c,0.467156688444122,0.7742848160955674,0.8578029465930019,0.2762096774193548
|
| 101 |
+
Twins_PCPVT,d,0.23605295665344328,0.9245520276642565,0.9460681399631676,0.5330739299610895
|
| 102 |
+
Twins_PCPVT,e,0.4829401325446714,0.7694840834248079,0.8557481268447741,0.5661157024793388
|
| 103 |
+
Twins_PCPVT,f,0.3720619066015847,0.8374254511656727,0.8931208308558979,0.11546565528866413
|
| 104 |
+
Twins_PCPVT,g,0.2366939251422882,0.9141666666666667,0.9769166666666668,0.9182928764080597
|
| 105 |
+
Twins_PCPVT,h,0.3045526731014252,0.87,0.9672449999999999,0.8812423873325214
|
| 106 |
+
Twins_PCPVT,i,0.18202911043167114,0.9496666666666667,0.9892225555555555,0.9504105090311987
|
| 107 |
+
Twins_PCPVT,j,1.547880547761917,0.4866666666666667,0.43741700000000006,0.1760299625468165
|
| 108 |
+
Twins_PCPVT,k,1.4932157402038575,0.5221666666666667,0.47360899999999995,0.18666666666666668
|
| 109 |
+
Twins_PCPVT,l,0.7149772635186269,0.7421606472423458,0.6893006569431472,0.579510175922732
|
| 110 |
+
PiT,a,0.44727815774318297,0.8010059729644766,0.9252882136279927,0.3329820864067439
|
| 111 |
+
PiT,b,0.3666396583074745,0.8374724929267526,0.9387182320441989,0.3793517406962785
|
| 112 |
+
PiT,c,0.587441599189167,0.7419050613014775,0.9032854511970534,0.27792436235708
|
| 113 |
+
PiT,d,0.033625746908820454,0.9855391386356491,0.9966703499079189,0.8729281767955801
|
| 114 |
+
PiT,e,0.631099864955006,0.7387486278814489,0.8873495799591311,0.5703971119133574
|
| 115 |
+
PiT,f,0.3817344946660815,0.8324684377662458,0.9379145273921177,0.127470754336426
|
| 116 |
+
PiT,g,0.19083814918994904,0.9163333333333333,0.9971276666666666,0.9226025285229725
|
| 117 |
+
PiT,h,0.30789998161792753,0.8656666666666667,0.9943662222222222,0.8812960235640648
|
| 118 |
+
PiT,i,0.01428527906537056,0.9948333333333333,0.9999181111111111,0.9948462177888612
|
| 119 |
+
PiT,j,5.1374336256980895,0.427,0.18367044444444444,0.031549295774647886
|
| 120 |
+
PiT,k,4.960880736589432,0.5055,0.6351591666666667,0.03637544657356284
|
| 121 |
+
PiT,l,1.8333862431563666,0.7295225001321982,0.6306325748279862,0.5562592174893728
|
| 122 |
+
Ensemble,a,,0.9239232945614586,0.9640395948434622,0.56
|
| 123 |
+
Ensemble,b,,0.926752593524049,0.9696408839779006,0.5693160813308688
|
| 124 |
+
Ensemble,c,,0.8563344860106885,0.9442707182320442,0.40261437908496733
|
| 125 |
+
Ensemble,d,,0.9874253379440427,0.9976040515653776,0.8850574712643678
|
| 126 |
+
Ensemble,e,,0.8254665203073546,0.9312873685007189,0.6595289079229122
|
| 127 |
+
Ensemble,f,,0.9207652389435366,0.9667325628328263,0.23140495867768596
|
| 128 |
+
Ensemble,g,,0.9656666666666667,0.999739111111111,0.9668063164679342
|
| 129 |
+
Ensemble,h,,0.9283333333333333,0.9995118888888889,0.9331259720062208
|
| 130 |
+
Ensemble,i,,0.9978333333333333,0.9999951111111111,0.9978380176284717
|
| 131 |
+
Ensemble,j,,0.47883333333333333,0.20988066666666666,0.04809741248097413
|
| 132 |
+
Ensemble,k,,0.511,0.48139255555555555,0.05109961190168176
|
| 133 |
+
Ensemble,l,,0.79144413304426,0.6240045673759117,0.6211335254562921
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e50d176c3ff9296a3ecd94bc719db684360b19a03d9d21e19110d04ab1a319ad
|
| 3 |
+
size 343214864
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:410c6c52b3124c455a228d8864b71f5ee82f888f67ad52ddd4c748362a397c14
|
| 3 |
+
size 343259038
|
roc_confusion_matrix/ViT_roc_confusion_matrix_a.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_b.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_c.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_d.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_e.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_f.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_g.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_h.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_i.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_j.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_k.png
ADDED
|
roc_confusion_matrix/ViT_roc_confusion_matrix_l.png
ADDED
|
roc_curves/ViT_ROC_a.png
ADDED
|
roc_curves/ViT_ROC_b.png
ADDED
|
roc_curves/ViT_ROC_c.png
ADDED
|
roc_curves/ViT_ROC_d.png
ADDED
|
roc_curves/ViT_ROC_e.png
ADDED
|
roc_curves/ViT_ROC_f.png
ADDED
|
roc_curves/ViT_ROC_g.png
ADDED
|
roc_curves/ViT_ROC_h.png
ADDED
|
roc_curves/ViT_ROC_i.png
ADDED
|
roc_curves/ViT_ROC_j.png
ADDED
|
roc_curves/ViT_ROC_k.png
ADDED
|
roc_curves/ViT_ROC_l.png
ADDED
|
training_curves/ViT_accuracy.png
ADDED
|
training_curves/ViT_auc.png
ADDED
|
training_curves/ViT_combined_metrics.png
ADDED
|
Git LFS Details
|
training_curves/ViT_f1.png
ADDED
|
training_curves/ViT_loss.png
ADDED
|
training_curves/ViT_metrics.csv
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.2571897825062275,0.16868997523188592,0.8909625,0.946,0.9619713853125,0.982836,0.8921635286991136,0.9458917835671342
|
| 3 |
+
2,0.10249610567688942,0.08976531195640564,0.959125,0.964,0.9935048543749999,0.995018,0.9591464481147398,0.9636363636363636
|
| 4 |
+
3,0.08254435644596815,0.12431702297925949,0.9676,0.95,0.9957464584375,0.994376,0.9676105265788618,0.9514563106796117
|
| 5 |
+
4,0.07725174049884082,0.07002760744094849,0.9704625,0.971,0.9962299949999999,0.997188,0.9704842678524588,0.9710867397806581
|
| 6 |
+
5,0.06949511439502239,0.08071183396875858,0.97285,0.967,0.9969555387499999,0.9960100000000001,0.9728581425572328,0.9669669669669669
|
| 7 |
+
6,0.06195828217118979,0.06686169847846031,0.97555,0.97,0.9975565184374999,0.997854,0.9755432744004602,0.9703557312252964
|
| 8 |
+
7,0.05968754206672311,0.05395021122694015,0.9770875,0.976,0.9977075931249999,0.99835,0.9771009531900009,0.976
|
| 9 |
+
8,0.05582731359452009,0.09855114138126374,0.9784875,0.959,0.9980155462499999,0.997438,0.9784974449317191,0.9603864734299516
|
| 10 |
+
9,0.05249064586088061,0.12695906853675842,0.9802875,0.964,0.9982350368749999,0.997152,0.9802961204473043,0.965183752417795
|
| 11 |
+
10,0.05202052024230361,0.0683318508900702,0.9794375,0.977,0.998266334375,0.9976320000000001,0.979444437502343,0.9767911200807265
|
| 12 |
+
11,0.04922362358346581,0.06005106982588768,0.980925,0.98,0.998434173125,0.9980680000000001,0.9809359618219524,0.9799599198396793
|
| 13 |
+
12,0.04954018256664276,0.0496393184363842,0.9809875,0.983,0.998417308125,0.998878,0.9809915392978992,0.9831849653808111
|
| 14 |
+
13,0.04701879207864404,0.06633866775035858,0.9822625,0.976,0.9985571665625002,0.998196,0.9822675980655562,0.9763779527559056
|
| 15 |
+
14,0.04468993845023215,0.05411972495913506,0.9832375,0.977,0.9986921206249999,0.997852,0.9832490163012929,0.9771598808341608
|
| 16 |
+
15,0.04467305731698871,0.052378923520445826,0.9830875,0.983,0.9987308056249999,0.998612,0.9830978525653038,0.9829488465396189
|
| 17 |
+
16,0.04416244664862752,0.061337709560990335,0.9831375,0.972,0.9987252821875,0.997954,0.9831423465753596,0.9719438877755511
|
| 18 |
+
17,0.04224556235820055,0.05293858873844147,0.983925,0.976,0.9988342181250001,0.99875,0.9839310258653005,0.9761904761904762
|
| 19 |
+
18,0.04184411993473768,0.04035755966603756,0.9844125,0.983,0.9988489365625,0.9992139999999999,0.9844142533964929,0.983016983016983
|
| 20 |
+
19,0.039481629332900045,0.0380017871260643,0.985075,0.984,0.9989661159375,0.999196,0.9850779844031193,0.9838709677419355
|
| 21 |
+
20,0.039050778500176964,0.05595802050828934,0.9852,0.978,0.9989906596875,0.998618,0.9852040688810577,0.9777327935222672
|
| 22 |
+
21,0.038018626151047646,0.03781456935405731,0.985725,0.984,0.999055713125,0.999188,0.985720359116713,0.9839034205231388
|
| 23 |
+
22,0.03786576530188322,0.07107540512271225,0.9856875,0.975,0.9990700378125001,0.997752,0.9856898254033719,0.9754178957718781
|
| 24 |
+
23,0.03550247654877603,0.04472261417657137,0.986525,0.981,0.99917039375,0.9989760000000001,0.9865276944611078,0.9811320754716981
|
| 25 |
+
24,0.03683767461553216,0.08104223525524139,0.9863375,0.971,0.999118265,0.99753,0.9863400612385178,0.971028971028971
|
| 26 |
+
25,0.03603814362697303,0.054750813365913927,0.9862375,0.977,0.9991466175,0.9986539999999999,0.9862428309030251,0.9767911200807265
|
| 27 |
+
26,0.033973789243213835,0.034988305181264875,0.9874875,0.987,0.9992454203125,0.99852,0.9874957840430715,0.9870388833499502
|
| 28 |
+
27,0.0336566893408075,0.037800674140453336,0.9871375,0.982,0.9992725715625002,0.99936,0.9871331574406362,0.9822134387351779
|
| 29 |
+
28,0.03516939798332751,0.05907388961315155,0.9867375,0.975,0.9991722771875,0.998462,0.9867413118728365,0.9750747756729811
|
| 30 |
+
29,0.03391784453056753,0.06921996867563575,0.9873375,0.976,0.9992547612500001,0.997552,0.987342404818133,0.9761431411530815
|
| 31 |
+
30,0.032122378143295646,0.07228873765468598,0.98765,0.974,0.9993077275,0.9974299999999999,0.9876567887661786,0.9743589743589743
|
| 32 |
+
31,0.032853978098742666,0.03264096333086491,0.9875625,0.986,0.9992887515625,0.999444,0.9875607896085712,0.9860279441117764
|
| 33 |
+
32,0.03169495304683223,0.04754947853460908,0.9878,0.984,0.9993381837499999,0.9988980000000001,0.9878042685060229,0.984
|
| 34 |
+
33,0.0325634615611285,0.07620548486709594,0.9877625,0.976,0.9992953393749999,0.99867,0.9877614291250485,0.9764705882352941
|
| 35 |
+
34,0.03284200843628496,0.053336817268282176,0.9876625,0.983,0.9992856396875,0.9982260000000001,0.9876728240098918,0.9831516352824579
|
| 36 |
+
35,0.030358187890425326,0.026614241600029346,0.9887625,0.991,0.9993883709375,0.998726,0.9887626404669941,0.990990990990991
|
| 37 |
+
36,0.03145117684975266,0.03880496996641159,0.9881625,0.984,0.9993390918749999,0.9991,0.988167380955356,0.9839034205231388
|
| 38 |
+
37,0.03075786703492049,0.043761813044548036,0.988625,0.983,0.9993586171875,0.999152,0.988628695673906,0.9831516352824579
|
| 39 |
+
38,0.029451999595947565,0.03821960013359785,0.988775,0.989,0.9994363434374999,0.9993380000000001,0.9887758418118641,0.989010989010989
|
| 40 |
+
39,0.02943100941479206,0.046573635160923006,0.9888,0.984,0.999430351875,0.999074,0.9888,0.9838709677419355
|
| 41 |
+
40,0.028862864456977694,0.033900146007537844,0.9894875,0.984,0.9994478037499999,0.999448,0.9894907841299594,0.984
|
| 42 |
+
41,0.027756637436151504,0.0390649971216917,0.9895875,0.985,0.9995044921875001,0.99941,0.9895894519777542,0.9850448654037887
|
| 43 |
+
42,0.021649938562326134,0.031183642968535423,0.992075,0.988,0.9996990165625,0.999484,0.9920763866323393,0.9879518072289156
|
| 44 |
+
43,0.018418092272244394,0.0391297372598201,0.9933375,0.988,0.99976330375,0.998498,0.9933389154804604,0.988
|
| 45 |
+
44,0.018358768828725443,0.034538188990205526,0.992775,0.987,0.9997714343749999,0.9986020000000001,0.9927746387319366,0.986986986986987
|
| 46 |
+
45,0.017306528286472896,0.03354046154068783,0.993575,0.989,0.9998045315625,0.9986320000000001,0.9935756424357565,0.988988988988989
|
| 47 |
+
46,0.017489736422430725,0.032686231285333635,0.993675,0.989,0.9997868325,0.999558,0.9936757905261843,0.9889669007021064
|
| 48 |
+
47,0.01589218409322202,0.03669157077372074,0.9937625,0.988,0.9998243946875001,0.9986219999999999,0.9937642927658298,0.988
|
| 49 |
+
48,0.016967264903057368,0.03582666659355164,0.9938625,0.987,0.999815451875,0.998578,0.9938615024941553,0.987012987012987
|
| 50 |
+
49,0.015907147341710514,0.034647640466686426,0.99395,0.988,0.9998361521875001,0.9985879999999999,0.9939521167591343,0.9879759519038076
|
| 51 |
+
50,0.015000127985933795,0.033812878728145736,0.9942125,0.988,0.9998465996874999,0.9986179999999999,0.9942119935494356,0.9879759519038076
|
| 52 |
+
51,0.01568684279639274,0.03433457591384649,0.994325,0.988,0.999837119375,0.998626,0.9943262765877677,0.9879759519038076
|
| 53 |
+
52,0.015734441786119713,0.03452299890294671,0.99385,0.988,0.9998309865625001,0.998634,0.993851537115721,0.9879759519038076
|
| 54 |
+
53,0.015493583727488294,0.03554822874069214,0.99405,0.988,0.9998134003125001,0.99861,0.9940517844646606,0.9879759519038076
|
| 55 |
+
54,0.01566599233309971,0.03530654576420784,0.9943,0.988,0.9998142143750001,0.998606,0.9943009973254681,0.9879759519038076
|
| 56 |
+
55,0.015907046043593435,0.035406885087490084,0.9939125,0.987,0.999830001875,0.998598,0.9939130326096467,0.986986986986987
|
training_metrics.csv
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.2571897825062275,0.16868997523188592,0.8909625,0.946,0.9619713853125,0.982836,0.8921635286991136,0.9458917835671342
|
| 3 |
+
2,0.10249610567688942,0.08976531195640564,0.959125,0.964,0.9935048543749999,0.995018,0.9591464481147398,0.9636363636363636
|
| 4 |
+
3,0.08254435644596815,0.12431702297925949,0.9676,0.95,0.9957464584375,0.994376,0.9676105265788618,0.9514563106796117
|
| 5 |
+
4,0.07725174049884082,0.07002760744094849,0.9704625,0.971,0.9962299949999999,0.997188,0.9704842678524588,0.9710867397806581
|
| 6 |
+
5,0.06949511439502239,0.08071183396875858,0.97285,0.967,0.9969555387499999,0.9960100000000001,0.9728581425572328,0.9669669669669669
|
| 7 |
+
6,0.06195828217118979,0.06686169847846031,0.97555,0.97,0.9975565184374999,0.997854,0.9755432744004602,0.9703557312252964
|
| 8 |
+
7,0.05968754206672311,0.05395021122694015,0.9770875,0.976,0.9977075931249999,0.99835,0.9771009531900009,0.976
|
| 9 |
+
8,0.05582731359452009,0.09855114138126374,0.9784875,0.959,0.9980155462499999,0.997438,0.9784974449317191,0.9603864734299516
|
| 10 |
+
9,0.05249064586088061,0.12695906853675842,0.9802875,0.964,0.9982350368749999,0.997152,0.9802961204473043,0.965183752417795
|
| 11 |
+
10,0.05202052024230361,0.0683318508900702,0.9794375,0.977,0.998266334375,0.9976320000000001,0.979444437502343,0.9767911200807265
|
| 12 |
+
11,0.04922362358346581,0.06005106982588768,0.980925,0.98,0.998434173125,0.9980680000000001,0.9809359618219524,0.9799599198396793
|
| 13 |
+
12,0.04954018256664276,0.0496393184363842,0.9809875,0.983,0.998417308125,0.998878,0.9809915392978992,0.9831849653808111
|
| 14 |
+
13,0.04701879207864404,0.06633866775035858,0.9822625,0.976,0.9985571665625002,0.998196,0.9822675980655562,0.9763779527559056
|
| 15 |
+
14,0.04468993845023215,0.05411972495913506,0.9832375,0.977,0.9986921206249999,0.997852,0.9832490163012929,0.9771598808341608
|
| 16 |
+
15,0.04467305731698871,0.052378923520445826,0.9830875,0.983,0.9987308056249999,0.998612,0.9830978525653038,0.9829488465396189
|
| 17 |
+
16,0.04416244664862752,0.061337709560990335,0.9831375,0.972,0.9987252821875,0.997954,0.9831423465753596,0.9719438877755511
|
| 18 |
+
17,0.04224556235820055,0.05293858873844147,0.983925,0.976,0.9988342181250001,0.99875,0.9839310258653005,0.9761904761904762
|
| 19 |
+
18,0.04184411993473768,0.04035755966603756,0.9844125,0.983,0.9988489365625,0.9992139999999999,0.9844142533964929,0.983016983016983
|
| 20 |
+
19,0.039481629332900045,0.0380017871260643,0.985075,0.984,0.9989661159375,0.999196,0.9850779844031193,0.9838709677419355
|
| 21 |
+
20,0.039050778500176964,0.05595802050828934,0.9852,0.978,0.9989906596875,0.998618,0.9852040688810577,0.9777327935222672
|
| 22 |
+
21,0.038018626151047646,0.03781456935405731,0.985725,0.984,0.999055713125,0.999188,0.985720359116713,0.9839034205231388
|
| 23 |
+
22,0.03786576530188322,0.07107540512271225,0.9856875,0.975,0.9990700378125001,0.997752,0.9856898254033719,0.9754178957718781
|
| 24 |
+
23,0.03550247654877603,0.04472261417657137,0.986525,0.981,0.99917039375,0.9989760000000001,0.9865276944611078,0.9811320754716981
|
| 25 |
+
24,0.03683767461553216,0.08104223525524139,0.9863375,0.971,0.999118265,0.99753,0.9863400612385178,0.971028971028971
|
| 26 |
+
25,0.03603814362697303,0.054750813365913927,0.9862375,0.977,0.9991466175,0.9986539999999999,0.9862428309030251,0.9767911200807265
|
| 27 |
+
26,0.033973789243213835,0.034988305181264875,0.9874875,0.987,0.9992454203125,0.99852,0.9874957840430715,0.9870388833499502
|
| 28 |
+
27,0.0336566893408075,0.037800674140453336,0.9871375,0.982,0.9992725715625002,0.99936,0.9871331574406362,0.9822134387351779
|
| 29 |
+
28,0.03516939798332751,0.05907388961315155,0.9867375,0.975,0.9991722771875,0.998462,0.9867413118728365,0.9750747756729811
|
| 30 |
+
29,0.03391784453056753,0.06921996867563575,0.9873375,0.976,0.9992547612500001,0.997552,0.987342404818133,0.9761431411530815
|
| 31 |
+
30,0.032122378143295646,0.07228873765468598,0.98765,0.974,0.9993077275,0.9974299999999999,0.9876567887661786,0.9743589743589743
|
| 32 |
+
31,0.032853978098742666,0.03264096333086491,0.9875625,0.986,0.9992887515625,0.999444,0.9875607896085712,0.9860279441117764
|
| 33 |
+
32,0.03169495304683223,0.04754947853460908,0.9878,0.984,0.9993381837499999,0.9988980000000001,0.9878042685060229,0.984
|
| 34 |
+
33,0.0325634615611285,0.07620548486709594,0.9877625,0.976,0.9992953393749999,0.99867,0.9877614291250485,0.9764705882352941
|
| 35 |
+
34,0.03284200843628496,0.053336817268282176,0.9876625,0.983,0.9992856396875,0.9982260000000001,0.9876728240098918,0.9831516352824579
|
| 36 |
+
35,0.030358187890425326,0.026614241600029346,0.9887625,0.991,0.9993883709375,0.998726,0.9887626404669941,0.990990990990991
|
| 37 |
+
36,0.03145117684975266,0.03880496996641159,0.9881625,0.984,0.9993390918749999,0.9991,0.988167380955356,0.9839034205231388
|
| 38 |
+
37,0.03075786703492049,0.043761813044548036,0.988625,0.983,0.9993586171875,0.999152,0.988628695673906,0.9831516352824579
|
| 39 |
+
38,0.029451999595947565,0.03821960013359785,0.988775,0.989,0.9994363434374999,0.9993380000000001,0.9887758418118641,0.989010989010989
|
| 40 |
+
39,0.02943100941479206,0.046573635160923006,0.9888,0.984,0.999430351875,0.999074,0.9888,0.9838709677419355
|
| 41 |
+
40,0.028862864456977694,0.033900146007537844,0.9894875,0.984,0.9994478037499999,0.999448,0.9894907841299594,0.984
|
| 42 |
+
41,0.027756637436151504,0.0390649971216917,0.9895875,0.985,0.9995044921875001,0.99941,0.9895894519777542,0.9850448654037887
|
| 43 |
+
42,0.021649938562326134,0.031183642968535423,0.992075,0.988,0.9996990165625,0.999484,0.9920763866323393,0.9879518072289156
|
| 44 |
+
43,0.018418092272244394,0.0391297372598201,0.9933375,0.988,0.99976330375,0.998498,0.9933389154804604,0.988
|
| 45 |
+
44,0.018358768828725443,0.034538188990205526,0.992775,0.987,0.9997714343749999,0.9986020000000001,0.9927746387319366,0.986986986986987
|
| 46 |
+
45,0.017306528286472896,0.03354046154068783,0.993575,0.989,0.9998045315625,0.9986320000000001,0.9935756424357565,0.988988988988989
|
| 47 |
+
46,0.017489736422430725,0.032686231285333635,0.993675,0.989,0.9997868325,0.999558,0.9936757905261843,0.9889669007021064
|
| 48 |
+
47,0.01589218409322202,0.03669157077372074,0.9937625,0.988,0.9998243946875001,0.9986219999999999,0.9937642927658298,0.988
|
| 49 |
+
48,0.016967264903057368,0.03582666659355164,0.9938625,0.987,0.999815451875,0.998578,0.9938615024941553,0.987012987012987
|
| 50 |
+
49,0.015907147341710514,0.034647640466686426,0.99395,0.988,0.9998361521875001,0.9985879999999999,0.9939521167591343,0.9879759519038076
|
| 51 |
+
50,0.015000127985933795,0.033812878728145736,0.9942125,0.988,0.9998465996874999,0.9986179999999999,0.9942119935494356,0.9879759519038076
|
| 52 |
+
51,0.01568684279639274,0.03433457591384649,0.994325,0.988,0.999837119375,0.998626,0.9943262765877677,0.9879759519038076
|
| 53 |
+
52,0.015734441786119713,0.03452299890294671,0.99385,0.988,0.9998309865625001,0.998634,0.993851537115721,0.9879759519038076
|
| 54 |
+
53,0.015493583727488294,0.03554822874069214,0.99405,0.988,0.9998134003125001,0.99861,0.9940517844646606,0.9879759519038076
|
| 55 |
+
54,0.01566599233309971,0.03530654576420784,0.9943,0.988,0.9998142143750001,0.998606,0.9943009973254681,0.9879759519038076
|
| 56 |
+
55,0.015907046043593435,0.035406885087490084,0.9939125,0.987,0.999830001875,0.998598,0.9939130326096467,0.986986986986987
|
training_notebook_a2.ipynb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc5d440a78b5d761bd1d79a4e23ab72bfc8a4744ca67cb8ad7dc260bdf23e550
|
| 3 |
+
size 21151683
|