Upload CvT model from experiment c1
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +2 -0
- README.md +166 -0
- config.json +76 -0
- confusion_matrices/CvT_Confusion_Matrix_a.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_b.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_c.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_d.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_e.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_f.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_g.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_h.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_i.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_j.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_k.png +0 -0
- confusion_matrices/CvT_Confusion_Matrix_l.png +0 -0
- cvt-gravit-c1.pth +3 -0
- evaluation_results.csv +133 -0
- model.safetensors +3 -0
- pytorch_model.bin +3 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_a.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_b.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_c.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_d.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_e.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_f.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_g.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_h.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_i.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_j.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_k.png +0 -0
- roc_confusion_matrix/CvT_roc_confusion_matrix_l.png +0 -0
- roc_curves/CvT_ROC_a.png +0 -0
- roc_curves/CvT_ROC_b.png +0 -0
- roc_curves/CvT_ROC_c.png +0 -0
- roc_curves/CvT_ROC_d.png +0 -0
- roc_curves/CvT_ROC_e.png +0 -0
- roc_curves/CvT_ROC_f.png +0 -0
- roc_curves/CvT_ROC_g.png +0 -0
- roc_curves/CvT_ROC_h.png +0 -0
- roc_curves/CvT_ROC_i.png +0 -0
- roc_curves/CvT_ROC_j.png +0 -0
- roc_curves/CvT_ROC_k.png +0 -0
- roc_curves/CvT_ROC_l.png +0 -0
- training_curves/CvT_accuracy.png +0 -0
- training_curves/CvT_auc.png +0 -0
- training_curves/CvT_combined_metrics.png +3 -0
- training_curves/CvT_f1.png +0 -0
- training_curves/CvT_loss.png +0 -0
- training_curves/CvT_metrics.csv +38 -0
- training_metrics.csv +38 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
training_curves/CvT_combined_metrics.png filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
training_notebook_c1.ipynb filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
|
@@ -0,0 +1,166 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
tags:
|
| 4 |
+
- image-classification
|
| 5 |
+
- pytorch
|
| 6 |
+
- timm
|
| 7 |
+
- cvt
|
| 8 |
+
- vision-transformer
|
| 9 |
+
- transformer
|
| 10 |
+
- gravitational-lensing
|
| 11 |
+
- strong-lensing
|
| 12 |
+
- astronomy
|
| 13 |
+
- astrophysics
|
| 14 |
+
datasets:
|
| 15 |
+
- parlange/gravit-c21-j24
|
| 16 |
+
metrics:
|
| 17 |
+
- accuracy
|
| 18 |
+
- auc
|
| 19 |
+
- f1
|
| 20 |
+
paper:
|
| 21 |
+
- title: "GraViT: A Gravitational Lens Discovery Toolkit with Vision Transformers"
|
| 22 |
+
url: "https://arxiv.org/abs/2509.00226"
|
| 23 |
+
authors: "Parlange et al."
|
| 24 |
+
model-index:
|
| 25 |
+
- name: CvT-c1
|
| 26 |
+
results:
|
| 27 |
+
- task:
|
| 28 |
+
type: image-classification
|
| 29 |
+
name: Strong Gravitational Lens Discovery
|
| 30 |
+
dataset:
|
| 31 |
+
type: common-test-sample
|
| 32 |
+
name: Common Test Sample (More et al. 2024)
|
| 33 |
+
metrics:
|
| 34 |
+
- type: accuracy
|
| 35 |
+
value: 0.6955
|
| 36 |
+
name: Average Accuracy
|
| 37 |
+
- type: auc
|
| 38 |
+
value: 0.7354
|
| 39 |
+
name: Average AUC-ROC
|
| 40 |
+
- type: f1
|
| 41 |
+
value: 0.4422
|
| 42 |
+
name: Average F1-Score
|
| 43 |
+
---
|
| 44 |
+
|
| 45 |
+
# 🌌 cvt-gravit-c1
|
| 46 |
+
|
| 47 |
+
🔭 This model is part of **GraViT**: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery
|
| 48 |
+
|
| 49 |
+
🔗 **GitHub Repository**: [https://github.com/parlange/gravit](https://github.com/parlange/gravit)
|
| 50 |
+
|
| 51 |
+
## 🛰️ Model Details
|
| 52 |
+
|
| 53 |
+
- **🤖 Model Type**: CvT
|
| 54 |
+
- **🧪 Experiment**: C1 - C21+J24-classification-head
|
| 55 |
+
- **🌌 Dataset**: C21+J24
|
| 56 |
+
- **🪐 Fine-tuning Strategy**: classification-head
|
| 57 |
+
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
## 💻 Quick Start
|
| 61 |
+
|
| 62 |
+
```python
|
| 63 |
+
import torch
|
| 64 |
+
import timm
|
| 65 |
+
|
| 66 |
+
# Load the model directly from the Hub
|
| 67 |
+
model = timm.create_model(
|
| 68 |
+
'hf-hub:parlange/cvt-gravit-c1',
|
| 69 |
+
pretrained=True
|
| 70 |
+
)
|
| 71 |
+
model.eval()
|
| 72 |
+
|
| 73 |
+
# Example inference
|
| 74 |
+
dummy_input = torch.randn(1, 3, 224, 224)
|
| 75 |
+
with torch.no_grad():
|
| 76 |
+
output = model(dummy_input)
|
| 77 |
+
predictions = torch.softmax(output, dim=1)
|
| 78 |
+
print(f"Lens probability: {predictions[0][1]:.4f}")
|
| 79 |
+
```
|
| 80 |
+
|
| 81 |
+
## ⚡️ Training Configuration
|
| 82 |
+
|
| 83 |
+
**Training Dataset:** C21+J24 (Cañameras et al. 2021 + Jaelani et al. 2024)
|
| 84 |
+
**Fine-tuning Strategy:** classification-head
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
| 🔧 Parameter | 📝 Value |
|
| 88 |
+
|--------------|----------|
|
| 89 |
+
| Batch Size | 192 |
|
| 90 |
+
| Learning Rate | AdamW with ReduceLROnPlateau |
|
| 91 |
+
| Epochs | 100 |
|
| 92 |
+
| Patience | 10 |
|
| 93 |
+
| Optimizer | AdamW |
|
| 94 |
+
| Scheduler | ReduceLROnPlateau |
|
| 95 |
+
| Image Size | 224x224 |
|
| 96 |
+
| Fine Tune Mode | classification_head |
|
| 97 |
+
| Stochastic Depth Probability | 0.1 |
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
## 📈 Training Curves
|
| 101 |
+
|
| 102 |
+

|
| 103 |
+
|
| 104 |
+
|
| 105 |
+
## 🏁 Final Epoch Training Metrics
|
| 106 |
+
|
| 107 |
+
| Metric | Training | Validation |
|
| 108 |
+
|:---------:|:-----------:|:-------------:|
|
| 109 |
+
| 📉 Loss | 0.5187 | 0.4767 |
|
| 110 |
+
| 🎯 Accuracy | 0.7170 | 0.7624 |
|
| 111 |
+
| 📊 AUC-ROC | 0.8122 | 0.8618 |
|
| 112 |
+
| ⚖️ F1 Score | 0.7171 | 0.7779 |
|
| 113 |
+
|
| 114 |
+
|
| 115 |
+
## ☑️ Evaluation Results
|
| 116 |
+
|
| 117 |
+
### ROC Curves and Confusion Matrices
|
| 118 |
+
|
| 119 |
+
Performance across all test datasets (a through l) in the Common Test Sample (More et al. 2024):
|
| 120 |
+
|
| 121 |
+

|
| 122 |
+

|
| 123 |
+

|
| 124 |
+

|
| 125 |
+

|
| 126 |
+

|
| 127 |
+

|
| 128 |
+

|
| 129 |
+

|
| 130 |
+

|
| 131 |
+

|
| 132 |
+

|
| 133 |
+
|
| 134 |
+
### 📋 Performance Summary
|
| 135 |
+
|
| 136 |
+
Average performance across 12 test datasets from the Common Test Sample (More et al. 2024):
|
| 137 |
+
|
| 138 |
+
| Metric | Value |
|
| 139 |
+
|-----------|----------|
|
| 140 |
+
| 🎯 Average Accuracy | 0.6955 |
|
| 141 |
+
| 📈 Average AUC-ROC | 0.7354 |
|
| 142 |
+
| ⚖️ Average F1-Score | 0.4422 |
|
| 143 |
+
|
| 144 |
+
|
| 145 |
+
## 📘 Citation
|
| 146 |
+
|
| 147 |
+
If you use this model in your research, please cite:
|
| 148 |
+
|
| 149 |
+
```bibtex
|
| 150 |
+
@misc{parlange2025gravit,
|
| 151 |
+
title={GraViT: Transfer Learning with Vision Transformers and MLP-Mixer for Strong Gravitational Lens Discovery},
|
| 152 |
+
author={René Parlange and Juan C. Cuevas-Tello and Octavio Valenzuela and Omar de J. Cabrera-Rosas and Tomás Verdugo and Anupreeta More and Anton T. Jaelani},
|
| 153 |
+
year={2025},
|
| 154 |
+
eprint={2509.00226},
|
| 155 |
+
archivePrefix={arXiv},
|
| 156 |
+
primaryClass={cs.CV},
|
| 157 |
+
url={https://arxiv.org/abs/2509.00226},
|
| 158 |
+
}
|
| 159 |
+
```
|
| 160 |
+
|
| 161 |
+
---
|
| 162 |
+
|
| 163 |
+
|
| 164 |
+
## Model Card Contact
|
| 165 |
+
|
| 166 |
+
For questions about this model, please contact the author through: https://github.com/parlange/
|
config.json
ADDED
|
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architecture": "cvt_13_224",
|
| 3 |
+
"num_classes": 2,
|
| 4 |
+
"num_features": 1000,
|
| 5 |
+
"global_pool": "avg",
|
| 6 |
+
"crop_pct": 0.875,
|
| 7 |
+
"interpolation": "bicubic",
|
| 8 |
+
"mean": [
|
| 9 |
+
0.485,
|
| 10 |
+
0.456,
|
| 11 |
+
0.406
|
| 12 |
+
],
|
| 13 |
+
"std": [
|
| 14 |
+
0.229,
|
| 15 |
+
0.224,
|
| 16 |
+
0.225
|
| 17 |
+
],
|
| 18 |
+
"first_conv": "conv1",
|
| 19 |
+
"classifier": "fc",
|
| 20 |
+
"input_size": [
|
| 21 |
+
3,
|
| 22 |
+
224,
|
| 23 |
+
224
|
| 24 |
+
],
|
| 25 |
+
"pool_size": [
|
| 26 |
+
7,
|
| 27 |
+
7
|
| 28 |
+
],
|
| 29 |
+
"pretrained_cfg": {
|
| 30 |
+
"tag": "gravit_c1",
|
| 31 |
+
"custom_load": false,
|
| 32 |
+
"input_size": [
|
| 33 |
+
3,
|
| 34 |
+
224,
|
| 35 |
+
224
|
| 36 |
+
],
|
| 37 |
+
"fixed_input_size": true,
|
| 38 |
+
"interpolation": "bicubic",
|
| 39 |
+
"crop_pct": 0.875,
|
| 40 |
+
"crop_mode": "center",
|
| 41 |
+
"mean": [
|
| 42 |
+
0.485,
|
| 43 |
+
0.456,
|
| 44 |
+
0.406
|
| 45 |
+
],
|
| 46 |
+
"std": [
|
| 47 |
+
0.229,
|
| 48 |
+
0.224,
|
| 49 |
+
0.225
|
| 50 |
+
],
|
| 51 |
+
"num_classes": 2,
|
| 52 |
+
"pool_size": [
|
| 53 |
+
7,
|
| 54 |
+
7
|
| 55 |
+
],
|
| 56 |
+
"first_conv": "conv1",
|
| 57 |
+
"classifier": "fc"
|
| 58 |
+
},
|
| 59 |
+
"model_name": "cvt_gravit_c1",
|
| 60 |
+
"experiment": "c1",
|
| 61 |
+
"training_strategy": "classification-head",
|
| 62 |
+
"dataset": "C21+J24",
|
| 63 |
+
"hyperparameters": {
|
| 64 |
+
"batch_size": "192",
|
| 65 |
+
"learning_rate": "AdamW with ReduceLROnPlateau",
|
| 66 |
+
"epochs": "100",
|
| 67 |
+
"patience": "10",
|
| 68 |
+
"optimizer": "AdamW",
|
| 69 |
+
"scheduler": "ReduceLROnPlateau",
|
| 70 |
+
"image_size": "224x224",
|
| 71 |
+
"fine_tune_mode": "classification_head",
|
| 72 |
+
"stochastic_depth_probability": "0.1"
|
| 73 |
+
},
|
| 74 |
+
"hf_hub_id": "parlange/cvt-gravit-c1",
|
| 75 |
+
"license": "apache-2.0"
|
| 76 |
+
}
|
confusion_matrices/CvT_Confusion_Matrix_a.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_b.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_c.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_d.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_e.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_f.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_g.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_h.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_i.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_j.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_k.png
ADDED
|
confusion_matrices/CvT_Confusion_Matrix_l.png
ADDED
|
cvt-gravit-c1.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d90b10f1f52fb487f40690e541fa23f7199b45e424f342243533bd98b7bdabd
|
| 3 |
+
size 125471131
|
evaluation_results.csv
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Model,Dataset,Loss,Accuracy,AUCROC,F1
|
| 2 |
+
ViT,a,0.4575504291374072,0.7920353982300885,0.8256229797574562,0.5688073394495413
|
| 3 |
+
ViT,b,0.35027294130266856,0.854133920150896,0.8602099447513812,0.34831460674157305
|
| 4 |
+
ViT,c,0.5165887196384036,0.7544797233574347,0.786182320441989,0.24101068999028183
|
| 5 |
+
ViT,d,0.2510994746836054,0.9016032694121345,0.9034125230202578,0.44206773618538325
|
| 6 |
+
ViT,e,0.4143968988721117,0.8177826564215148,0.8443351244986,0.5990338164251208
|
| 7 |
+
ViT,f,0.3670418868224962,0.841640022569118,0.8478622806743998,0.12836438923395446
|
| 8 |
+
ViT,g,0.2497027156352997,0.9033333333333333,0.9734338888888888,0.9069618222649984
|
| 9 |
+
ViT,h,0.3378777855237325,0.8505,0.9529537222222222,0.8630743397954511
|
| 10 |
+
ViT,i,0.19712424822648367,0.9285,0.9841957777777779,0.9294755877034359
|
| 11 |
+
ViT,j,0.7854293506145478,0.5983333333333334,0.6951524444444445,0.4527702089009991
|
| 12 |
+
ViT,k,0.7328508781194687,0.6235,0.7771344444444446,0.46884552080884084
|
| 13 |
+
ViT,l,0.487411656158395,0.7679451725381748,0.8238313924596774,0.6716570261993875
|
| 14 |
+
MLP-Mixer,a,0.3996943971224591,0.8086283185840708,0.8391867831243361,0.5685785536159601
|
| 15 |
+
MLP-Mixer,b,0.3412993300058526,0.8704809808236403,0.8364769797421732,0.35625
|
| 16 |
+
MLP-Mixer,c,0.3957880755634062,0.8242690977679975,0.80653591160221,0.289707750952986
|
| 17 |
+
MLP-Mixer,d,0.30740130081636957,0.8868280414963848,0.8505101289134438,0.3877551020408163
|
| 18 |
+
MLP-Mixer,e,0.4771155259938978,0.7793633369923162,0.7879361235147202,0.5314685314685315
|
| 19 |
+
MLP-Mixer,f,0.3381275081289247,0.8648674064321986,0.8287089328554598,0.13693693693693693
|
| 20 |
+
MLP-Mixer,g,0.26878669277826944,0.9036666666666666,0.9679408888888887,0.9054319371727748
|
| 21 |
+
MLP-Mixer,h,0.29767482439676923,0.8791666666666667,0.9592756666666666,0.8841667998082761
|
| 22 |
+
MLP-Mixer,i,0.2508150982062022,0.9123333333333333,0.9736637777777777,0.9132013201320132
|
| 23 |
+
MLP-Mixer,j,0.6541107538541158,0.6641666666666667,0.7743557777777779,0.5689839572192513
|
| 24 |
+
MLP-Mixer,k,0.6361391656398773,0.6728333333333333,0.7940376666666666,0.5753839498161367
|
| 25 |
+
MLP-Mixer,l,0.4342376798394926,0.7992064446314777,0.8634144024748618,0.716034687978235
|
| 26 |
+
CvT,a,0.6301151535152334,0.6548672566371682,0.7049165921612678,0.41353383458646614
|
| 27 |
+
CvT,b,0.6147656698190952,0.6692863879283244,0.7044972375690608,0.17295597484276728
|
| 28 |
+
CvT,c,0.6944268911336811,0.6001257466205596,0.6591786372007367,0.14745308310991956
|
| 29 |
+
CvT,d,0.2650476947487693,0.9000314366551398,0.8999226519337017,0.40892193308550184
|
| 30 |
+
CvT,e,0.6509704625828991,0.6465422612513722,0.6885869976538257,0.4059040590405904
|
| 31 |
+
CvT,f,0.5340885097132559,0.7186383298852737,0.7464956265694429,0.0684931506849315
|
| 32 |
+
CvT,g,0.42030701192220055,0.8028333333333333,0.9186497222222223,0.825490485322319
|
| 33 |
+
CvT,h,0.4625407474835714,0.7661666666666667,0.8991485555555555,0.799542791827404
|
| 34 |
+
CvT,i,0.2348982002735138,0.9251666666666667,0.9787946666666666,0.9257237386269644
|
| 35 |
+
CvT,j,1.2669325167338052,0.438,0.3431898888888889,0.265359477124183
|
| 36 |
+
CvT,k,1.08152370317777,0.5603333333333333,0.6091516666666666,0.3158713692946058
|
| 37 |
+
CvT,l,0.7296624132198569,0.6642419141517374,0.6727565192585612,0.5574134242016008
|
| 38 |
+
Swin,a,0.3518127097492724,0.8573008849557522,0.92351161137984,0.7074829931972789
|
| 39 |
+
Swin,b,0.3482047942909419,0.8509902546369066,0.9245285451197054,0.3969465648854962
|
| 40 |
+
Swin,c,0.3979874673372393,0.8255265639735933,0.9125395948434623,0.35986159169550175
|
| 41 |
+
Swin,d,0.17529730912425615,0.9415278214397989,0.9671712707182321,0.6265060240963856
|
| 42 |
+
Swin,e,0.5681350649908005,0.7387486278814489,0.8579505032922122,0.5672727272727273
|
| 43 |
+
Swin,f,0.329795688788694,0.8606356968215159,0.9286062369152528,0.17391304347826086
|
| 44 |
+
Swin,g,0.2527379404703776,0.9035,0.9724114444444445,0.9083715777812945
|
| 45 |
+
Swin,h,0.2791310551166534,0.89,0.9682955555555556,0.896875
|
| 46 |
+
Swin,i,0.16106815997759502,0.9515,0.9900552222222222,0.9517492953075776
|
| 47 |
+
Swin,j,0.907479268391927,0.5803333333333334,0.6390204444444445,0.42511415525114155
|
| 48 |
+
Swin,k,0.8158094821770986,0.6283333333333333,0.7901601111111112,0.4550342130987292
|
| 49 |
+
Swin,l,0.503966703916006,0.7787062642779848,0.8190682128691764,0.682535575679172
|
| 50 |
+
CaiT,a,0.42890300766556666,0.8130530973451328,0.8644880523906682,0.6060606060606061
|
| 51 |
+
CaiT,b,0.3192856568973119,0.8893429739075762,0.903804788213628,0.42483660130718953
|
| 52 |
+
CaiT,c,0.48953707897974164,0.7730273498899717,0.8254953959484346,0.26476578411405294
|
| 53 |
+
CaiT,d,0.31145547537899587,0.867966048412449,0.9025340699815839,0.38235294117647056
|
| 54 |
+
CaiT,e,0.4548946529397849,0.7870472008781558,0.8492999318852645,0.5726872246696035
|
| 55 |
+
CaiT,f,0.3738367850840574,0.8445552003009216,0.8744395460236903,0.1359121798222687
|
| 56 |
+
CaiT,g,0.3011642297108968,0.8948333333333334,0.9556216666666666,0.8943225590353374
|
| 57 |
+
CaiT,h,0.39142585277557373,0.8331666666666667,0.9202581666666666,0.8421384639646743
|
| 58 |
+
CaiT,i,0.2970129141012828,0.8835,0.9525210000000002,0.8842523596621957
|
| 59 |
+
CaiT,j,0.7035526345570882,0.661,0.7364068888888889,0.5547285464098074
|
| 60 |
+
CaiT,k,0.6994013291200002,0.6496666666666666,0.7619980555555556,0.5465918895599655
|
| 61 |
+
CaiT,l,0.4910164130440666,0.7766021401947818,0.8205470707913558,0.6864135021097046
|
| 62 |
+
DeiT,a,0.4880231645254962,0.7798672566371682,0.8260661913604304,0.5664488017429193
|
| 63 |
+
DeiT,b,0.31241170634304494,0.8758252121974222,0.8908674033149171,0.3969465648854962
|
| 64 |
+
DeiT,c,0.5076334938917547,0.7563659226658284,0.8106298342541435,0.25120772946859904
|
| 65 |
+
DeiT,d,0.21307693347358284,0.9264382269726501,0.9316758747697976,0.5263157894736842
|
| 66 |
+
DeiT,e,0.41276343642814756,0.8155872667398463,0.8614016498902596,0.6074766355140186
|
| 67 |
+
DeiT,f,0.347494895568754,0.852642467556893,0.8730114223466999,0.1423097974822113
|
| 68 |
+
DeiT,g,0.21456254084904988,0.9195,0.9812000000000001,0.9221595487510073
|
| 69 |
+
DeiT,h,0.31806263717015587,0.8561666666666666,0.9641542222222222,0.8689445709946849
|
| 70 |
+
DeiT,i,0.1618985648949941,0.9463333333333334,0.9902026666666668,0.9467240238252813
|
| 71 |
+
DeiT,j,0.7059943490028381,0.6493333333333333,0.7624447777777776,0.5410122164048866
|
| 72 |
+
DeiT,k,0.6533303724129995,0.6761666666666667,0.8324687777777777,0.5607054035722361
|
| 73 |
+
DeiT,l,0.4469874652336176,0.7916315979319466,0.8543558192306258,0.7094232059020792
|
| 74 |
+
DeiT3,a,0.4202683992617953,0.8163716814159292,0.8467825130097888,0.5829145728643216
|
| 75 |
+
DeiT3,b,0.262984538234768,0.9012889028607356,0.8990220994475138,0.4249084249084249
|
| 76 |
+
DeiT3,c,0.472756382524124,0.7783715812637535,0.7983001841620626,0.24759871931696906
|
| 77 |
+
DeiT3,d,0.33147037092037673,0.862936183590066,0.8655248618784529,0.3473053892215569
|
| 78 |
+
DeiT3,e,0.38950261964363536,0.814489571899012,0.86056913645652,0.5785536159600998
|
| 79 |
+
DeiT3,f,0.34437218432882033,0.8561218732367877,0.8542026846822373,0.13166855845629966
|
| 80 |
+
DeiT3,g,0.21329958017667133,0.9278333333333333,0.9798244444444444,0.9286067600989283
|
| 81 |
+
DeiT3,h,0.32451361187299094,0.8626666666666667,0.9527946111111111,0.872366790582404
|
| 82 |
+
DeiT3,i,0.24960849436124166,0.9075,0.9715454444444445,0.9102957814772911
|
| 83 |
+
DeiT3,j,0.5672995191415151,0.7125,0.8462789444444444,0.6385920804525456
|
| 84 |
+
DeiT3,k,0.6036084276040395,0.6921666666666667,0.800755,0.62267620020429
|
| 85 |
+
DeiT3,l,0.4151913604799968,0.8082241192737766,0.8763639318146963,0.7364072054205917
|
| 86 |
+
Twins_SVT,a,0.4067131507713183,0.8396017699115044,0.8804971611532671,0.6472019464720195
|
| 87 |
+
Twins_SVT,b,0.3369882810569267,0.8682804149638479,0.8973793738489871,0.3883211678832117
|
| 88 |
+
Twins_SVT,c,0.44650013615072465,0.808236403646652,0.8531749539594844,0.3036529680365297
|
| 89 |
+
Twins_SVT,d,0.11182147282515634,0.9789374410562716,0.9822780847145488,0.7987987987987988
|
| 90 |
+
Twins_SVT,e,0.48889580673234523,0.7694840834248079,0.8373949897827897,0.5588235294117647
|
| 91 |
+
Twins_SVT,f,0.30908069689963935,0.8816061688922324,0.903702339279268,0.17442622950819672
|
| 92 |
+
Twins_SVT,g,0.25526987624168396,0.9126666666666666,0.9751330000000001,0.9157285300739788
|
| 93 |
+
Twins_SVT,h,0.3133294117450714,0.8808333333333334,0.9663294444444444,0.8884381338742393
|
| 94 |
+
Twins_SVT,i,0.13589392483234405,0.9713333333333334,0.9974112222222222,0.9706784861916127
|
| 95 |
+
Twins_SVT,j,1.2727711579799652,0.49183333333333334,0.44312866666666667,0.1743839696723531
|
| 96 |
+
Twins_SVT,k,1.1533952040274937,0.5505,0.7285452777777779,0.19275665968272973
|
| 97 |
+
Twins_SVT,l,0.6311305313460368,0.7541180714199832,0.7295359094064597,0.6175425472227417
|
| 98 |
+
Twins_PCPVT,a,0.47225049058947943,0.7533185840707964,0.8392020662830595,0.5363825363825364
|
| 99 |
+
Twins_PCPVT,b,0.3277859269766941,0.8707953473750393,0.9017642725598527,0.38565022421524664
|
| 100 |
+
Twins_PCPVT,c,0.5325269242833524,0.7318453316567117,0.8134834254143647,0.23222322232223222
|
| 101 |
+
Twins_PCPVT,d,0.34038559299375754,0.8726815466834329,0.8983720073664826,0.3891402714932127
|
| 102 |
+
Twins_PCPVT,e,0.43271679039451605,0.8024149286498353,0.8658064027851359,0.589041095890411
|
| 103 |
+
Twins_PCPVT,f,0.40193462960922294,0.8247131841263871,0.8686157929759784,0.12158341187558906
|
| 104 |
+
Twins_PCPVT,g,0.28006591550509136,0.9015,0.9626466666666665,0.9035417006691693
|
| 105 |
+
Twins_PCPVT,h,0.3886127746899923,0.8278333333333333,0.9278962222222221,0.8427462323032425
|
| 106 |
+
Twins_PCPVT,i,0.28674582862854003,0.9025,0.9641564444444444,0.9044273811468714
|
| 107 |
+
Twins_PCPVT,j,0.5901691876252493,0.6993333333333334,0.8068135555555556,0.6328856328856329
|
| 108 |
+
Twins_PCPVT,k,0.5968491020202636,0.7003333333333334,0.7921438888888889,0.6336593317033414
|
| 109 |
+
Twins_PCPVT,l,0.4564983546913021,0.7871227606107971,0.8525991250300863,0.7154680594616312
|
| 110 |
+
PiT,a,0.3924222976233052,0.831858407079646,0.8797864942726362,0.6513761467889908
|
| 111 |
+
PiT,b,0.31899622544594736,0.8758252121974222,0.9015782688766113,0.4182621502209131
|
| 112 |
+
PiT,c,0.454398755605998,0.7959761081420936,0.8498250460405157,0.30439442658092175
|
| 113 |
+
PiT,d,0.14505015389513048,0.9597610814209369,0.9646924493554327,0.6893203883495146
|
| 114 |
+
PiT,e,0.3891645081087734,0.827661909989023,0.8838870809051691,0.6439909297052154
|
| 115 |
+
PiT,f,0.3070308875887381,0.8753996614632311,0.9020960965500401,0.17650714729645742
|
| 116 |
+
PiT,g,0.22340119377772014,0.9178333333333333,0.9791551111111111,0.9207268049525648
|
| 117 |
+
PiT,h,0.2951871022383372,0.8755,0.9678674444444444,0.8845975590916113
|
| 118 |
+
PiT,i,0.13118078410625458,0.9623333333333334,0.9941626666666668,0.9620295698924731
|
| 119 |
+
PiT,j,1.1235656504631042,0.5436666666666666,0.5612956111111111,0.3110216406643181
|
| 120 |
+
PiT,k,1.0313452566862107,0.5881666666666666,0.7460636666666667,0.33342325330455896
|
| 121 |
+
PiT,l,0.5711348351591465,0.7689070578333533,0.7784629383878745,0.6533814247069432
|
| 122 |
+
Ensemble,a,,0.8628318584070797,0.9054087098721564,0.6836734693877551
|
| 123 |
+
Ensemble,b,,0.9182646966362779,0.931486187845304,0.5075757575757576
|
| 124 |
+
Ensemble,c,,0.8340144608613643,0.8768360957642726,0.33668341708542715
|
| 125 |
+
Ensemble,d,,0.9682489783087079,0.9682486187845304,0.7262872628726287
|
| 126 |
+
Ensemble,e,,0.8309549945115258,0.8953757662907742,0.6350710900473934
|
| 127 |
+
Ensemble,f,,0.9079368064698138,0.922026931389281,0.214915797914996
|
| 128 |
+
Ensemble,g,,0.9488333333333333,0.9909529999999999,0.9498284033338781
|
| 129 |
+
Ensemble,h,,0.9041666666666667,0.9819358888888887,0.9099733834351026
|
| 130 |
+
Ensemble,i,,0.9753333333333334,0.9972660000000001,0.9751677852348993
|
| 131 |
+
Ensemble,j,,0.5953333333333334,0.7440818888888889,0.39269634817408705
|
| 132 |
+
Ensemble,k,,0.6218333333333333,0.8487693333333335,0.408960666840323
|
| 133 |
+
Ensemble,l,,0.802332571840808,0.8579947764427092,0.6993965990126165
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7eb14026cd0166e64202650df4d632785fd86a388f21a29685474cce09d7588
|
| 3 |
+
size 125239576
|
pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d90b10f1f52fb487f40690e541fa23f7199b45e424f342243533bd98b7bdabd
|
| 3 |
+
size 125471131
|
roc_confusion_matrix/CvT_roc_confusion_matrix_a.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_b.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_c.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_d.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_e.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_f.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_g.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_h.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_i.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_j.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_k.png
ADDED
|
roc_confusion_matrix/CvT_roc_confusion_matrix_l.png
ADDED
|
roc_curves/CvT_ROC_a.png
ADDED
|
roc_curves/CvT_ROC_b.png
ADDED
|
roc_curves/CvT_ROC_c.png
ADDED
|
roc_curves/CvT_ROC_d.png
ADDED
|
roc_curves/CvT_ROC_e.png
ADDED
|
roc_curves/CvT_ROC_f.png
ADDED
|
roc_curves/CvT_ROC_g.png
ADDED
|
roc_curves/CvT_ROC_h.png
ADDED
|
roc_curves/CvT_ROC_i.png
ADDED
|
roc_curves/CvT_ROC_j.png
ADDED
|
roc_curves/CvT_ROC_k.png
ADDED
|
roc_curves/CvT_ROC_l.png
ADDED
|
training_curves/CvT_accuracy.png
ADDED
|
training_curves/CvT_auc.png
ADDED
|
training_curves/CvT_combined_metrics.png
ADDED
|
Git LFS Details
|
training_curves/CvT_f1.png
ADDED
|
training_curves/CvT_loss.png
ADDED
|
training_curves/CvT_metrics.csv
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.6177887970854768,0.5793780776571602,0.6403957314361939,0.6909620991253644,0.7103180007841088,0.7776202942651447,0.638594078975637,0.7111716621253406
|
| 3 |
+
2,0.5600605519329119,0.5366562573624769,0.6928036392242706,0.717201166180758,0.7802340546587921,0.8174856989859668,0.6933438038820697,0.7309292649098474
|
| 4 |
+
3,0.5450726019238122,0.5260146274163494,0.7018332934295585,0.7295918367346939,0.7923471509698521,0.8348158505384662,0.7023728234892455,0.7564018384766907
|
| 5 |
+
4,0.5377111861895539,0.515728648654226,0.7062369600164176,0.7325072886297376,0.7982963378804904,0.8381446081139662,0.7050490654807386,0.7548430193720775
|
| 6 |
+
5,0.5344073905922644,0.5070405717146292,0.7074169716455176,0.7427113702623906,0.800385995657201,0.8475072036311402,0.7060319424040139,0.7666886979510905
|
| 7 |
+
6,0.5278249657072128,0.49534227531783437,0.7135222492047748,0.7485422740524781,0.8062911270525801,0.8547554590349259,0.7119731084345635,0.7686116700201208
|
| 8 |
+
7,0.52661867241537,0.48591154153423477,0.7119831036016007,0.7514577259475219,0.8061202233273697,0.8569016736223852,0.7117982767619553,0.7656357388316152
|
| 9 |
+
8,0.5269551495237357,0.4852170724389157,0.711350343742518,0.7623906705539358,0.8059306758146515,0.8547235845608547,0.7140230936708432,0.7742382271468145
|
| 10 |
+
9,0.5241837057421633,0.5155514045116157,0.7129322433902248,0.7223032069970845,0.8079216695342756,0.8542667170991679,0.7111042269034834,0.7599243856332704
|
| 11 |
+
10,0.521815941439537,0.4836132715811882,0.7153179190751445,0.7543731778425656,0.8101999091031789,0.8620069018861188,0.7173457172693082,0.7754830113257828
|
| 12 |
+
11,0.5230052669707127,0.480255791553603,0.7142148647262031,0.7616618075801749,0.8091502292385969,0.8604992392625521,0.7156687594643799,0.7758738862234407
|
| 13 |
+
12,0.5216906908562527,0.48589079173243777,0.7142576187707357,0.7631195335276968,0.8100854571825495,0.855647944308919,0.7135546583691208,0.77906186267845
|
| 14 |
+
13,0.5203107965177227,0.4852381767406408,0.7161131443034511,0.7660349854227405,0.8113744171196645,0.8569664850529967,0.7149676333728258,0.7808873720136519
|
| 15 |
+
14,0.5208332048752697,0.48790633791389687,0.7156257481957793,0.7638483965014577,0.8107133624095892,0.854133906790538,0.7157108297786858,0.7786885245901639
|
| 16 |
+
15,0.5203544425219017,0.49315251284020983,0.7173444607859903,0.7660349854227405,0.8115763053532916,0.8530108628207634,0.717392791190753,0.7811860940695297
|
| 17 |
+
16,0.5220811562625076,0.4836661697476668,0.7158224168006293,0.7667638483965015,0.8098662726954311,0.8597087523055869,0.7162445996482301,0.782312925170068
|
| 18 |
+
17,0.517030687437675,0.47370702149916666,0.7187895474911926,0.7602040816326531,0.814179251339733,0.863257443752178,0.7192959994537338,0.7729468599033816
|
| 19 |
+
18,0.5199297091511746,0.4831294265328621,0.7165321339398707,0.7623906705539358,0.8117216783173604,0.8624372072860798,0.7160440953514865,0.7826666666666666
|
| 20 |
+
19,0.5194016830492318,0.49196437235823876,0.7185073707972774,0.75,0.8126051711913708,0.8589958265688616,0.7164952892747033,0.7744904667981591
|
| 21 |
+
20,0.5196415349980257,0.4765674379754692,0.715326469884051,0.7602040816326531,0.8115310097850255,0.8650232896157213,0.7160015696176616,0.7781523937963587
|
| 22 |
+
21,0.5198505823339112,0.4802529356570007,0.7166860485001881,0.760932944606414,0.8116370341708854,0.8594314443811678,0.7168821403241932,0.7762619372442019
|
| 23 |
+
22,0.5190914454642646,0.4846779636321888,0.7167972090159729,0.7529154518950437,0.8121206001525717,0.8580831541279568,0.7169085594133032,0.769857433808554
|
| 24 |
+
23,0.5193884302193528,0.4794780606778648,0.7161729999657968,0.7529154518950437,0.812006350179254,0.8606756113524128,0.7157038242473556,0.7714093054619016
|
| 25 |
+
24,0.5207682174807287,0.4809463116239876,0.7147450148784075,0.7572886297376094,0.810329750257103,0.8642944266419604,0.7143493226927885,0.7778519012675117
|
| 26 |
+
25,0.5181710218936363,0.4774894813406919,0.7165406847487772,0.7696793002915452,0.8123784661044771,0.8585825208884054,0.7162106632880183,0.780250347705146
|
| 27 |
+
26,0.5183458806784481,0.4876498219396908,0.7175154769641208,0.7478134110787172,0.8127058757790775,0.8617763431903372,0.7175830939679935,0.7711640211640212
|
| 28 |
+
27,0.5177797832925942,0.484501314319605,0.71796011902726,0.7558309037900874,0.813426613012022,0.8633913165432772,0.718292537109475,0.7782925215089345
|
| 29 |
+
28,0.5183137668326064,0.485816216156017,0.716258508054862,0.7580174927113703,0.8127243228836496,0.8566169283206828,0.7165905111671008,0.7741496598639456
|
| 30 |
+
29,0.5166353127896617,0.5278777311564187,0.7189947669049492,0.728134110787172,0.8147622951632917,0.8542210303529992,0.7214740357151938,0.7673112913287585
|
| 31 |
+
30,0.5173938232590775,0.48899602212294196,0.7186783869754079,0.7558309037900874,0.8139822627778102,0.8526666185007947,0.7199237239077877,0.7719537100068073
|
| 32 |
+
31,0.5168906647594641,0.4785742049835861,0.7193539008790232,0.7645772594752187,0.8144676413133902,0.862389395574973,0.7201698368986009,0.7795221843003413
|
| 33 |
+
32,0.5155035974095677,0.48033515739718957,0.7180028730717926,0.7587463556851312,0.8152104316022155,0.8612217273414989,0.7195162401449239,0.7771043771043771
|
| 34 |
+
33,0.5188441814453824,0.48279932697382333,0.7155915449601532,0.7514577259475219,0.8120172104406536,0.8601496825302383,0.7168770588786081,0.7716008037508373
|
| 35 |
+
34,0.5178068923747143,0.48123584839762473,0.7180627287341382,0.7631195335276968,0.8133943546708526,0.8598245628947122,0.7182265673070349,0.77906186267845
|
| 36 |
+
35,0.516852009711083,0.5108656356703088,0.7180370763074186,0.739067055393586,0.8139105369712744,0.8489840542631046,0.7173531050443578,0.7666232073011734
|
| 37 |
+
36,0.5173892301459864,0.47859706979451305,0.7169425727673838,0.7674927113702624,0.8133955893132507,0.8577208476060145,0.7169449931166577,0.779848171152519
|
| 38 |
+
37,0.5186981641580009,0.4767029957938125,0.717028080856449,0.7623906705539358,0.812201361675534,0.8617561560234256,0.7170982329859716,0.7779291553133515
|
training_metrics.csv
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
epoch,train_loss,val_loss,train_accuracy,val_accuracy,train_auc,val_auc,train_f1,val_f1
|
| 2 |
+
1,0.6177887970854768,0.5793780776571602,0.6403957314361939,0.6909620991253644,0.7103180007841088,0.7776202942651447,0.638594078975637,0.7111716621253406
|
| 3 |
+
2,0.5600605519329119,0.5366562573624769,0.6928036392242706,0.717201166180758,0.7802340546587921,0.8174856989859668,0.6933438038820697,0.7309292649098474
|
| 4 |
+
3,0.5450726019238122,0.5260146274163494,0.7018332934295585,0.7295918367346939,0.7923471509698521,0.8348158505384662,0.7023728234892455,0.7564018384766907
|
| 5 |
+
4,0.5377111861895539,0.515728648654226,0.7062369600164176,0.7325072886297376,0.7982963378804904,0.8381446081139662,0.7050490654807386,0.7548430193720775
|
| 6 |
+
5,0.5344073905922644,0.5070405717146292,0.7074169716455176,0.7427113702623906,0.800385995657201,0.8475072036311402,0.7060319424040139,0.7666886979510905
|
| 7 |
+
6,0.5278249657072128,0.49534227531783437,0.7135222492047748,0.7485422740524781,0.8062911270525801,0.8547554590349259,0.7119731084345635,0.7686116700201208
|
| 8 |
+
7,0.52661867241537,0.48591154153423477,0.7119831036016007,0.7514577259475219,0.8061202233273697,0.8569016736223852,0.7117982767619553,0.7656357388316152
|
| 9 |
+
8,0.5269551495237357,0.4852170724389157,0.711350343742518,0.7623906705539358,0.8059306758146515,0.8547235845608547,0.7140230936708432,0.7742382271468145
|
| 10 |
+
9,0.5241837057421633,0.5155514045116157,0.7129322433902248,0.7223032069970845,0.8079216695342756,0.8542667170991679,0.7111042269034834,0.7599243856332704
|
| 11 |
+
10,0.521815941439537,0.4836132715811882,0.7153179190751445,0.7543731778425656,0.8101999091031789,0.8620069018861188,0.7173457172693082,0.7754830113257828
|
| 12 |
+
11,0.5230052669707127,0.480255791553603,0.7142148647262031,0.7616618075801749,0.8091502292385969,0.8604992392625521,0.7156687594643799,0.7758738862234407
|
| 13 |
+
12,0.5216906908562527,0.48589079173243777,0.7142576187707357,0.7631195335276968,0.8100854571825495,0.855647944308919,0.7135546583691208,0.77906186267845
|
| 14 |
+
13,0.5203107965177227,0.4852381767406408,0.7161131443034511,0.7660349854227405,0.8113744171196645,0.8569664850529967,0.7149676333728258,0.7808873720136519
|
| 15 |
+
14,0.5208332048752697,0.48790633791389687,0.7156257481957793,0.7638483965014577,0.8107133624095892,0.854133906790538,0.7157108297786858,0.7786885245901639
|
| 16 |
+
15,0.5203544425219017,0.49315251284020983,0.7173444607859903,0.7660349854227405,0.8115763053532916,0.8530108628207634,0.717392791190753,0.7811860940695297
|
| 17 |
+
16,0.5220811562625076,0.4836661697476668,0.7158224168006293,0.7667638483965015,0.8098662726954311,0.8597087523055869,0.7162445996482301,0.782312925170068
|
| 18 |
+
17,0.517030687437675,0.47370702149916666,0.7187895474911926,0.7602040816326531,0.814179251339733,0.863257443752178,0.7192959994537338,0.7729468599033816
|
| 19 |
+
18,0.5199297091511746,0.4831294265328621,0.7165321339398707,0.7623906705539358,0.8117216783173604,0.8624372072860798,0.7160440953514865,0.7826666666666666
|
| 20 |
+
19,0.5194016830492318,0.49196437235823876,0.7185073707972774,0.75,0.8126051711913708,0.8589958265688616,0.7164952892747033,0.7744904667981591
|
| 21 |
+
20,0.5196415349980257,0.4765674379754692,0.715326469884051,0.7602040816326531,0.8115310097850255,0.8650232896157213,0.7160015696176616,0.7781523937963587
|
| 22 |
+
21,0.5198505823339112,0.4802529356570007,0.7166860485001881,0.760932944606414,0.8116370341708854,0.8594314443811678,0.7168821403241932,0.7762619372442019
|
| 23 |
+
22,0.5190914454642646,0.4846779636321888,0.7167972090159729,0.7529154518950437,0.8121206001525717,0.8580831541279568,0.7169085594133032,0.769857433808554
|
| 24 |
+
23,0.5193884302193528,0.4794780606778648,0.7161729999657968,0.7529154518950437,0.812006350179254,0.8606756113524128,0.7157038242473556,0.7714093054619016
|
| 25 |
+
24,0.5207682174807287,0.4809463116239876,0.7147450148784075,0.7572886297376094,0.810329750257103,0.8642944266419604,0.7143493226927885,0.7778519012675117
|
| 26 |
+
25,0.5181710218936363,0.4774894813406919,0.7165406847487772,0.7696793002915452,0.8123784661044771,0.8585825208884054,0.7162106632880183,0.780250347705146
|
| 27 |
+
26,0.5183458806784481,0.4876498219396908,0.7175154769641208,0.7478134110787172,0.8127058757790775,0.8617763431903372,0.7175830939679935,0.7711640211640212
|
| 28 |
+
27,0.5177797832925942,0.484501314319605,0.71796011902726,0.7558309037900874,0.813426613012022,0.8633913165432772,0.718292537109475,0.7782925215089345
|
| 29 |
+
28,0.5183137668326064,0.485816216156017,0.716258508054862,0.7580174927113703,0.8127243228836496,0.8566169283206828,0.7165905111671008,0.7741496598639456
|
| 30 |
+
29,0.5166353127896617,0.5278777311564187,0.7189947669049492,0.728134110787172,0.8147622951632917,0.8542210303529992,0.7214740357151938,0.7673112913287585
|
| 31 |
+
30,0.5173938232590775,0.48899602212294196,0.7186783869754079,0.7558309037900874,0.8139822627778102,0.8526666185007947,0.7199237239077877,0.7719537100068073
|
| 32 |
+
31,0.5168906647594641,0.4785742049835861,0.7193539008790232,0.7645772594752187,0.8144676413133902,0.862389395574973,0.7201698368986009,0.7795221843003413
|
| 33 |
+
32,0.5155035974095677,0.48033515739718957,0.7180028730717926,0.7587463556851312,0.8152104316022155,0.8612217273414989,0.7195162401449239,0.7771043771043771
|
| 34 |
+
33,0.5188441814453824,0.48279932697382333,0.7155915449601532,0.7514577259475219,0.8120172104406536,0.8601496825302383,0.7168770588786081,0.7716008037508373
|
| 35 |
+
34,0.5178068923747143,0.48123584839762473,0.7180627287341382,0.7631195335276968,0.8133943546708526,0.8598245628947122,0.7182265673070349,0.77906186267845
|
| 36 |
+
35,0.516852009711083,0.5108656356703088,0.7180370763074186,0.739067055393586,0.8139105369712744,0.8489840542631046,0.7173531050443578,0.7666232073011734
|
| 37 |
+
36,0.5173892301459864,0.47859706979451305,0.7169425727673838,0.7674927113702624,0.8133955893132507,0.8577208476060145,0.7169449931166577,0.779848171152519
|
| 38 |
+
37,0.5186981641580009,0.4767029957938125,0.717028080856449,0.7623906705539358,0.812201361675534,0.8617561560234256,0.7170982329859716,0.7779291553133515
|