Parveshiiii commited on
Commit
31aedae
·
verified ·
1 Parent(s): 4350516

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +119 -0
README.md CHANGED
@@ -21,3 +21,122 @@ tags:
21
  />
22
  </p>
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  />
22
  </p>
23
 
24
+ ---
25
+
26
+ # XenArcAI/AIRealNet
27
+
28
+ **Model type:** Image Classification (Binary)
29
+ **Task:** AI-generated vs Human image detection
30
+ **Base model:** `Microsoft/swinv2-tiny-patch4-window16-256`
31
+ **Fine-tuned by:** `Parveshiiii/AI-vs-Real` dataset split
32
+
33
+ ---
34
+
35
+ ## Overview
36
+
37
+ In an era of rapidly advancing AI-generated imagery, deepfakes, and synthetic media, the need for reliable detection tools has never been higher. **XenArcAI/AIRealNet** is a binary image classifier explicitly designed to distinguish **AI-generated images** from **real human photographs**. This model is optimized to detect conventional AI-generated content while adhering to strict privacy standards—avoiding personal or sensitive images.
38
+
39
+ * **Class 0:** AI-generated image
40
+ * **Class 1:** Real human image
41
+
42
+ By leveraging the robust **SwinV2 Tiny** architecture as its backbone, AIRealNet achieves a high degree of accuracy while remaining lightweight enough for practical deployment.
43
+
44
+ ---
45
+
46
+ ## Key Features
47
+
48
+ 1. **High Accuracy on Public Datasets:**
49
+ Despite using a **14k-image fine-tuning split(Part of main fine tuning split)**, AIRealNet demonstrates exceptional accuracy and robustness in detecting AI-generated images.
50
+
51
+ 2. **Balanced Training Split:**
52
+ The dataset contains a balanced number of AI-generated and real images, ensuring unbiased training and minimizing class imbalance issues.
53
+
54
+ * **AI-Generated:** 60%
55
+ * **Human-Images:** 40%
56
+
57
+ 4. **Ethical Design:**
58
+ No personal photos were included, even if edited or AI-modified, respecting privacy and ethical AI principles.
59
+
60
+ 5. **Fast and Scalable:**
61
+ Based on a transformer vision model, AIRealNet can be deployed efficiently in both research and production environments.
62
+
63
+ ---
64
+
65
+ ## Training Data
66
+
67
+ * **Dataset:** `Parveshiiii/AI-vs-Real` (open-sourced subset of main dataset )
68
+ * **Size:** 14k images (balanced between AI and human)
69
+ * **Split:** Used the train split for fine-tuning; validation performed on a separate balanced subset.
70
+ * **Notes:** Images sourced from public datasets and AI generation tools. Edited personal photos were intentionally excluded.
71
+
72
+ ---
73
+
74
+ ## Limitations
75
+
76
+ While AIRealNet performs exceptionally well on typical AI-generated images, users should note:
77
+
78
+ 1. **Subtle Edits:** The model struggles with nano-scale edits or ultra-precise modifications, like “nano banana” edits.
79
+ 2. **Edited Personal Images(over precise):** Images of real people that have been AI-modified are **not detected**, aligning with privacy and ethical guidelines.
80
+ 3. **Domain Generalization:** Performance may vary on images from completely unseen AI generators or extremely unconventional content.
81
+
82
+ ---
83
+
84
+ ## Performance Metrics
85
+
86
+ > Metrics shown are from **Epoch 2**, chosen to illustrate stable performance after fine-tuning.
87
+
88
+ <p align="center">
89
+ <img
90
+ src="https://cdn-uploads.huggingface.co/production/uploads/677fcdf29b9a9863eba3f29f/3NVa0KLX0iAxTP2e6IlGH.png"
91
+ alt="AIRealNet Banner"
92
+ width="90%"
93
+ style="border-radius:15px;"
94
+ />
95
+ </p>
96
+
97
+ **Note:** Extremely low loss and high accuracy are due to the controlled dataset environment. Real-world performance may be lower depending on the image domain.(In our testing this is model is over accurate despite it can't detect Nano-Banana images(only edited fully generated images can be detected over accurately))
98
+
99
+ ---
100
+
101
+ ## Intended Use
102
+
103
+ * Detect AI-generated imagery on social media, research publications, and digital media platforms.
104
+ * Assist content moderators, researchers, and fact-checkers in identifying synthetic media.
105
+ * **Not intended** for legal verification without human corroboration.
106
+
107
+ ---
108
+
109
+ ## Ethical Considerations
110
+
111
+ * **Privacy-first Approach:** Personal photos, even if AI-edited, were excluded.
112
+ * **Responsible Deployment:** Users should combine model predictions with human review to avoid false positives or negatives.
113
+ * **Transparency:** The model card openly communicates its limitations and dataset design to prevent misuse.
114
+
115
+ ---
116
+
117
+ ## How It Works
118
+
119
+ 1. Images are preprocessed and resized to `256x256`.
120
+ 2. Features are extracted using the **SwinV2 Tiny** vision transformer backbone.
121
+ 3. A binary classification head outputs probabilities for AI-generated vs real human images.
122
+ 4. Predictions are interpreted as class 0 (AI) or class 1 (Human).
123
+
124
+ ---
125
+
126
+ ## Future Work
127
+
128
+ Future iterations aim to:
129
+
130
+ * Improve detection of subtle AI-generated edits and “nano banana” modifications.
131
+ * Expand training data with diverse AI generators to enhance generalization.
132
+ * Explore multi-modal detection capabilities (e.g., video, metadata, and image combined).
133
+
134
+ ---
135
+
136
+ ## References
137
+
138
+ * Microsoft SwinV2 Tiny: [https://github.com/microsoft/Swin-Transformer](https://github.com/microsoft/Swin-Transformer)
139
+ * Parveshiiii/AI-vs-Real dataset (subset): Open-sourced by our team member
140
+
141
+ ---
142
+