anubhavmaity commited on
Commit
21d61b2
1 Parent(s): c5aaf8b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -1
README.md CHANGED
@@ -36,4 +36,52 @@ dataset_info:
36
  ---
37
  # Dataset Card for "notMNIST"
38
 
39
- [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  ---
37
  # Dataset Card for "notMNIST"
38
 
39
+ ## Overview
40
+
41
+ The notMNIST dataset is a collection of images of letters from A to J in various fonts. It is designed as a more challenging alternative to the traditional MNIST dataset, which consists of handwritten digits. The notMNIST dataset is commonly used in machine learning and computer vision tasks for character recognition.
42
+
43
+ ## Dataset Information
44
+
45
+ Number of Classes: 10 (A to J)
46
+ Number of Samples: 187,24
47
+ Image Size: 28 x 28 pixels
48
+ Color Channels: Grayscale
49
+
50
+ ## Dataset Structure
51
+
52
+ The dataset is split into a training set and a test set. Each class has its own subdirectory containing images of that class. The directory structure is as follows:
53
+
54
+ notMNIST/
55
+ |-- train/
56
+ | |-- A/
57
+ | |-- B/
58
+ | |-- ...
59
+ | |-- J/
60
+ |
61
+ |-- test/
62
+ | |-- A/
63
+ | |-- B/
64
+ | |-- ...
65
+ | |-- J/
66
+
67
+
68
+ ## Acknowledgements
69
+
70
+ http://yaroslavvb.blogspot.com/2011/09/notmnist-dataset.html
71
+ https://www.kaggle.com/datasets/lubaroli/notmnist
72
+
73
+
74
+ ## Inspiration
75
+
76
+ This is a pretty good dataset to train classifiers! According to Yaroslav:
77
+
78
+ Judging by the examples, one would expect this to be a harder task
79
+ than MNIST. This seems to be the case -- logistic regression on top of
80
+ stacked auto-encoder with fine-tuning gets about 89% accuracy whereas
81
+ same approach gives got 98% on MNIST. Dataset consists of small
82
+ hand-cleaned part, about 19k instances, and large uncleaned dataset,
83
+ 500k instances. Two parts have approximately 0.5% and 6.5% label error
84
+ rate. I got this by looking through glyphs and counting how often my
85
+ guess of the letter didn't match it's unicode value in the font file.
86
+
87
+