Update README.md
Browse files
README.md
CHANGED
@@ -1,8 +1,9 @@
|
|
1 |
---
|
2 |
license: creativeml-openrail-m
|
3 |
---
|
|
|
4 |
|
5 |
-
|
6 |
```
|
7 |
ETL9G
|
8 |
: 607200 samples
|
@@ -13,49 +14,41 @@ ETL9G
|
|
13 |
: image_height: 64px
|
14 |
```
|
15 |
|
16 |
-
I was testing a few more samples locally with the below. Note: The results of the model are encoded.
|
17 |
|
18 |
-
|
19 |
-
Model Details
|
20 |
-
Model Name
|
21 |
-
Version
|
22 |
-
Model Type
|
23 |
-
Framework
|
24 |
-
|
|
|
25 |
This model is trained on a dataset derived from the ETL9G dataset to recognize Kanji characters from 64x64 grayscale images. The primary use-case is for optical character recognition (OCR) for handwritten Kanji characters.
|
26 |
|
27 |
-
Intended Use
|
28 |
The primary application of this model is for OCR tasks to recognize handwritten Kanji characters in images, with potential extensions for applications like smart dictionary lookup, handwriting-based user authentication, and so on.
|
29 |
|
30 |
-
Limitations
|
31 |
This model might have limitations regarding:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
-
|
34 |
-
|
35 |
-
Characters written in unconventional ways.
|
36 |
-
Data Details
|
37 |
-
Training Data:
|
38 |
-
Dataset: Derived from the ETL9G dataset
|
39 |
-
Size: (Specify the number of samples in the training dataset)
|
40 |
-
Data Type: 64x64 grayscale images of handwritten Kanji characters
|
41 |
-
Labels: Kanji characters (Specify the total number of unique characters)
|
42 |
-
Validation and Test Data:
|
43 |
-
Validation Size: (Specify the number of samples in the validation dataset)
|
44 |
-
Test Size: (Specify the number of samples in the test dataset)
|
45 |
-
Data Type: Same as training
|
46 |
-
Labels: Same as training
|
47 |
-
Metrics
|
48 |
-
Accuracy: (Your test accuracy here, e.g., 95.4%)
|
49 |
-
Precision: (If calculated)
|
50 |
-
Recall: (If calculated)
|
51 |
-
F1 Score: (If calculated)
|
52 |
-
Model Files
|
53 |
-
PyTorch Model: Kanji_ETL9G.pth
|
54 |
-
ONNX Model: (If you've saved it in ONNX format)
|
55 |
-
CoreML Model: (If you've converted it to CoreML)
|
56 |
-
Usage
|
57 |
-
python
|
58 |
-
Copy code
|
59 |
import torch
|
60 |
model = torch.load('Kanji_ETL9G.pth')
|
61 |
model.eval()
|
@@ -63,7 +56,3 @@ model.eval()
|
|
63 |
# Assuming input image tensor is `input_tensor`
|
64 |
output = model(input_tensor)
|
65 |
predicted_label = torch.argmax(output).item()
|
66 |
-
Maintainers
|
67 |
-
Your Name (Your email/contact details)
|
68 |
-
Licensing
|
69 |
-
Specify the license under which you're releasing the model. If unsure, consider common licenses like MIT, Apache 2.0, etc.
|
|
|
1 |
---
|
2 |
license: creativeml-openrail-m
|
3 |
---
|
4 |
+
# Model Card for Kanji_ETL9G
|
5 |
|
6 |
+
### Summary:
|
7 |
```
|
8 |
ETL9G
|
9 |
: 607200 samples
|
|
|
14 |
: image_height: 64px
|
15 |
```
|
16 |
|
|
|
17 |
|
18 |
+
|
19 |
+
## Model Details
|
20 |
+
- **Model Name:** Kanji_ETL9G
|
21 |
+
- **Version:** 1.0.0
|
22 |
+
- **Model Type:** Neural Network
|
23 |
+
- **Framework:** PyTorch
|
24 |
+
|
25 |
+
## Model Description
|
26 |
This model is trained on a dataset derived from the ETL9G dataset to recognize Kanji characters from 64x64 grayscale images. The primary use-case is for optical character recognition (OCR) for handwritten Kanji characters.
|
27 |
|
28 |
+
## Intended Use
|
29 |
The primary application of this model is for OCR tasks to recognize handwritten Kanji characters in images, with potential extensions for applications like smart dictionary lookup, handwriting-based user authentication, and so on.
|
30 |
|
31 |
+
## Limitations
|
32 |
This model might have limitations regarding:
|
33 |
+
- Variability in handwriting styles not present in the training set. (200 samples per character/class were used)
|
34 |
+
- Noises and artifacts in input images.
|
35 |
+
- Characters written in unconventional ways.
|
36 |
+
|
37 |
+
## Data Details
|
38 |
+
|
39 |
+
### Training Data:
|
40 |
+
- **Dataset:** Derived from the ETL9G dataset (http://etlcdb.db.aist.go.jp/specification-of-etl-9)
|
41 |
+
- **Size:** 607200 samples
|
42 |
+
- **Data Type:** 64x64 grayscale images of handwritten Kanji characters (images were resized from 128x127 due to technical limitations)
|
43 |
+
- **Labels:** 3036 unique characters (classes)
|
44 |
+
|
45 |
+
## Model Files
|
46 |
+
- **PyTorch Model:** Kanji_ETL9G.pth
|
47 |
+
- **ONNX Model:** Kanji_ETL9G.onnx
|
48 |
+
- **CoreML Model:** next effort....
|
49 |
|
50 |
+
## Usage
|
51 |
+
```python
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
import torch
|
53 |
model = torch.load('Kanji_ETL9G.pth')
|
54 |
model.eval()
|
|
|
56 |
# Assuming input image tensor is `input_tensor`
|
57 |
output = model(input_tensor)
|
58 |
predicted_label = torch.argmax(output).item()
|
|
|
|
|
|
|
|