|
--- |
|
license: mit |
|
task_categories: |
|
- image-classification |
|
- zero-shot-image-classification |
|
- text-to-image |
|
language: |
|
- en |
|
tags: |
|
- art |
|
- anime |
|
- not-for-all-audiences |
|
size_categories: |
|
- 1M<n<10M |
|
--- |
|
|
|
# Danbooru 2023 webp: A space-efficient version of Danbooru 2023 |
|
|
|
This dataset is a resized/re-encoded version of [danbooru2023](https://huggingface.co/datasets/nyanko7/danbooru2023).<br> |
|
Which removed the non-image/truncated files and resize all of them into smaller size. |
|
|
|
This dataset already be updated to latest_id = 7,832,883. |
|
Thx to DeepGHS! |
|
|
|
**Notice**: content of updates folder and deepghs/danbooru_newest-webp-4Mpixel have been merged to 2000~2999.tar, You can ignore all the content in updates folder safely! |
|
|
|
--- |
|
|
|
## Details |
|
This dataset employs few method to reduce the size and improve the efficiency. |
|
|
|
### Size and Format |
|
This dataset resize all the image which have more than 2048x2048 pixel into near 2048x2048 pixels with bicubic algorithm.<br> |
|
And remove all the image with longer edge larger than 16383 after resize.<br> |
|
(one reason is beacuse webp doesn't allow that, another is that aspect ratio is too large/small.) |
|
|
|
This dataset encode/save all the image with 90% quality webp with pillow library in Python. |
|
Which is half size of the 100% quality lossy webp. |
|
|
|
The total size of this dataset is around 1.3~1.4TB. Which is less than the 20% of original file size. |
|
|
|
### Webdataset |
|
This dataset use webdataset library to save all the tarfile, therefore, you can also use webdataset to load them easily. This is also a recommended way. |
|
|
|
The `__key__` of each files is the id of it. You can use this id to query the [metadata database](https://huggingface.co/datasets/KBlueLeaf/danbooru2023-sqlite) easily. |
|
|