danbooru2023 / README.md
nyanko7's picture
Better explanation about continuous updates (#7)
046a4f1 verified
|
raw
history blame
3.69 kB
---
license: mit
task_categories:
- image-classification
- image-to-image
- text-to-image
language:
- en
- ja
pretty_name: danbooru2023
size_categories:
- 1M<n<10M
viewer: false
---
<img src="https://huggingface.co/datasets/nyanko7/danbooru2023/resolve/main/cover.webp" alt="cover" width="750"/>
# Danbooru2023: A Large-Scale Crowdsourced and Tagged Anime Illustration Dataset
<!-- Provide a quick summary of the dataset. -->
Danbooru2023 is a large-scale anime image dataset with over 5 million images contributed and annotated in detail by an enthusiast community. Image tags cover aspects like characters, scenes, copyrights, artists, etc with an average of 30 tags per image.
Danbooru is a veteran anime image board with high-quality images and extensive tag metadata. The dataset can be used to train image classification, multi-label tagging, character detection, generative models, and other computer vision tasks.
- **Shared by:** Nyanko Devs
- **Language(s):** English, Japanese
- **License:** MIT
This dataset is built on the top of [danbooru2021](https://gwern.net/danbooru2021). We expands the dataset to include images up to ID #6,857,737, adding over 1.8 million additional images and total size is now approximately 8 terabytes (8,000 GB).
## Use
## Format
The goal of the dataset is to be as easy as possible to use immediately, avoiding obscure file formats, while allowing simultaneous research & seeding of the torrent, with easy updates.
Images are provided in the full original form (be that JPG, PNG, GIF or otherwise) for reference/archival purposes, and bucketed into 1000 subdirectories 0000–0999 (0-padded), which is the Danbooru ID modulo 1000 (ie. all images in 0999/ have an ID ending in β€˜999’); IDs can be turned into paths by dividing & padding (eg. in Bash, BUCKET=$(printf "%04d" $(( ID % 1000 )) )) and then the file is at {original,512px}/$BUCKET/$ID.$EXT.
The reason for the bucketing is that a single directory would cause pathological filesystem performance, and modulo ID is a simple hash which spreads images evenly without requiring additional future directories to be made or a filesystem IO to check where the file is. The ID is not zero-padded and files end in the relevant extension, hence the file layout looks like this:
```bash
$ tree / | less
/
β”œβ”€β”€ danbooru2023 -> /mnt/diffusionstorage/workspace/danbooru/
β”‚ β”œβ”€β”€ metadata
β”‚ β”œβ”€β”€ readme.md
β”‚ β”œβ”€β”€ original
β”‚ β”‚ β”œβ”€β”€ 0000 -> data-0000.tar
β”‚ β”‚ β”œβ”€β”€ 0001 -> data-0001.tar
β”‚ β”‚ β”‚ β”œβ”€β”€ 10001.jpg
β”‚ β”‚ β”‚ β”œβ”€β”€ 210001.png
β”‚ β”‚ β”‚ β”œβ”€β”€ 3120001.webp
β”‚ β”‚ β”‚ β”œβ”€β”€ 6513001.jpg
β”‚ β”‚
β”‚ β”œβ”€β”€ recent
β”‚ β”‚ β”œβ”€β”€ 0000 -> data-1000.tar
β”‚ β”‚ β”œβ”€β”€ 0001 -> data-1001.tar
β”‚ β”‚
β”‚ β”œβ”€β”€ updates
β”‚ β”‚ β”œβ”€β”€ 20240319
β”‚ β”‚ β”‚ β”œβ”€β”€ dataset-0.tar
β”‚ β”‚ β”‚ β”œβ”€β”€ dataset-1.tar
β”‚ β”‚ β”‚
β”‚ β”‚ β”œβ”€β”€ 2024xxxx
β”‚ β”‚ β”‚ β”œβ”€β”€ dataset-0.tar
β”‚ β”‚ β”‚ β”œβ”€β”€ dataset-1.tar
```
Where `data-{1000..1999}.tar` refer to recent update files (should be updated every few months) and `updates` refer to fast patches (should be updated every few days to few weeks).
Currently represented file extensions are: avi/bmp/gif/html/jpeg/jpg/mp3/mp4/mpg/pdf/png/rar/swf/webm/wmv/zip.
Raw original files are treacherous. Be careful if working with the original dataset. There are many odd files: truncated, non-sRGB colorspace, wrong file extensions (eg. some PNGs have .jpg extensions like original/0146/1525146.jpg or original/0558/1422558.jpg), etc.