Update README.md
Browse files
README.md
CHANGED
|
@@ -6,7 +6,7 @@
|
|
| 6 |
|
| 7 |
# Overview
|
| 8 |
|
| 9 |
-
Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising
|
| 10 |
|
| 11 |

|
| 12 |
|
|
@@ -14,8 +14,6 @@ Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising
|
|
| 14 |
|
| 15 |
To construct a comprehensive dataset for document parsing, we integrate both real-world and synthetic data generation pipelines. Our real-world data pipeline collects diverse scanned documents from various practical domains (such as financial reports, medical records, and academic papers), employing a multi-expert strategy with cross-validation to generate reliable pseudo-ground-truth annotations for structural elements like text, tables, and formulas. Complementing this, our synthetic data pipeline programmatically creates a wide array of documents by injecting content from sources like Wikipedia into predefined HTML layouts, rendering them into scanned formats, and extracting precise ground-truth annotations directly from the original HTML. This dual approach yields a rich, diverse, and cost-effective dataset with accurate and well-aligned supervision, effectively overcoming common issues of imprecise or inconsistent labeling found in other datasets and enabling robust training for end-to-end document parsing models.
|
| 16 |
|
| 17 |
-
Note that the elements in 'form' are sorted in reading order.
|
| 18 |
-
|
| 19 |

|
| 20 |
|
| 21 |
# Data Statistics
|
|
@@ -32,7 +30,6 @@ Note that the elements in 'form' are sorted in reading order.
|
|
| 32 |
| All | 55k |||
|
| 33 |
|
| 34 |
# Data Structure
|
| 35 |
-
|
| 36 |
- id: The MD5 hash of the image, which serves as its unique identifier.
|
| 37 |
- image: The document image.
|
| 38 |
- gt: The content of the document, formatted in Markdown/HTML.
|
|
|
|
| 6 |
|
| 7 |
# Overview
|
| 8 |
|
| 9 |
+
Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising 55K real-world and synthetic scanned documents. The dataset features rich layout variations and comprehensive structural annotations, enabling robust training of document parsing models. Additionally, this dataset encompasses a broad spectrum of document types, including financial reports, medical reports, academic reports, books, magazines, web pages, and synthetic documents.
|
| 10 |
|
| 11 |

|
| 12 |
|
|
|
|
| 14 |
|
| 15 |
To construct a comprehensive dataset for document parsing, we integrate both real-world and synthetic data generation pipelines. Our real-world data pipeline collects diverse scanned documents from various practical domains (such as financial reports, medical records, and academic papers), employing a multi-expert strategy with cross-validation to generate reliable pseudo-ground-truth annotations for structural elements like text, tables, and formulas. Complementing this, our synthetic data pipeline programmatically creates a wide array of documents by injecting content from sources like Wikipedia into predefined HTML layouts, rendering them into scanned formats, and extracting precise ground-truth annotations directly from the original HTML. This dual approach yields a rich, diverse, and cost-effective dataset with accurate and well-aligned supervision, effectively overcoming common issues of imprecise or inconsistent labeling found in other datasets and enabling robust training for end-to-end document parsing models.
|
| 16 |
|
|
|
|
|
|
|
| 17 |

|
| 18 |
|
| 19 |
# Data Statistics
|
|
|
|
| 30 |
| All | 55k |||
|
| 31 |
|
| 32 |
# Data Structure
|
|
|
|
| 33 |
- id: The MD5 hash of the image, which serves as its unique identifier.
|
| 34 |
- image: The document image.
|
| 35 |
- gt: The content of the document, formatted in Markdown/HTML.
|