Suggest some additional metadata
#2
by
davanstrien
HF staff
- opened
README.md
CHANGED
@@ -1,5 +1,12 @@
|
|
1 |
---
|
2 |
license: cc0-1.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
**French-Public Domain-Book** or **French-PD-Books** is a large collection aiming to agregate all the French monographies in the public domain.
|
5 |
|
@@ -21,4 +28,4 @@ The collection also aims to expand the availability of open works for the traini
|
|
21 |
This dataset is not a one time work but will continue to evolve significantly on two directions:
|
22 |
* Correction of computer generated errors in the text. All the texts have been transcribed automatically through the use of Optical Character Recognition (OCR) software. The original files have been digitized over a long time period (since the mid-2000s) and some documents should be. Future versions will strive either to re-OCRize the original text or use experimental LLM models for partial OCR correction.
|
23 |
* Enhancement of the structure/editorial presentation of the original text. Some parts of the original documents are likely unwanted for large scale analysis or model training (header, page count…). Additionally, some advanced document structures like tables or multi-column layout are unlikely to be well formatted. Major enhancements could be experted through applying new SOTA layout recognition models (like COLAF) on the original PDF files.
|
24 |
-
* Expansion of the collection to other cultural heritage holdings, especially coming from Hathi Trust, Internet Archive and Google Books.
|
|
|
1 |
---
|
2 |
license: cc0-1.0
|
3 |
+
task_categories:
|
4 |
+
- text-generation
|
5 |
+
language:
|
6 |
+
- fr
|
7 |
+
tags:
|
8 |
+
- ocr
|
9 |
+
pretty_name: French-Public Domain-Book
|
10 |
---
|
11 |
**French-Public Domain-Book** or **French-PD-Books** is a large collection aiming to agregate all the French monographies in the public domain.
|
12 |
|
|
|
28 |
This dataset is not a one time work but will continue to evolve significantly on two directions:
|
29 |
* Correction of computer generated errors in the text. All the texts have been transcribed automatically through the use of Optical Character Recognition (OCR) software. The original files have been digitized over a long time period (since the mid-2000s) and some documents should be. Future versions will strive either to re-OCRize the original text or use experimental LLM models for partial OCR correction.
|
30 |
* Enhancement of the structure/editorial presentation of the original text. Some parts of the original documents are likely unwanted for large scale analysis or model training (header, page count…). Additionally, some advanced document structures like tables or multi-column layout are unlikely to be well formatted. Major enhancements could be experted through applying new SOTA layout recognition models (like COLAF) on the original PDF files.
|
31 |
+
* Expansion of the collection to other cultural heritage holdings, especially coming from Hathi Trust, Internet Archive and Google Books.
|