You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

By completing the form below, you acknowledge that the provided data is offered as is. Although we anticipate no problems, you accept full responsibility for any repercussions resulting from the use of this data. Furthermore, you agree that the data must not be utilized for malicious or harmful purposes towards humanity.

Log in or Sign Up to review the conditions and access this dataset content.

BUT-LCC Corpus

BUT-LCC (Brno University of Technology Large Czech Collection) is a corpus of Czech texts. It was cleaned using exact deduplication, fuzzy deduplication (using minhashlsh), n-gram language model, and an SVM classifier that filters inappropriate content (we manually labelled).

Data Sources

Part GB of text GB of titles %
CulturaX 157.79 3.85 49
TenTen-cs-2017 48.97 0.95 15
BUT_Crawl 25.15 0.8 8
cswiki-20230101 1.05 0.01 0
historical 13.47 0.00 4
hplt 65.55 3.20 21
idnes_comments 7.38 0.03 2
Sum 319.36 8.84

Format

The corpus consists of train and test splits. It uses jsonl format, which means that every sample is JSON on its own line.

Sample Format

{
  "id": unique identifier, 		
  "part": original source, 
  "title": source document title, 
  "text": the context,
  "ugly": (type: bool) inappropriate content,
  "ugly_score": (type: float) score from SVM classifier that filters inappropriate content
}

License Information

  • We do not own any of the text from which these text data has been extracted.
  • We license the actual packaging of these text data under the Creative Commons CC0 license ("no rights reserved").

Detailed licensing information for contained corpora (not crawled by us) is below.

Corpus Licensing Information
CulturaX uonlp/CulturaX
TenTen-cs-2017 NLP Centre Web Corpus License Agreement
Czech Wikipedia CC BY-SA 4.0 DEED
Historical OCR'd documents since 1850, publicly available from the Czech Digital Library
HPLT https://hplt-project.org/datasets/v1.2

Our Models Linked to This Dataset

Statistics

Split Samples
Train 176 780 582
Test 20 000

ID 2 URL mapping

If you need to recover original webpages, we provide ID to source URL mapping where possible in id2url.csv file.

Acknowledgement

This work was supported by NAKI III program of Ministry of Culture Czech Republic, project semANT --- "Sémantický průzkumník textového kulturního dědictví" grant no. DH23P03OVV060 and by the Ministry of Education, Youth and Sports of the Czech Republic through the e-INFRA CZ (ID:90254).

Contributors

  • Jan Doležal developed cleaning pipeline for text processing, collected data for cleaning, and analyzed cutoff threshold for pruning.
  • Martin Dočkal uploaded data to Huggingface, and helped with cutoff analysis.
  • Martin Fajčík reviewed existing corpora, planned pipeline steps, processed data for LM training, and verified their usefullness.
  • Martin Kišš downloaded historical documents, and ran our PeroOCR on the collection.
  • Karel Beneš performed cleaning of historical documents, and created n-gram lm for document filtering.
  • Karel Ondřej who wrote a crawler for collecting BUT_Crawl and prepared preliminary clean corpus version.
  • Michal Hradiš managed the work, and pushed the members when necessary.
Downloads last month
10
Edit dataset card

Models trained or fine-tuned on BUT-FIT/BUT-LCC