--- annotations_creators: - no-annotation language: - de - fr - el - et - fi - hr - ji - pl - ru - sr - sv - uk language_creators: - machine-generated multilinguality: - multilingual pretty_name: 'Europeana Newspapers ' size_categories: - 1M - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses ### Direct Use To download the full dataset using the `Datasets` library you can do the following ```python from datasets import load_dataset dataset = load_dataset("biglam/europeana_newspapers") ``` You can also access a subset based on language or decade ranges using the following function. ```python from typing import List, Optional, Literal, Union from huggingface_hub import hf_hub_url, list_repo_files LanguageOption = Literal[ "et", "pl", "sr", "ru", "sv", "no_language_found", "ji", "hr", "el", "uk", "fr", "fi", "de", "multi_language", ] def get_files_for_lang_and_years( languages: Union[None, List[LanguageOption]] = None, min_year: Optional[int] = None, max_year: Optional[int] = None, ): files = list_repo_files("biglam/europeana_newspapers", repo_type="dataset") parquet_files = [f for f in files if f.endswith(".parquet")] parquet_files_filtered_for_lang = [ f for f in parquet_files if any(lang in f for lang in ["uk", "fr"]) ] filtered_files = [ f for f in parquet_files if (min_year is None or min_year <= int(f.split("-")[1].split(".")[0])) and (max_year is None or int(f.split("-")[1].split(".")[0]) <= max_year) ] return [ hf_hub_url("biglam/europeana_newspapers", f, repo_type="dataset") for f in filtered_files ] ``` This function takes a list of language codes, and a min, max value for decades you want to include. You can can use this function to get the URLs for files you want to download from the Hub: ```python ds = load_dataset("parquet", data_files=get_files_for_lang_and_years(['fr']), num_proc=4) ``` [More Information Needed] ### Out-of-Scope Use [More Information Needed] ## Dataset Structure [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Data Collection and Processing [More Information Needed] #### Who are the source data producers? [More Information Needed] ### Annotations [optional] #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] #### Personal and Sensitive Information [More Information Needed] ## Bias, Risks, and Limitations [More Information Needed] ### Recommendations Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations. ## Citation [optional] **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] [More Information Needed] ## More Information [optional] [More Information Needed] ## Dataset Card Authors [optional] [More Information Needed] ## Dataset Card Contact [More Information Needed]