metadata
dataset_info:
features:
- name: section
dtype: string
- name: filename
dtype: string
- name: text
dtype: string
splits:
- name: train
num_bytes: 73734751.97826087
num_examples: 43
- name: validation
num_bytes: 1714761.6739130435
num_examples: 1
- name: test
num_bytes: 3429523.347826087
num_examples: 2
download_size: 42120770
dataset_size: 78879037.00000001
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
license: odc-by
task_categories:
- text-generation
- fill-mask
language:
- en
size_categories:
- n<1K
law books (nougat-small)
A decent chunk of: https://www.survivorlibrary.com/index.php/8-category/173-library-law
(ki) ➜ primerdata-for-LLMs python push_dataset_from_text.py /home/pszemraj/Dropbox/programming-projects/primerdata-for-LLMs/utils/output-hf-nougat-space/law -e .md -r BEE-spoke-data/survivorslib-law-books INFO:__main__:Looking for files with extensions: ['md'] Processing md files: 100%|███████████████████████████████| 46/46 [00:00<00:00, 778.32it/s] INFO:__main__:Found 46 text files. INFO:__main__:Performing train-test split... INFO:__main__:Performing validation-test split... INFO:__main__:Train size: 43 INFO:__main__:Validation size: 1 INFO:__main__:Test size: 2 INFO:__main__:Pushing dataset