| | import random |
| | from glob import glob |
| |
|
| | from datasets import load_dataset, Dataset, DatasetDict |
| |
|
| | |
| | token_files = glob('tokenized/*.tokens') |
| | total_files = len(token_files) |
| |
|
| | print(f"Found {total_files} token files") |
| |
|
| | |
| | train_size = 23 |
| | dev_size = 8 |
| | test_size = 8 |
| |
|
| | |
| | if total_files < (train_size + dev_size + test_size): |
| | print(f"Warning: Not enough files ({total_files}) for the requested split sizes.") |
| | |
| | total_requested = train_size + dev_size + test_size |
| | train_size = int(total_files * (train_size / total_requested)) |
| | dev_size = int(total_files * (dev_size / total_requested)) |
| | test_size = total_files - train_size - dev_size |
| |
|
| | |
| | random.seed(42) |
| | random.shuffle(token_files) |
| |
|
| | |
| | train_files = token_files[:train_size] |
| | dev_files = token_files[train_size:train_size + dev_size] |
| | test_files = token_files[train_size + dev_size:train_size + dev_size + test_size] |
| |
|
| | |
| | def process_files(file_list): |
| | result = [] |
| | for file in file_list: |
| | tokens = [] |
| | ner_tags = [] |
| | |
| | with open(file, 'r') as f: |
| | for line in f: |
| | line = line.strip() |
| | |
| | |
| | if not line: |
| | if tokens: |
| | result.append({ |
| | "tokens": tokens, |
| | "ner_tags": ner_tags, |
| | "file_name": file |
| | }) |
| | tokens = [] |
| | ner_tags = [] |
| | continue |
| | |
| | |
| | parts = line.split() |
| | |
| | |
| | if len(parts) >= 3: |
| | token = parts[0] |
| | ner_tag = parts[2] |
| | |
| | tokens.append(token) |
| | ner_tags.append(ner_tag) |
| | |
| | |
| | if tokens: |
| | result.append({ |
| | "tokens": tokens, |
| | "ner_tags": ner_tags, |
| | "file_name": file |
| | }) |
| | |
| | return result |
| |
|
| | |
| | train_data = process_files(train_files) |
| | dev_data = process_files(dev_files) |
| | test_data = process_files(test_files) |
| |
|
| | |
| | train_dataset = Dataset.from_list(train_data) |
| | dev_dataset = Dataset.from_list(dev_data) |
| | test_dataset = Dataset.from_list(test_data) |
| |
|
| | |
| | dataset_dict = DatasetDict({ |
| | "train": train_dataset, |
| | "validation": dev_dataset, |
| | "test": test_dataset |
| | }) |
| |
|
| | |
| | print(f"Train split: {len(train_data)} files") |
| | print(f"Validation split: {len(dev_data)} files") |
| | print(f"Test split: {len(test_data)} files") |
| | print(f"Dataset features: {train_dataset.features}") |
| |
|
| | |
| | dataset_dict.push_to_hub('extraordinarylab/malware-text-db') |