Datasets:
metadata
license: apache-2.0
task_categories:
- text-retrieval
language:
- en
tags:
- information-retrieval
- benchmark
- clinical-trials
- code-search
- legal-qa
size_categories:
- 10K<n<100K
configs:
- config_name: clinical_trial
data_files:
- split: queries
path: clinical_trial/queries.jsonl
- split: documents
path: clinical_trial/documents.jsonl
- split: qrels
path: clinical_trial/qrels.jsonl
- config_name: code_retrieval
data_files:
- split: queries
path: code_retrieval/queries.jsonl
- split: documents
path: code_retrieval/documents.jsonl
- split: qrels
path: code_retrieval/qrels.jsonl
- config_name: legal_qa
data_files:
- split: queries
path: legal_qa/queries.jsonl
- split: documents
path: legal_qa/documents.jsonl
- split: qrels
path: legal_qa/qrels.jsonl
- config_name: paper_retrieval
data_files:
- split: queries
path: paper_retrieval/queries.jsonl
- split: documents
path: paper_retrieval/documents.jsonl
- split: qrels
path: paper_retrieval/qrels.jsonl
- config_name: set_operation_entity_retrieval
data_files:
- split: queries
path: set_operation_entity_retrieval/queries.jsonl
- split: documents
path: set_operation_entity_retrieval/documents.jsonl
- split: qrels
path: set_operation_entity_retrieval/qrels.jsonl
- config_name: stack_exchange
data_files:
- split: queries
path: stack_exchange/queries.jsonl
- split: documents
path: stack_exchange/documents.jsonl
- split: qrels
path: stack_exchange/qrels.jsonl
- config_name: theorem_retrieval
data_files:
- split: queries
path: theorem_retrieval/queries.jsonl
- split: documents
path: theorem_retrieval/documents.jsonl
- split: qrels
path: theorem_retrieval/qrels.jsonl
- config_name: tip_of_the_tongue
data_files:
- split: queries
path: tip_of_the_tongue/queries.jsonl
- split: documents
path: tip_of_the_tongue/documents.jsonl
- split: qrels
path: tip_of_the_tongue/qrels.jsonl
NanoCrumb Dataset
A curated subset of the Crumb retrieval dataset, designed for rapid experimentation and evaluation of information retrieval systems.
Dataset Summary
NanoCrumb distills the large Crumb dataset (10.5 GB, 6.36M rows) into a manageable benchmark while maintaining task diversity across 8 different retrieval domains.
- Total Size: ~125 MB (JSONL format)
- Queries: 400 (50 per task split)
- Documents: 30,040 unique passages
- Query-Document Pairs: 31,754
- Configs: 8 task-specific configs
Configs (Task Splits)
Each config represents a different retrieval domain:
| Config Name | Queries | Documents | Docs/Query (avg) | Description |
|---|---|---|---|---|
clinical_trial |
50 | 22,251 | 464 | Match patients to clinical trials |
paper_retrieval |
50 | 4,402 | 102 | Find relevant academic papers |
set_operation_entity_retrieval |
50 | 1,533 | 31 | Entity-based retrieval |
code_retrieval |
50 | 1,206 | 24 | Find relevant code snippets |
tip_of_the_tongue |
50 | 363 | 7 | Recall items from vague descriptions |
stack_exchange |
50 | 125 | 3 | Find relevant Q&A posts |
legal_qa |
50 | 86 | 2 | Legal question answering |
theorem_retrieval |
50 | 74 | 2 | Find mathematical theorems |
Dataset Structure
Each config contains three splits:
queries
query_id: Unique query identifier (string)query_content: The query text (string)instruction: Task-specific instructions (string)passage_qrels: List of relevant passages with graded relevance scores (list)task_split: Task domain name (string)metadata: Additional task-specific information (string)use_max_p: Boolean flag for MaxP aggregation (bool)
documents
document_id: Unique document identifier (string)document_content: The passage text (string)parent_id: Links passages to source documents (string)task_split: Task domain name (string)metadata: Document metadata (string)
qrels
query_id: Query identifier (string)document_id: Document identifier (string)relevance_score: Graded relevance 0.0-2.0 (float)binary_relevance: Binary relevance 0 or 1 (int)task_split: Task domain name (string)
Usage
from datasets import load_dataset
# Load a specific config (task split)
clinical_data = load_dataset("YOUR_USERNAME/nanocrumb", "clinical_trial")
# Access the splits
queries = clinical_data['queries']
documents = clinical_data['documents']
qrels = clinical_data['qrels']
# Load all configs
all_configs = [
"clinical_trial", "code_retrieval", "legal_qa", "paper_retrieval",
"set_operation_entity_retrieval", "stack_exchange",
"theorem_retrieval", "tip_of_the_tongue"
]
for config_name in all_configs:
data = load_dataset("YOUR_USERNAME/nanocrumb", config_name)
print(f"{config_name}: {len(data['queries'])} queries")
Sampling Methodology
For each task split:
- Query Selection: Randomly sampled 50 queries from evaluation set (seed=42)
- Document Selection:
- Include ALL positive documents (binary_relevance=1)
- Fill remainder with hard negatives (relevance=0) to reach ~100 docs per query
- Target: ~5,000 documents per task split
- Deduplication: Documents shared across queries are deduplicated within each config
Use Cases
- 🚀 Rapid prototyping of retrieval models
- 🧪 Quick benchmarking without downloading large datasets
- 📚 Educational purposes for learning IR techniques
- 🔬 Ablation studies across diverse domains
Citation
If you use NanoCrumb, please cite the original Crumb dataset:
@misc{crumb2024,
title={Crumb: A Comprehensive Retrieval Benchmark},
author={[Original Crumb Authors]},
year={2024},
url={https://huggingface.co/datasets/jfkback/crumb}
}
License
This dataset inherits the license from the original Crumb dataset.