Datasets:

DLDLD111 commited on
Commit
7a3de80
·
verified ·
1 Parent(s): 341ed36

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +142 -54
README.md CHANGED
@@ -1,70 +1,158 @@
1
  ---
2
  configs:
3
- - config_name: engineering_human_factor
4
- data_files:
5
- - split: train
6
- path: data/engineering_human_factor.json
7
- - config_name: material_science_additive_manufacturing
8
- data_files:
9
- - split: train
10
- path: data/material_science_additive_manufacturing.json
11
- - config_name: physics_surface_enhanced_raman_spectroscopy
12
- data_files:
13
- - split: train
14
- path: data/physics_surface_enhanced_raman_spectroscopy.json
15
- - config_name: public_health_infectious_disease_modeling
16
- data_files:
17
- - split: train
18
- path: data/public_health_infectious_disease_modeling.json
19
- - config_name: earth_science_remote_sensing
20
- data_files:
21
- - split: train
22
- path: data/earth_science_remote_sensing.json
23
- pretty_name: Intrabench
24
  task_categories:
25
- - question-answering
26
- - text-classification
27
  size_categories:
28
- - n<1K
29
- license: mit
30
- language:
31
- - en
32
  ---
33
 
34
- # IntraBench Spaceholding
35
 
36
- This dataset is a five-domain benchmark of multiple-choice research questions derived from scientific literature.
37
- Each example represents one question-paper pair with a gold letter answer and supporting metadata.
38
 
39
- ## Subsets
40
- - `engineering_human_factor`: Engineering - Human Factor
41
- - `material_science_additive_manufacturing`: Material Science - Additive Manufacturing
42
- - `physics_surface_enhanced_raman_spectroscopy`: Physics - Surface Enhanced Raman Spectroscopy
43
- - `public_health_infectious_disease_modeling`: Public Health - Infectious-disease Modeling
44
- - `earth_science_remote_sensing`: Earth Science - Remote Sensing
45
 
46
- ## Data Schema
47
 
48
- Each record contains:
49
 
50
- - `subject`: human-readable subject name
51
- - `paper_id`: paper index within the subject subset
52
- - `paper_title`: canonical benchmark paper title
53
- - `question_id`: question identifier
54
- - `question`: canonical benchmark question wording
55
- - `choices`: answer choice dictionary keyed by `A`-`F`
56
- - `answer`: gold answer letter
57
- - `metadata`: extra metadata with:
58
- - `Task-oriented Category`
59
- - `question_key_term`
60
- - `term_explanation`
61
 
62
- ## Notes
63
 
64
- - The published dataset excludes source PDFs, markdown paper extracts, and internal evaluation outputs.
65
- - The benchmark uses canonical paper titles and normalized question wording across all five domains.
66
- - Data files are stored as JSON arrays and exposed as separate Hugging Face configs.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
 
68
  ## Citation
69
 
70
- If you publish this dataset, add the benchmark paper citation here.
 
 
 
 
 
 
 
 
1
  ---
2
  configs:
3
+ - config_name: engineering_human_factor
4
+ data_files:
5
+ - split: train
6
+ path: data/engineering_human_factor.json
7
+ - config_name: material_science_additive_manufacturing
8
+ data_files:
9
+ - split: train
10
+ path: data/material_science_additive_manufacturing.json
11
+ - config_name: physics_surface_enhanced_raman_spectroscopy
12
+ data_files:
13
+ - split: train
14
+ path: data/physics_surface_enhanced_raman_spectroscopy.json
15
+ - config_name: public_health_infectious_disease_modeling
16
+ data_files:
17
+ - split: train
18
+ path: data/public_health_infectious_disease_modeling.json
19
+ - config_name: earth_science_remote_sensing
20
+ data_files:
21
+ - split: train
22
+ path: data/earth_science_remote_sensing.json
23
+ pretty_name: IntraBench
24
  task_categories:
25
+ - question-answering
26
+ - text-classification
27
  size_categories:
28
+ - n<1K
 
 
 
29
  ---
30
 
31
+ # IntrAgent: An LLM Agent for Content-Grounded Information Retrieval through Literature Review
32
 
33
+ Fengbo Ma, Zixin Rao, Xiaoting Li, Zhetao Chen, Hongyue Sun, Yiping Zhao, Xianyan Chen, Zhen Xiang
34
+ University of Georgia, Athens, GA, USA
35
 
36
+ [![Paper](https://img.shields.io/badge/Paper-Coming%20Soon-2f6f4f?style=for-the-badge)](#resources)
37
+ [![GitHub](https://img.shields.io/badge/GitHub-Code-181717?style=for-the-badge&logo=github&logoColor=white)](https://github.com/FengboMa/IntrAgent)
38
+ [![Hugging%20Face](https://img.shields.io/badge/Hugging%20Face-Dataset-ffcc4d?style=for-the-badge)](https://huggingface.co/datasets/IntrAgent/IntraBench)
 
 
 
39
 
40
+ ## Key Contributions
41
 
42
+ IntrAgent introduces **IntraView**, a new task for content-grounded information retrieval from a provided scientific paper, and proposes an LLM agent that mimics human literature reading through structure-aware section ranking and iterative evidence gathering. To evaluate this setting, the work presents **IntraBench**, a 315-instance benchmark across five STEM domains on which IntrAgent improves average cross-domain accuracy by **13.2%** over strong RAG and literature-agent baselines.
43
 
44
+ - Defines **IntraView**, a new task for faithful information retrieval from a provided scientific paper rather than from external search.
45
+ - Introduces **IntrAgent**, a specialized LLM agent that follows a human-like workflow: identify promising sections, extract evidence, and stop when support is sufficient.
46
+ - Develops two core mechanisms for the task: **hierarchy-preserving section ranking** and **iterative reading with information sufficiency checking**.
47
+ - Builds **IntraBench**, the first benchmark for this setting, covering 315 instances across five impactful STEM domains.
 
 
 
 
 
 
 
48
 
49
+ ## IntraView Task
50
 
51
+ IntraView is formulated as a content question answering problem over a full scientific paper. Given a literature document *C* and a research-driven query *Q*, the system must return an answer *A* that is accurate, concise, and explicitly grounded in the provided paper.
52
+
53
+ Compared with standard content QA, the task is harder because scientific papers are long, structurally complex, and filled with domain-specific terminology. The relevant evidence may appear anywhere in the document, may require cross-referencing multiple sections, and may sometimes be absent entirely, making hallucination control central to the task.
54
+
55
+ ## IntrAgent Method
56
+
57
+ ### Stage 1: Section Ranking
58
+
59
+ IntrAgent first parses section titles and preserves the paper hierarchy so the model can reason over the document as a structured artifact rather than a flat list of chunks. The LLM then ranks sections by likely relevance to the question, producing a reordered reading path.
60
+
61
+ This hierarchy-aware step is designed to better align scientific questions with the parts of a paper most likely to contain supporting evidence.
62
+
63
+ ### Stage 2: Iterative Reading
64
+
65
+ The agent reads the ranked sections sequentially, extracts anchored details such as terminology, measurements, results, and comparisons, and stores them in short-term memory for answer synthesis.
66
+
67
+ After each reading step, IntrAgent performs an explicit **information sufficiency check**. If the evidence is still incomplete, it continues reading; otherwise it stops and synthesizes a grounded answer.
68
+
69
+ ![IntrAgent pipeline](./element/pipeline.png)
70
+
71
+ Figure 1. Overview of the IntrAgent pipeline containing two stages: Section Ranking (top) reorders the paper’s sections by relevance to the Research Question Q, while Iterative Reading (bottom) steps through ranked sections, extracting information until gathered information is sufficient.
72
+
73
+ ## IntraBench Benchmark
74
+
75
+ To evaluate IntraView, the paper introduces **IntraBench**, the first benchmark specifically designed for literature-grounded information retrieval. It contains **315 test instances** derived from expert-authored questions paired with research papers.
76
+
77
+ The benchmark spans five high-impact domains and is intended to capture technical depth, conceptual complexity, and domain-specific phrasing encountered in real literature review workflows.
78
+
79
+ Domains: Physics, Earth Science, Public Health, Engineering, Material Science
80
+
81
+ Questions are organized around four research-oriented categories described in the paper: study subject and experimental setup, data characteristics and collection, technical approach and details, and conclusions and results. Evaluation is performed through LLM-grounded multiple-choice mapping to handle synonyms, abbreviations, and scientific terminology variation.
82
+
83
+ ![IntrAgent example](./element/example.png)
84
+
85
+ Figure 2. An example of IntrAgent executing a question-paper pair from IntraBench. For an input question Q regarding paper C, vanilla RAG fails to extract the correct chunk, resulting in an incorrect answer. In contrast, IntrAgent ranks the sections through reasoning and retrieves the correct details that pass the sufficiency check in the first iteration, leading to a correct answer. Details for section ranking and sufficiency check are also presented.
86
+
87
+ ### Benchmark Papers Used in IntraBench
88
+
89
+ | Domain | Title |
90
+ |---|---|
91
+ | Public Health - Infectious-disease Modeling | Mathematical modeling and analysis of COVID-19: A study of new variant Omicron |
92
+ | Public Health - Infectious-disease Modeling | COVID-19 pandemic in India: a mathematical model study |
93
+ | Public Health - Infectious-disease Modeling | A mathematical COVID-19 model considering asymptomatic and symptomatic classes with waning immunity |
94
+ | Public Health - Infectious-disease Modeling | Mathematical modeling and analysis of COVID-19 pandemic in Nigeria |
95
+ | Public Health - Infectious-disease Modeling | Mathematical modeling of COVID-19 transmission dynamics with a case study of Wuhan |
96
+ | Physics - Surface Enhanced Raman Spectroscopy | Quantification of Analyte Concentration in the Single Molecule Regime Using Convolutional Neural Networks |
97
+ | Physics - Surface Enhanced Raman Spectroscopy | Machine learning enabled multiplex detection of periodontal pathogens by surface-enhanced Raman spectroscopy |
98
+ | Physics - Surface Enhanced Raman Spectroscopy | Rapid Detection of SARS-CoV-2 Variants Using an Angiotensin-Converting Enzyme 2-Based Surface-Enhanced Raman Spectroscopy Sensor Enhanced by CoVari Deep Learning Algorithms |
99
+ | Physics - Surface Enhanced Raman Spectroscopy | Rapid Detection of SARS-CoV-2 RNA in Human Nasopharyngeal Specimens Using Surface-Enhanced Raman Spectroscopy and Deep Learning Algorithms |
100
+ | Physics - Surface Enhanced Raman Spectroscopy | Quantitative detection of α1-acid glycoprotein (AGP) level in blood plasma using SERS and CNN transfer learning approach |
101
+ | Earth Science - Remote Sensing | Annual maps of global artificial impervious area (GAIA) between 1985 and 2018 |
102
+ | Earth Science - Remote Sensing | Annual dynamics of global land cover and its long-term changes from 1982 to 2015 |
103
+ | Earth Science - Remote Sensing | Finer resolution observation and monitoring of global land cover: first mapping results with Landsat TM and ETM+ data |
104
+ | Earth Science - Remote Sensing | Mapping Essential Urban Land Use Categories in Beijing with a Fast Area of Interest (AOI)-Based Method |
105
+ | Earth Science - Remote Sensing | Mapping essential urban land use categories in China (EULUC-China): preliminary results for 2018 |
106
+ | Engineering - Human Factor | A data analytic end-to-end framework for the automated quantification of ergonomic risk factors across multiple tasks using a single wearable sensor |
107
+ | Engineering - Human Factor | Assessing human situation awareness reliability considering fatigue and mood using EEG data: A Bayesian neural network-Bayesian network approach |
108
+ | Engineering - Human Factor | Automatic driver cognitive fatigue detection based on upper body posture variations |
109
+ | Engineering - Human Factor | Enhancing Data Privacy in Human Factors Studies with Federated Learning |
110
+ | Engineering - Human Factor | Worker’s physical fatigue classification using neural networks |
111
+ | Material Science - Additive Manufacturing | Autonomous optimization of process parameters and in-situ anomaly detection in aerosol jet printing by an integrated machine learning approach |
112
+ | Material Science - Additive Manufacturing | Geometrical defect detection for additive manufacturing with machine learning models |
113
+ | Material Science - Additive Manufacturing | Layer-Wise Modeling and Anomaly Detection for LaserBased Additive Manufacturing |
114
+ | Material Science - Additive Manufacturing | Online droplet anomaly detection from streaming videos in inkjet printing |
115
+ | Material Science - Additive Manufacturing | Toward the digital twin of additive manufacturing- Integrating thermal simulations, sensing, and analytics to detect process faults |
116
+
117
+ ## Experiments and Results
118
+
119
+ The experiments compare IntrAgent against a broad set of RAG-based retrieval systems and literature-oriented agents, including vanilla RAG variants, contextual RAG, DRAGIN, R2AG, LongRAG, LUMOS, PaperQA2, Agentic-Hybrid-RAG, and SciMaster.
120
+
121
+ On IntraBench, IntrAgent sets a new state of the art across all five domains and seven backbone LLMs. Reported average accuracies include **70.0%** with GPT-4o, **75.8%** with GPT-4.1, **74.4%** with DeepSeek-R1, **73.4%** with o3, **73.8%** with o4-mini, **75.9%** with Gemini-2.5 Pro, and **68.8%** with Llama-3.1-70B.
122
+
123
+ The paper attributes these gains to two main design choices: hierarchy-aware section ranking and the sufficiency check that stops reading once evidence is complete. In contrast, flat RAG pipelines often inject irrelevant chunks, while literature agents designed for online search degrade into static retrieval pipelines when constrained to a provided paper.
124
+
125
+ ### Cross-Domain Accuracy on IntraBench
126
+
127
+ | Group | Method | GPT-4o | GPT-4.1 | DS-R1 | o3 | o4-mini | Gemini-2.5 Pro | Llama-3.1-70B |
128
+ |---|---|---:|---:|---:|---:|---:|---:|---:|
129
+ | RAG | Vanilla RAG all-MiniLM-L6-v2 | 60.3 | 61.2 | 64.3 | 60.4 | 61.5 | 61.8 | 59.2 |
130
+ | RAG | Vanilla RAG E5-mistral-7b-instruct | 59.4 | 64.2 | 63.8 | 60.3 | 61.4 | 59.9 | 60.5 |
131
+ | RAG | Vanilla RAG GritLM-7B | 60.4 | 63.2 | 63.2 | 59.7 | 58.4 | 58.4 | 61.4 |
132
+ | RAG | Context. RAG E5-mistral-7b-instruct | 60.7 | 63.8 | 62.8 | 59.1 | 58.3 | 58.9 | 58.9 |
133
+ | RAG | Context. RAG GritLM-7B | 60.8 | 62.8 | 61.6 | 58.4 | 60.7 | 61.6 | 59.2 |
134
+ | RAG | DRAGIN | 42.5 | 44.6 | 46.9 | 44.0 | 46.9 | 45.9 | 45.4 |
135
+ | RAG | R²AG | 59.4 | 59.5 | 61.5 | 56.6 | 55.3 | 55.6 | 56.1 |
136
+ | RAG | LongRAG | 62.1 | 64.7 | 65.5 | 57.0 | 58.3 | 57.1 | 57.4 |
137
+ | Agent | LUMOS | 50.2 | 52.1 | 55.4 | 55.2 | 56.4 | 54.9 | 54.4 |
138
+ | Agent | PaperQA2 | 47.7 | 48.9 | 54.0 | 51.8 | 49.2 | 51.2 | 53.8 |
139
+ | Agent | Agentic-Hybrid-RAG | 59.8 | 60.2 | 62.3 | 57.5 | 57.8 | 57.2 | 56.6 |
140
+ | Agent | SciMaster | 59.0 | 57.6 | 63.3 | 57.2 | 58.1 | 57.2 | 57.0 |
141
+ | Agent | **IntrAgent (Ours)** | **70.0** | **75.8** | **74.4** | **73.4** | **73.8** | **75.9** | **68.8** |
142
+
143
+ ## Resources
144
+
145
+ - Paper: coming soon
146
+ - GitHub: https://github.com/FengboMa/IntrAgent
147
+ - Hugging Face: https://huggingface.co/datasets/IntrAgent/IntraBench
148
 
149
  ## Citation
150
 
151
+ ```bibtex
152
+ @inproceedings{ma2026intragent,
153
+ title={IntrAgent: An {LLM} Agent for Content-Grounded Information Retrieval through Literature Review},
154
+ author={Fengbo Ma and Zixin Rao and Xiaoting Li and Zhetao Chen and Hongyue Sun and Xianyan Chen and Yiping Zhao and Zhen Xiang},
155
+ booktitle={The 64th Annual Meeting of the Association for Computational Linguistics},
156
+ year={2026},
157
+ }
158
+ ```