|
defaults: |
|
- _self_ |
|
|
|
- retriever: related_span_retriever_with_relations_from_other_docs |
|
|
|
- argumentation_model: pipelined_new |
|
|
|
- pdf_fulltext_extractor: none |
|
|
|
|
|
|
|
|
|
handle_parts_of_same: true |
|
|
|
default_split_regex: "\n\n\n+" |
|
|
|
|
|
default_min_similarity: 0.95 |
|
default_top_k: 10 |
|
|
|
|
|
default_arxiv_id: "1706.03762" |
|
default_load_pie_dataset_kwargs: |
|
path: "pie/sciarg" |
|
name: "resolve_parts_of_same" |
|
split: "train" |
|
|
|
|
|
|
|
|
|
|
|
acl_anthology_pdf_dir: "data/acl-anthology/pdf" |
|
|
|
|
|
render_mode_captions: |
|
displacy: "displaCy + highlighted arguments" |
|
pretty_table: "Pretty Table" |
|
layer_caption_mapping: |
|
labeled_multi_spans: "adus" |
|
binary_relations: "relations" |
|
labeled_partitions: "partitions" |
|
relation_name_mapping: |
|
supports_reversed: "supported by" |
|
contradicts_reversed: "contradicts" |
|
|
|
default_render_mode: "displacy" |
|
default_render_kwargs: |
|
entity_options: |
|
|
|
colors: |
|
OWN_CLAIM: "#009933" |
|
BACKGROUND_CLAIM: "#99ccff" |
|
DATA: "#993399" |
|
colors_hover: |
|
selected: "#ffa" |
|
|
|
tail: |
|
|
|
supports: "#9f9" |
|
|
|
contradicts: "#f99" |
|
|
|
parts_of_same: null |
|
head: null |
|
other: null |
|
|
|
example_text: > |
|
Scholarly Argumentation Mining (SAM) has recently gained attention due to its |
|
potential to help scholars with the rapid growth of published scientific literature. |
|
It comprises two subtasks: argumentative discourse unit recognition (ADUR) and |
|
argumentative relation extraction (ARE), both of which are challenging since they |
|
require e.g. the integration of domain knowledge, the detection of implicit statements, |
|
and the disambiguation of argument structure. |
|
|
|
While previous work focused on dataset construction and baseline methods for |
|
specific document sections, such as abstract or results, full-text scholarly argumentation |
|
mining has seen little progress. In this work, we introduce a sequential pipeline model |
|
combining ADUR and ARE for full-text SAM, and provide a first analysis of the |
|
performance of pretrained language models (PLMs) on both subtasks. |
|
|
|
We establish a new SotA for ADUR on the Sci-Arg corpus, outperforming the previous best |
|
reported result by a large margin (+7% F1). We also present the first results for ARE, and |
|
thus for the full AM pipeline, on this benchmark dataset. Our detailed error analysis reveals |
|
that non-contiguous ADUs as well as the interpretation of discourse connectors pose major |
|
challenges and that data annotation needs to be more consistent. |
|
|