id
stringlengths 1
4
| tokens
sequence | ner_tags
sequence |
---|---|---|
2100 | [
"To",
"verify",
"this",
"guess,",
"we",
"test",
"the",
"fine-tuning",
"performance",
"of",
"a",
"random",
"mapping",
"with",
"different",
"labels",
"selected."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2101 | [
"Allowing",
"multiple",
"labels",
"can",
"resolve",
"mishaps",
"like",
"this",
"and",
"thus",
"improve",
"the",
"final",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2102 | [
"This",
"is",
"due",
"to",
"the",
"format",
"processing",
"in",
"the",
"pretraining",
"of",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2103 | [
"One",
"example",
"is",
"the",
"meaningless",
"</s>",
"(endof-sequence",
"marker)",
"label",
"found",
"by",
"AMuLaP,",
"as",
"shown",
"in",
"Table",
"1."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0
] |
2104 | [
"Under",
"a",
"few-shot",
"setting,",
"the",
"limited",
"number",
"of",
"training",
"examples",
"n",
"and",
"complex",
"training",
"procedure",
"of",
"the",
"backbone",
"model",
"can",
"often",
"bring",
"noise",
"to",
"both",
"automatic",
"label",
"selection",
"and",
"inference."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2105 | [
"However,",
"we",
"find",
"this",
"explanation",
"insufficient",
"for",
"understanding",
"the",
"mechanism",
"behind",
"the",
"improved",
"performance",
"with",
"multiple",
"labels."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2106 | [
"Why",
"Does",
"AMuLaP",
"Work?"
] | [
0,
0,
1,
0
] |
2107 | [
"7",
"Discussion"
] | [
0,
0
] |
2108 | [
"With",
"only",
"one",
"example,",
"AMuLaP",
"achieves",
"decent",
"performance",
"while",
"standard",
"fine-tuning",
"is",
"close",
"to",
"random."
] | [
0,
0,
0,
0,
1,
0,
0,
0,
0,
1,
2,
0,
0,
0,
0
] |
2109 | [
"In",
"addition,",
"the",
"results",
"demonstrate",
"the",
"effectiveness",
"of",
"AMuLaP",
"especially",
"for",
"extreme",
"few-shot",
"settings."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2110 | [
"For",
"MRPC,",
"although",
"the",
"performance",
"of",
"our",
"AMuLaP",
"and",
"standard",
"fine-tuning",
"fluctuate",
"as",
"n",
"increases,",
"in",
"general,",
"AMuLaP",
"with",
"fine-tuning",
"can",
"still",
"achieve",
"comparable",
"performance",
"to",
"standard",
"fine-tuning."
] | [
0,
13,
0,
0,
0,
0,
0,
1,
0,
1,
2,
0,
0,
3,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
1,
2
] |
2111 | [
"For",
"a",
"harder",
"task",
"like",
"MNLI,",
"although",
"the",
"performance",
"of",
"AMuLaP",
"under",
"nonfinetuning",
"setting",
"gradually",
"becomes",
"saturated",
"as",
"n",
"increases,",
"AMuLaP",
"under",
"fine-tuning",
"settings",
"continues",
"to",
"improve",
"as",
"n",
"increases",
"and",
"continues",
"to",
"outperform",
"the",
"standard",
"fine-tuning."
] | [
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0
] |
2112 | [
"For",
"an",
"easier",
"task",
"like",
"SST-2,",
"although",
"only",
"32",
"training",
"examples",
"are",
"used,",
"the",
"performance",
"of",
"our",
"AMuLaP",
"with",
"non-finetuning",
"and",
"fine-tuning",
"is",
"close",
"to",
"saturation",
"and",
"can",
"be",
"comparable",
"to",
"standard",
"fine-tuning",
"on",
"the",
"entire",
"dataset."
] | [
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2113 | [
"When",
"using",
"more",
"than",
"16",
"training",
"examples,",
"AMuLaP",
"under",
"fine-tuning",
"setting",
"still",
"out"
] | [
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0
] |
2114 | [
"For",
"MNLI",
"and",
"SST-2",
"task,",
"AMuLaP",
"outperforms",
"standard",
"fine-tuning",
"when",
"we",
"use",
"no",
"more",
"than",
"16",
"training",
"examples",
"for",
"non-finetuning",
"and",
"fine-tuning",
"setting."
] | [
0,
13,
0,
13,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2115 | [
"Figure",
"2",
"illustrates",
"how",
"standard",
"fine-tuning",
"and",
"our",
"AMuLaP",
"with",
"non-finetuning",
"and",
"fine-tuning",
"compare",
"as",
"n",
"increases."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
3,
0
] |
2116 | [
"Similarly,",
"in",
"this",
"section,",
"we",
"aim",
"to",
"test",
"how",
"AMuLaP",
"scales",
"to",
"different",
"training",
"set",
"sizes",
"n."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
3,
4,
4,
3
] |
2117 | [
"Le",
"Scao",
"and",
"Rush",
"(2021)",
"explore",
"the",
"scaling",
"law",
"of",
"PET",
"(Schick",
"and",
"Schütze,",
"2021a)",
"when",
"using",
"more",
"examples",
"for",
"training."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2118 | [
"6.3",
"Scaling",
"Few-Shot",
"Learning"
] | [
0,
0,
0,
0
] |
2119 | [
"An",
"interesting",
"exception",
"is",
"that",
"for",
"CoLA,",
"the",
"random",
"mapping",
"outperforms",
"all",
"label",
"selection",
"methods",
"in",
"Table",
"2",
"(both",
"manual",
"and",
"automatic)",
"and",
"is",
"close",
"to",
"the",
"fine-tuning",
"baseline."
] | [
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2120 | [
"Moreover,",
"a",
"random",
"label",
"mapping",
"often",
"leads",
"to",
"lower",
"performance",
"than",
"a",
"label",
"mapping",
"selected",
"based",
"on",
"the",
"training",
"set."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2121 | [
"Also,",
"our",
"multi-label",
"strategy",
"is",
"shown",
"to",
"be",
"effective",
"at",
"improving",
"the",
"average",
"GLUE",
"scores",
"by",
"3.6",
"and",
"1.1",
"for",
"non-finetuning",
"and",
"fine-tuning",
"settings,",
"respectively."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
9,
0,
9,
0,
0,
0,
0,
0,
0
] |
2122 | [
"maps",
"to",
"two",
"classes,",
"optimization",
"would",
"be",
"difficult",
"due",
"to",
"the",
"contradiction",
"of",
"supervision",
"signals."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2123 | [
"Notably,",
"deduplication",
"is",
"especially",
"important",
"for",
"prompt-based",
"fine-tuning",
"since",
"if",
"the",
"same",
"label"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2124 | [
"For",
"both",
"non-finetuning",
"and",
"prompt-based",
"fine-tuning",
"settings,",
"our",
"deduplication",
"algorithm",
"can",
"effectively",
"improve",
"the",
"overall",
"performance",
"by",
"1.1",
"and",
"9.9",
"in",
"terms",
"of",
"the",
"GLUE",
"average",
"score,",
"respectively."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
9,
0,
9,
0,
0,
0,
0,
13,
0,
0,
0
] |
2125 | [
"As",
"shown",
"in",
"Table",
"4,",
"we",
"evaluate",
"the",
"effect",
"of",
"each",
"design",
"choice",
"on",
"the",
"GLUE",
"benchmark."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0
] |
2126 | [
"6.2",
"Ablation",
"Study"
] | [
0,
0,
0
] |
2127 | [
"This",
"may",
"explain",
"why",
"AMuLaP",
"works",
"well,",
"especially",
"for",
"the",
"nonfinetuning",
"settings."
] | [
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0
] |
2128 | [
"Additionally,",
"we",
"would",
"like",
"to",
"highlight",
"that",
"AMuLaP",
"mainly",
"collects",
"common",
"words",
"while",
"other",
"methods",
"prefer",
"rare",
"words."
] | [
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2129 | [
"Thus,",
"introducing",
"top-k",
"truncation",
"can",
"resolve",
"the",
"problem."
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2130 | [
"Although",
"AMuLaP",
"outputs",
"three",
"labels",
"that",
"are",
"rated",
"not",
"suitable",
"by",
"the",
"human",
"annotator,",
"it",
"should",
"be",
"noted",
"that",
"all",
"three",
"tokens",
"are",
"ranked",
"low",
"in",
"the",
"candidate",
"set."
] | [
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2131 | [
"AMuLaP",
"achieves",
"interpretability",
"that",
"is",
"competitive",
"to",
"automatic",
"labels",
"obtained",
"by",
"a",
"fine-tuned",
"pretrained",
"language",
"model,",
"measured",
"by",
"the",
"human",
"agreement",
"ratio."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
7,
8,
8
] |
2132 | [
"PETAL-CE",
"suffers",
"from",
"incorrect",
"mappings",
"for",
"“negative”",
"while",
"PETAL-LR",
"occasionally",
"outputs",
"vague",
"labels."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0
] |
2133 | [
"We",
"report",
"the",
"average",
"of",
"5",
"runs",
"along",
"with",
"their",
"standard",
"deviation",
"in",
"the",
"parentheses."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2134 | [
"All",
"of",
"these",
"we",
"use",
"are",
"from",
"https://github.com/princeton-nlp/",
"LM-BFF/tree/main/auto_label_mapping."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2135 | [
"No",
"external",
"data",
"used."
] | [
0,
0,
0,
0
] |
2136 | [
"2The",
"validation",
"scores",
"of",
"all",
"fine-tuned",
"assignments",
"are",
"obtained",
"on",
"as",
"described",
"in",
"Gao",
"et",
"al.",
"(2021)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2137 | [
"We",
"shuffle",
"the",
"labels",
"from",
"each",
"model",
"and",
"ask",
"a",
"human",
"annotator"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2138 | [
"As",
"shown",
"in",
"Table",
"3,",
"we",
"list",
"the",
"10",
"most",
"likely",
"label",
"mappings",
"output",
"by",
"PETAL",
"(Schick",
"et",
"al.,",
"2020),",
"Auto-L",
"(Gao",
"et",
"al.,",
"2021)",
"and",
"AMuLaP",
"for",
"the",
"SST-2",
"dataset,",
"respectively."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
1,
0,
0,
13,
0,
0
] |
2139 | [
"6.1",
"Case",
"Study"
] | [
0,
0,
0
] |
2140 | [
"6",
"Analysis"
] | [
0,
0
] |
2141 | [
"On",
"all",
"tasks",
"except",
"CoLA,",
"AMuLaP",
"outperforms",
"direct",
"few-shot",
"fine-tuning,",
"suggesting",
"that",
"prompting",
"is",
"a",
"promising",
"method",
"for",
"exploiting",
"large",
"pretrained",
"LMs."
] | [
0,
0,
0,
0,
13,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2142 | [
"Under",
"Setting",
"3,",
"AMuLaP",
"FT",
"outperforms",
"all",
"baselines",
"including",
"Auto-L.",
"Generally",
"speaking,",
"methods",
"with",
"parameter",
"update",
"(Setting",
"3)",
"have",
"better",
"performance",
"than",
"those",
"that",
"do",
"not",
"require",
"access",
"to",
"parameters."
] | [
0,
0,
0,
1,
2,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2143 | [
"Additionally,",
"we",
"attempt",
"to",
"replace",
"the",
"predicted",
"token",
"distribution",
"of",
"AMuLaP",
"with",
"the",
"validation",
"score",
"of",
"all",
"fine-tuned",
"assignments",
"(Gao",
"et",
"al.,",
"2021).2",
"With",
"the",
"help",
"of",
"many",
"trials",
"in",
"automatic",
"search,",
"AMuLaP",
"outperforms",
"Auto-L",
"by",
"a",
"considerable",
"margin",
"of",
"3.8",
"in",
"terms",
"of",
"the",
"average",
"score,",
"verifying",
"the",
"versatility",
"of",
"our",
"multi-label",
"mechanism",
"and",
"label",
"selection",
"algorithm."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
1,
0,
0,
0,
0,
0,
9,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2144 | [
"Notably,",
"AMuLaP",
"even",
"outperforms",
"Auto-L",
"by",
"1.3",
"without",
"using",
"any",
"external",
"model",
"or",
"data."
] | [
0,
1,
0,
0,
1,
0,
9,
0,
0,
0,
0,
0,
0,
0
] |
2145 | [
"Under",
"Setting",
"2,",
"compared",
"to",
"variants",
"of",
"PETAL",
"(Schick",
"et",
"al.,",
"2020),",
"AMuLaP",
"has",
"an",
"advantage",
"of",
"5.8",
"and",
"8.5",
"in",
"terms",
"of",
"the",
"average",
"score",
"over",
"CE",
"and",
"LR,",
"respectively."
] | [
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
1,
0,
0,
0,
0,
9,
0,
9,
0,
0,
0,
0,
0,
0,
0,
1,
0,
1,
0
] |
2146 | [
"outperforms",
"GPT-3-style",
"in-context",
"learning",
"by",
"4.5",
"in",
"terms",
"of",
"the",
"average",
"score",
"and",
"outperforms",
"zero-shot",
"inference",
"with",
"manually",
"designed",
"labels",
"by",
"2.4."
] | [
0,
1,
0,
0,
0,
9,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
9
] |
2147 | [
"Under",
"Setting",
"1,",
"AMuLaP"
] | [
0,
0,
0,
1
] |
2148 | [
"We",
"demonstrate",
"experimental",
"results",
"under",
"three",
"settings",
"in",
"Table",
"2."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2149 | [
"5.2",
"Experimental",
"Results"
] | [
0,
0,
0
] |
2150 | [
"Following",
"(Gao",
"et",
"al.,",
"2021),",
"we",
"grid",
"search",
"the",
"learning",
"rate",
"from",
"{1e-5,",
"2e-5,",
"5e-5}",
"and",
"batch",
"size",
"from",
"{2,",
"4,",
"8}."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
0,
5,
5,
5,
0,
3,
3,
0,
5,
5,
5
] |
2151 | [
"For",
"prompt-based",
"fine-tuning",
"(Setting",
"3),",
"where",
"we",
",",
"we",
"search",
"k",
"in",
"a",
"smaller",
"fine-tune",
"the",
"model",
"L",
"1,",
"2,",
"4,",
"8,",
"16",
"due",
"to",
"the",
"increased",
"compuspace",
"}",
"tational",
"overhead."
] | [
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
5,
5,
5,
5,
5,
0,
0,
0,
0,
0,
0,
0,
0
] |
2152 | [
"Note",
"that",
"for",
"settings",
"that",
"do",
"}",
"{",
",",
"search",
"over",
"k",
"is",
"not",
"update",
"the",
"parameters",
"of",
"fast,",
"as",
"we",
"only",
"need",
"to",
"run",
"the",
"model",
"once",
"and",
"cache",
"the",
"distribution",
"of",
"the",
"[MASK]",
"token."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2153 | [
"When",
"selecting",
"k,",
"if",
"there",
"are",
"multiple",
"k",
"with",
"identical",
"performance",
"(which",
"happens",
"occasionally",
"given",
"there",
"are",
"only",
"16",
"examples",
"for",
"each",
"class",
"in",
"Ddev",
"),",
"we",
"always",
"choose",
"the",
"largest",
"k.",
"For",
"Settings",
"1",
"and",
"2,",
"we",
"search",
"k",
"over",
"1,",
"2,",
"4,",
".",
".",
"."
] | [
0,
0,
3,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0,
0,
0,
0,
0,
0,
0,
3,
0,
5,
5,
5,
0,
0,
0
] |
2154 | [
"Implementation",
"Details",
"We",
"implement",
"AMuLaP",
"based",
"on",
"Hugging",
"Face",
"Transformers",
"(Wolf",
"et",
"al.,",
"2020)."
] | [
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2155 | [
"This",
"setting",
"is",
"for",
"fair",
"comparison",
"with",
"conventional",
"finetuning,",
"prompt-based",
"fine-tuning",
"with",
"manual",
"prompts,",
"Auto-L",
"(Gao",
"et",
"al.,",
"2021)",
"and",
"PETAL",
"(Schick",
"et",
"al.,",
"2020)."
] | [
0,
0,
0,
0,
0,
0,
0,
1,
2,
0,
0,
0,
1,
2,
1,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0
] |
2156 | [
"Ddev",
"in",
"the",
"same",
"way",
"as",
"Setting",
"2",
"but",
"fine-tune",
"the",
"parameters",
"of",
"the",
"language",
"model",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2157 | [
"Dtrain",
"and"
] | [
0,
0
] |
2158 | [
"•",
"Setting",
"3:",
"We",
"use"
] | [
0,
0,
0,
0,
0
] |
2159 | [
"This",
"setting",
"is",
"for",
"fair",
"comparison",
"with",
"Auto-L",
"(Gao",
"et",
"al.,",
"2021)",
"and",
"PETAL",
"(Schick",
"et",
"al.,",
"2020)."
] | [
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0
] |
2160 | [
"The",
"paand",
"an",
"additional",
"rameters",
"of",
"are",
"not",
"updated."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2161 | [
"•",
"Setting",
"2:",
"We",
"use",
"Dtrain",
"for",
"label",
"selection",
"Ddev",
"for",
"k",
"tuning."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0
] |
2162 | [
"The",
"parameters",
"Ddev",
"is",
"not",
"used."
] | [
0,
0,
0,
0,
0,
0
] |
2163 | [
"Dtrain",
"alone",
"for",
"both",
"label",
"selection",
"and",
"tuning",
"k."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
3
] |
2164 | [
"•",
"Setting",
"1:",
"We",
"only",
"use"
] | [
0,
0,
0,
0,
0,
0
] |
2165 | [
"L"
] | [
0
] |
2166 | [
"To",
"fairly",
"compare",
"with",
"different",
"baselines,",
"we",
"consider",
"the",
"following",
"three",
"settings:"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2167 | [
"We",
"also",
"report",
"the",
"standard",
"deviation",
"for",
"each",
"result."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2168 | [
"Following",
"Gao",
"et",
"al.",
"(2021),",
"the",
"original",
"development",
"split",
"of",
"each",
"dataset",
"is",
"used",
"as",
"the",
"test",
"set",
"in",
"our",
"experiments."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2169 | [
"For",
"each",
"reported",
"result,",
"we",
"measure",
"average",
"performance",
"across",
"5",
"different",
"randomly",
"sampled",
"Dtrain",
"and",
"Ddev",
"splits."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2170 | [
"We",
"use",
"RoBERTalarge",
"(Liu",
"et",
"al.,",
"2019)",
"as",
"the",
"backbone",
"LM",
"."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2171 | [
"We",
"set",
"k",
"=",
"16",
"throughout",
"all",
"experiments."
] | [
0,
0,
3,
0,
5,
0,
0,
0
] |
2172 | [
"We",
"sample",
"n",
"training",
"examples",
"and",
"n",
"development",
"examples",
"per",
"class."
] | [
0,
0,
5,
3,
4,
0,
5,
3,
4,
0,
0
] |
2173 | [
"Task",
"Setup",
"We",
"closely",
"follow",
"the",
"setup",
"in",
"Gao",
"et",
"al.",
"(2021)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2174 | [
"We",
"use",
"the",
"officially",
"released",
"code",
"and",
"same",
"hyperparameters",
"for",
"this",
"baseline."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2175 | [
"The",
"detailed",
"description",
"can",
"be",
"found",
"in",
"Appendix",
"A.",
"Note",
"that",
"the",
"results",
"of",
"this",
"baseline",
"is",
"different",
"from",
"those",
"reported",
"in",
"Table",
"3",
"of",
"Gao",
"et",
"al.",
"(2021)",
"since",
"they",
"search",
"for",
"both",
"templates",
"and",
"label",
"mapping",
"whereas",
"we",
"fix",
"the",
"templates",
"and",
"search",
"for",
"the",
"label",
"mapping",
"alone,",
"for",
"the",
"sake",
"of",
"fair",
"comparison."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2176 | [
"•",
"Auto-L",
"(Gao",
"et",
"al.,",
"2021):",
"the",
"automatic",
"label",
"searching",
"method",
"with",
"an",
"external",
"pretrained",
"language",
"model,",
"RoBERTa-large",
"(Liu",
"et",
"al.,",
"2019)."
] | [
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2177 | [
"•",
"PETAL-LR",
"(Schick",
"et",
"al.,",
"2020):",
"the",
"variant",
"of",
"PETAL",
"using",
"the",
"likelihood",
"ratio",
"metric."
] | [
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2178 | [
"•",
"PETAL-CE",
"(Schick",
"et",
"al.,",
"2020):",
"the",
"variant",
"of",
"PETAL",
"using",
"the",
"cross-entropy",
"metric."
] | [
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2179 | [
"designed",
"prompts",
"in",
"Gao",
"et",
"al.",
"(2021)."
] | [
0,
0,
0,
0,
0,
0,
0
] |
2180 | [
"•",
"Manual",
"prompts:",
"we",
"use",
"the",
"human"
] | [
0,
1,
2,
0,
0,
0,
0
] |
2181 | [
"•",
"GPT-3-style",
"in-context",
"learning",
"(Brown",
"et",
"al.,",
"2020):",
"present",
"a",
"few",
"examples",
"to",
"the",
"language",
"model",
"and",
"make",
"it",
"directly",
"predict",
"the",
"next",
"token",
"as",
"the",
"prediction."
] | [
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2182 | [
"•",
"Majority:",
"always",
"predict",
"the",
"majority",
"class"
] | [
0,
1,
0,
0,
0,
0,
0
] |
2183 | [
"We",
"compare",
"our",
"method",
"to",
"various",
"baselines:"
] | [
0,
0,
0,
0,
0,
0,
0
] |
2184 | [
"Baselines"
] | [
0
] |
2185 | [
"The",
"metrics",
"for",
"each",
"dataset",
"are",
"indicated",
"in",
"Table",
"2."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2186 | [
"We",
"use",
"the",
"manual",
"templates",
"in",
"Gao",
"et",
"al.",
"(2021),",
"as",
"listed",
"in",
"Table",
"1."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2187 | [
"Natural",
"Language",
"Inference",
"Mismatched",
"(MNLImm)",
"(Williams",
"et",
"al.,",
"2018),",
"Question",
"Natural",
"Language",
"Inference",
"(QNLI)",
"(Rajpurkar",
"et",
"al.,",
"2016)",
"and",
"Recognizing",
"Textual",
"Entailment",
"(RTE)",
"(Wang",
"et",
"al.,",
"2019)",
"for",
"the",
"Natural",
"Language",
"Inference",
"(NLI)",
"task;",
"The",
"Corpus",
"of",
"Linguistic",
"Acceptability",
"(CoLA)",
"(Warstadt",
"et",
"al.,",
"2019)",
"for",
"Linguistic",
"Acceptability."
] | [
0,
0,
0,
0,
13,
0,
0,
0,
0,
13,
14,
14,
14,
13,
0,
0,
0,
0,
0,
13,
14,
14,
13,
0,
0,
0,
0,
0,
0,
11,
12,
12,
11,
0,
13,
14,
14,
14,
14,
13,
0,
0,
0,
0,
0,
11,
12
] |
2188 | [
"We",
"report",
"the",
"average",
"of",
"5",
"runs",
"along",
"with",
"their",
"standard",
"deviation",
"in",
"the",
"parentheses."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2189 | [
"Specifically,",
"we",
"test",
"on",
"Microsoft",
"Research",
"Paraphrase",
"Matching",
"(MRPC)",
"(Dolan",
"and",
"Brockett,",
"2005),",
"Quora",
"Question",
"Pairs",
"(QQP)",
"for",
"Paraphrase",
"Similarity",
"Matching;",
"Stanford",
"Sentiment",
"Treebank",
"(SST-2)",
"(Socher",
"et",
"al.,",
"2013)",
"for",
"Sentiment",
"Classification;",
"Multi-Genre",
"Natural",
"Language",
"Inference",
"Matched",
"(MNLI-m),",
"Multi-Genre"
] | [
0,
0,
0,
0,
13,
14,
14,
14,
13,
0,
0,
0,
0,
13,
14,
14,
13,
0,
11,
12,
12,
13,
14,
14,
13,
0,
0,
0,
0,
0,
11,
12,
13,
14,
14,
14,
14,
13,
0
] |
2190 | [
"Datasets",
"We",
"evaluate",
"seven",
"classification",
"tasks",
"of",
"the",
"GLUE",
"benchmark",
"(Wang",
"et",
"al.,",
"2019)."
] | [
0,
0,
0,
0,
11,
0,
0,
0,
13,
0,
0,
0,
0,
0
] |
2191 | [
"5.1",
"Experimental",
"Setting"
] | [
0,
0,
0
] |
2192 | [
"5",
"Experiments"
] | [
0,
0
] |
2193 | [
"Then,",
"we",
"can",
"truncate",
"the",
"candidate",
"sets",
"of",
"each",
"class",
"and",
"select",
"the",
"k",
"most",
"likely",
"tokens",
"from",
"each",
"set."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2194 | [
"To",
"ensure",
"this,",
"we",
"have",
"to",
"iterate",
"over",
"the",
"vocabulary",
"and",
"check",
"that",
"for",
"every",
"token."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2195 | [
"Thus,",
"we",
"want",
"to",
"make",
"sure",
"that",
"(2)",
"Each",
"token",
"only",
"belongs",
"to",
"at",
"most",
"one",
"label",
"set",
"where",
"it",
"has",
"the",
"highest",
"probability."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2196 | [
"For",
"example,",
"if",
"we",
"simply",
"take",
"the",
"10",
"most",
"likely",
"tokens",
"for",
"the",
"SST-2",
"dataset",
"(Socher",
"et",
"al.,",
"2013),",
"we",
"would",
"find",
"“good”",
"in",
"both",
"positive",
"and",
"negative",
"label",
"sets,",
"although",
"it",
"is",
"ranked",
"second",
"place",
"in",
"the",
"positive",
"set",
"and",
"ninth",
"in",
"the",
"negative",
"set."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2197 | [
"However,",
"in",
"practice,",
"we",
"would",
"find",
"common",
"words",
"in",
"more",
"than",
"one",
"label",
"set."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2198 | [
"A",
"simple",
"solution",
"is",
"to",
"select",
"the",
"k",
"most",
"likely",
"tokens",
"predicted",
"for",
"the",
"[MASK]",
"token",
"in",
"the",
"training",
"examples",
"of",
"each",
"class",
"y."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2199 | [
"For",
"example,",
"in",
"a",
"sentiment",
"classification",
"task,",
"we",
"would",
"like",
"to",
"see",
"positive",
"words",
"in",
"the",
"label",
"set",
"of",
"the",
"“positive”",
"class",
"while",
"negative",
"words",
"in",
"the",
"label",
"set",
"of",
"the",
"“negative”",
"class."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |