id
stringlengths 1
4
| tokens
sequence | ner_tags
sequence |
---|---|---|
2300 | [
"For",
"sentence-pair",
"tasks,",
"we",
"hypothesize",
"that",
"inserting",
"prompt",
"into",
"position",
"1",
"can",
"better",
"align",
"the",
"two",
"input",
"sentences."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2301 | [
"This",
"result",
"is",
"intuitive",
"for",
"single-sentence",
"tasks",
"since",
"prompt",
"in",
"position",
"0",
"can",
"be",
"regarded",
"as",
"the",
"premise",
"and",
"original",
"input",
"sentence",
"as",
"the",
"hypothesis."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2302 | [
"Looking",
"at",
"the",
"prompt",
"position",
"in",
"downstream",
"tasks",
"first,",
"Figure",
"7(a)",
"shows",
"that",
"for",
"both",
"standard",
"prompt",
"tuning",
"and",
"our",
"proposed",
"method,",
"the",
"best",
"position",
"is",
"0",
"for",
"single-sentence",
"tasks",
"and",
"1",
"for",
"sentence-pair",
"tasks."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2303 | [
"We",
"conduct",
"a",
"comprehensive",
"study",
"of",
"the",
"prompt",
"position",
"for",
"our",
"proposed",
"method",
"in",
"both",
"supplementary",
"training",
"and",
"downstream",
"fine-tuning",
"phases."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2304 | [
"A.5.2",
"Prompt",
"Position",
"As",
"we",
"discussed",
"in",
"Section",
"A.4,",
"the",
"prompt",
"position",
"has",
"a",
"direct",
"impact",
"on",
"the",
"prediction",
"results."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2305 | [
"One",
"surprising",
"result",
"is",
"that",
"the",
"mixed",
"model",
"of",
"Residual",
"and",
"LayerNorm",
"has",
"significantly",
"poorer",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2306 | [
"We",
"observe",
"that",
"adding",
"LayerNorm",
"slightly",
"improves",
"the",
"voting",
"results,",
"while",
"residual",
"performs",
"slightly",
"worse."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2307 | [
"Note",
"that,",
"to",
"balance",
"the",
"token",
"embedding",
"and",
"sentence",
"embedding,",
"we",
"apply",
"LayerNorm",
"to",
"each",
"embedding",
"first,",
"then",
"after",
"the",
"add-up,",
"use",
"LayerNorm",
"again",
"to",
"control",
"the",
"generated",
"tokens."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2308 | [
"et",
"al.,",
"2016)",
"is",
"also",
"added",
"to",
"normalize",
"the",
"generated",
"token",
"embedding;",
"(iii)",
"residual",
"+",
"layerNorm:",
"a",
"mixed",
"model",
"that",
"uses",
"both",
"the",
"residual",
"component",
"and",
"LayerNorm."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2309 | [
"We",
"explore",
"three",
"different",
"architectures",
"for",
"the",
"proposed",
"PHM-based",
"generator:",
"(i)",
"Residual:",
"a",
"residual",
"structure",
"(He",
"et",
"al.,",
"2016)",
"is",
"applied",
"to",
"add",
"the",
"sentence",
"representation",
"to",
"each",
"generated",
"tokens;",
"(ii)",
"LayerNorm:",
"layer",
"normalization",
"(Ba"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2310 | [
"A.5.1",
"Generator",
"Architecture",
"Exploration"
] | [
0,
0,
0,
0
] |
2311 | [
"A.5",
"Ablation",
"study",
"for",
"single-layer",
"IDPG"
] | [
0,
0,
0,
0,
0,
1
] |
2312 | [
"We",
"report",
"average",
"results",
"and",
"voting",
"results",
"across",
"5",
"runs."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2313 | [
"We",
"double",
"the",
"MNLI",
"dataset",
"by",
"reordering",
"the",
"two",
"sentences",
"on",
"one",
"shard,",
"and",
"use",
"the",
"doubled",
"dataset",
"during",
"intermediate",
"training."
] | [
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2314 | [
"We",
"further",
"reduce",
"the",
"distribution",
"difference",
"by",
"reconstructing",
"the",
"supplementary",
"training",
"data."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2315 | [
"As",
"shown",
"in",
"Figure",
"6,",
"we",
"test",
"five",
"different",
"insertion",
"positions",
"(Pos",
"0–4)",
"for",
"sentence",
"pair",
"tasks",
"and",
"three",
"different",
"positions",
"(Pos",
"0,",
"1,",
"4)",
"for",
"single",
"sentence",
"tasks."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2316 | [
"In",
"this",
"way,",
"the",
"length",
"difference",
"in",
"MNLI",
"and",
"MRPC",
"becomes",
"more",
"balanced:",
"4.8",
"vs.",
"0.6",
"+",
"5",
"=",
"5.6."
] | [
0,
0,
0,
0,
0,
0,
0,
13,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2317 | [
"Method"
] | [
0
] |
2318 | [
"For",
"all",
"the",
"other",
"tasks,",
"we",
"report",
"accuracy."
] | [
0,
0,
0,
0,
0,
0,
0,
7
] |
2319 | [
"We",
"report",
"the",
"average",
"of",
"accuracy",
"and",
"F1",
"for",
"both",
"MRPC",
"and",
"QQP,",
"and",
"average",
"of",
"Pearson",
"and",
"Spearman",
"correlation",
"coefficients",
"for",
"STS-B."
] | [
0,
0,
0,
0,
0,
7,
0,
7,
0,
0,
13,
0,
13,
0,
0,
0,
7,
8,
8,
8,
8,
0,
13
] |
2320 | [
"We",
"report",
"average",
"results",
"across",
"5",
"runs",
"with",
"different",
"initialization."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2321 | [
"Each",
"methods",
"are",
"evaluated",
"on",
"full",
"test",
"sets",
"(dev",
"sets",
"for",
"GLUE",
"tasks)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0
] |
2322 | [
"Then,",
"when",
"fine-tuning",
"downstream",
"tasks"
] | [
0,
0,
0,
0,
0
] |
2323 | [
"For",
"example,",
"assuming",
"that",
"we",
"are",
"adding",
"a",
"prompt",
"with",
"a",
"length",
"t",
"=",
"5",
"after",
"the",
"second",
"sentence",
"in",
"the",
"supplementary",
"training",
"stage",
"on",
"MNLI."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
3,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13
] |
2324 | [
"One",
"natural",
"solution",
"to",
"smooth",
"the",
"length",
"distribution",
"difference",
"between",
"tasks",
"is",
"to",
"insert",
"prompt",
"in",
"both",
"supplementary",
"training",
"and",
"downstream",
"finetuning",
"stage."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2325 | [
"For",
"example,",
"the",
"length",
"of",
"the",
"first",
"sentence",
"in",
"MNLI",
"is",
"9.8",
"longer",
"than",
"the",
"second",
"sentence",
"on",
"average,",
"while",
"this",
"length",
"difference",
"in",
"MRPC",
"is",
"only",
"0.6."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
0
] |
2326 | [
"Figure",
"5",
"provides",
"a",
"comprehensive",
"statistic",
"among",
"all",
"sentence",
"pair",
"tasks",
"in",
"GLUE",
"benchmark."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0
] |
2327 | [
"However,",
"a",
"drawback",
"of",
"supplementary",
"training",
"is",
"that",
"if",
"the",
"data",
"distribution",
"of",
"the",
"downstream",
"tasks",
"is",
"quite",
"different",
"from",
"the",
"supplementary",
"training",
"task,",
"i.e.,",
"MRPC",
"vs.",
"MNLI",
"(Wang",
"et",
"al.,",
"2019),",
"it",
"may",
"harm",
"the",
"downstream",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2328 | [
"Following",
"this",
"idea,",
"we",
"also",
"conduct",
"intermediate",
"training",
"for",
"single-layer",
"IDPG."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1
] |
2329 | [
"According",
"to",
"previous",
"works",
"(Phang",
"et",
"al.,",
"2018;",
"Wang",
"et",
"al.,",
"2021),",
"supplementing",
"pre-trained",
"LMs",
"with",
"rich",
"data",
"helps",
"tasks",
"with",
"limited",
"labels",
"and",
"stabilizes",
"downstream",
"fine-tuning."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2330 | [
"A.4",
"Supplementary",
"Training",
"for",
"Single-layer"
] | [
0,
0,
0,
0,
0
] |
2331 | [
"While",
"in",
"Table",
"1,",
"we",
"tune",
"the",
"learning",
"4",
"4,",
"1e−",
"3,",
"5e−",
"to",
"make",
"the",
"rate",
"from",
"}",
"fair",
"comparison",
"with",
"other",
"models."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
6,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2332 | [
"M-IDPG-PHM),",
"this",
"is",
"because",
"we",
"tune",
"the",
"learning",
"rate",
"more",
"carefully",
"in",
"Table",
"6",
"(lr",
"2,",
"7e−",
"4,",
"3,",
"1e−",
"1e−",
"{",
"4",
"1e−",
")",
"to",
"seek",
"the",
"best",
"performance",
"each",
"model",
"can",
"reach."
] | [
1,
0,
0,
0,
0,
0,
0,
3,
4,
0,
0,
0,
0,
0,
3,
5,
5,
6,
5,
5,
5,
0,
5,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2333 | [
"Note",
"that",
"the",
"M",
"version",
"model",
"with",
"m",
"=",
"16",
"and",
"previous",
"layer",
"as",
"input",
"one",
"is",
"slightly",
"higher",
"than",
"the",
"results",
"shown",
"in",
"Table",
"1(Cf."
] | [
0,
0,
0,
0,
0,
0,
0,
3,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2334 | [
"We",
"provide",
"a",
"detailed",
"result",
"table",
"for",
"all",
"compared",
"methods",
"in",
"Section",
"4.5.3."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2335 | [
"A.3",
"Detailed",
"results",
"for",
"Multi-layer",
"Architecture",
"Exploration"
] | [
0,
0,
0,
0,
0,
0,
0
] |
2336 | [
"We",
"provide",
"a",
"detailed",
"information",
"in",
"Table",
"5",
"for",
"10",
"NLU",
"datasets",
"we",
"used."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
0,
0,
0
] |
2337 | [
"A.2",
"Datasets"
] | [
0,
0
] |
2338 | [
"Total",
"parameters:",
"1024",
"×",
"24",
"=",
"216K.",
"1024"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2339 | [
"M-IDPG-DNN:",
"hidden",
"size",
"d",
"=",
"1024,",
"generator",
"hidden",
"size",
"m",
"=",
"16,",
"prompt",
"length",
"t",
"=",
"5."
] | [
1,
3,
4,
3,
0,
5,
3,
4,
4,
3,
5,
6,
3,
4,
3,
5,
6
] |
2340 | [
"First",
"PHM",
"layer",
"16",
"+",
"16",
"=",
"1K",
"paramW1",
"takes",
"1024/16",
"eters,",
"second",
"PHM",
"layer",
"W2",
"takes",
"16/16",
"1024/16",
"1024",
"the",
"shared",
"matrix",
"Ai",
"takes",
"16316",
"parameters:"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2341 | [
"=",
"5,",
"user",
"defined",
"n",
"=",
"16",
"(Cf.",
"Equation",
"4)."
] | [
0,
5,
0,
0,
3,
0,
5,
0,
0,
0
] |
2342 | [
"M-IDPG-PHM:",
"hidden",
"size",
"d",
"=",
"1024,",
"generator",
"hidden",
"size",
"m",
"=",
"16,",
"prompt",
"length",
"t"
] | [
1,
3,
4,
3,
0,
5,
3,
4,
4,
3,
0,
5,
3,
4,
3
] |
2343 | [
"First",
"PHM",
"4",
"+",
"16",
"=",
"1216",
"palayer",
"W1",
"takes",
"300/4",
"×",
"rameters,",
"second",
"PHM",
"layer",
"W2",
"takes",
"16/4",
"1024/4",
"1024",
"the",
"shared",
"matrix",
"Ai",
"takes",
"43",
"parameters:",
"141K."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2344 | [
"×",
"M-IDPG-PHM-GloVe:",
"input",
"vector",
"size",
"300,",
"generator",
"hidden",
"size",
"m",
"=",
"16,",
"prompt",
"length",
"t",
"=",
"5,",
"user",
"defined",
"n",
"=",
"4",
"(Cf.",
"Equation",
"4)."
] | [
0,
1,
3,
4,
4,
5,
3,
4,
4,
3,
0,
5,
3,
4,
3,
0,
5,
0,
3,
4,
0,
5,
0,
0,
0
] |
2345 | [
"S-IDPG-DNN:",
"hidden",
"size",
"d",
"=",
"1024,",
"generator",
"hidden",
"size",
"m",
"=",
"256,",
"prompt",
"length",
"t",
"=",
"5."
] | [
1,
3,
4,
3,
0,
5,
3,
4,
4,
3,
0,
5,
3,
4,
3,
0,
5
] |
2346 | [
"First",
"PHM",
"layer",
"W1",
"takes",
"1024/16",
"16",
"+",
"256",
"=",
"16.25K",
"256/16",
"parameters,",
"second",
"PHM",
"layer",
"W2",
"takes",
"256/16",
"×",
"1024",
"=",
"85K",
"parameters,",
"5",
"×",
"the",
"shared",
"matrix",
"Ai",
"takes",
"163",
"=",
"4K",
"(Note",
"we",
"use",
"one",
"shared",
"matrix",
"in",
"single",
"version",
"IDPG)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1
] |
2347 | [
"P-tuning",
"v2:",
"prompt",
"length",
"t",
"=",
"5,",
"inserted",
"lay1024",
"=",
"120K.",
"S-IDPG-PHM:",
"hidden",
"size",
"d",
"=",
"1024,",
"generator",
"hidden",
"size",
"m",
"=",
"256,",
"prompt",
"length",
"t",
"=",
"5,",
"user",
"defined",
"n",
"=",
"16",
"(Cf.",
"Equation",
"4)."
] | [
1,
2,
3,
4,
3,
0,
5,
0,
0,
0,
0,
1,
3,
4,
3,
0,
5,
3,
4,
4,
3,
0,
5,
3,
4,
3,
0,
5,
0,
0,
3,
0,
5,
0,
0,
0
] |
2348 | [
"Prompt-tuning-134:",
"prompt",
"length",
"t",
"=",
"134."
] | [
1,
3,
4,
3,
0,
5
] |
2349 | [
"Prompt-tuning:",
"prompt",
"length",
"t",
"=",
"5."
] | [
1,
3,
4,
3,
0,
5
] |
2350 | [
"Total",
"parameters:",
"(1024",
"2",
"=",
"1.55M.",
"16"
] | [
0,
0,
0,
0,
0,
0,
0
] |
2351 | [
"size",
"m",
"=",
"16."
] | [
0,
3,
0,
5
] |
2352 | [
"Adapter:",
"hidden",
"size",
"d",
"=",
"1024,",
"adapter",
"hidden",
"16",
"+",
"16",
"+"
] | [
1,
3,
4,
3,
0,
5,
1,
2,
5,
0,
0,
0
] |
2353 | [
"24",
"4"
] | [
0,
0
] |
2354 | [
"Down-project",
"2",
"=",
"48K,",
"downsi",
"matrix",
"takes",
"1024/4",
"4",
"×",
"project",
"ti",
"matrix",
"takes",
"16/4",
"2",
"=",
"0.75K,",
"24",
"×",
"×",
"2",
"=",
"0.75K,",
"up-project",
"hidden",
"bias",
"takes",
"16",
"24",
"si",
"and",
"ti",
"matrix",
"takes",
"the",
"same",
"number",
"of",
"parameters",
"as",
"down-projector,",
"the",
"output",
"bias",
"takes",
"2",
"=",
"48K,",
"the",
"shared",
"matrix",
"Ai",
"takes",
"1024",
"×",
"43",
"2",
"=",
"3K.",
"Total",
"parameters:",
"48",
"+",
"0.75",
"+",
"0.75",
"+",
"48",
"+",
"0.75",
"+",
"48",
"+",
"3",
"=",
"149.25K."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2355 | [
"Compacter:",
"hidden",
"size",
"d",
"=",
"1024,",
"adapter",
"hidden",
"size",
"m",
"=",
"16,",
"user",
"defined",
"n",
"=",
"4,",
"each",
"transformer",
"layer",
"inserts",
"2",
"compacters."
] | [
0,
0,
0,
3,
0,
5,
3,
4,
4,
3,
0,
5,
0,
0,
3,
0,
5,
0,
0,
0,
0,
5,
3
] |
2356 | [
"We",
"report",
"the",
"detailed",
"model",
"hyperparameters",
"for",
"each",
"method",
"in",
"Table",
"1",
"and",
"illustrate",
"how",
"numbers",
"in",
"Table",
"2",
"are",
"computed."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2357 | [
"A.1.2",
"Model",
"hyperparameters"
] | [
0,
0,
0
] |
2358 | [
"For",
"all",
"adapter-based",
"and",
"prompt-based",
"methods,",
"we",
"train",
"them",
"more",
"sufficiently",
"(with",
"fifty",
"epochs)",
"on",
"small",
"datasets",
"(i.e.,",
"MPQA,",
"Subj,",
"CR,",
"MR,",
"RTE,",
"MRPC,",
"STS-B)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
13,
13,
13,
13,
13,
13
] |
2359 | [
"Note",
"that",
"for",
"both",
"transformer",
"fine-tuning",
"methods,",
"including",
"RoBERTa",
"(Liu",
"et",
"al.,",
"2019)",
"and",
"EFL",
"(Wang",
"et",
"al.,",
"2021),",
"we",
"follow",
"their",
"official",
"training",
"instructions,",
"i.e.,",
"using",
"a",
"polynomial",
"learning",
"rate",
"scheduler",
"with",
"6%",
"steps",
"to",
"warm",
"up",
"and",
"tuning",
"for",
"ten",
"epochs."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
3,
0,
0,
0,
0,
0,
0,
5,
3
] |
2360 | [
"The",
"detailed",
"model",
"hyperparameters",
"are",
"listed",
"in",
"Table",
"4."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2361 | [
"We",
"use",
"RoBERTa-Large",
"(Liu",
"et",
"al.,",
"2019)",
"model",
"implemented",
"by",
"Fairseq",
"(Ott",
"et",
"al.,",
"2019)",
"as",
"our",
"basic",
"model."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2362 | [
"A",
"Appendix"
] | [
0,
0
] |
2363 | [
"References"
] | [
0
] |
2364 | [
"Acknowledgment"
] | [
0
] |
2365 | [
"This",
"work",
"provided",
"a",
"new",
"research",
"angle",
"for",
"prompt-tuning",
"of",
"a",
"pre-trained",
"language",
"model."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
11,
0,
0,
0,
0,
0
] |
2366 | [
"It",
"is",
"also",
"on",
"par",
"with",
"the",
"lightweight",
"adapter",
"tuning",
"methods",
"such",
"as",
"Compacter",
"while",
"using",
"a",
"similar",
"amount",
"of",
"trainable",
"parameters."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0
] |
2367 | [
"Despite",
"adding",
"fewer",
"parameters",
"than",
"prompt",
"tuning,",
"IDPG",
"shows",
"consistent",
"improvement."
] | [
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0
] |
2368 | [
"Parameterized",
"Hypercomplex",
"Multiplication",
"(PHM)",
"is",
"applied",
"to",
"shrink",
"the",
"training",
"parameters",
"in",
"our",
"prompt",
"generator,",
"which",
"helps",
"us",
"build",
"an",
"extreme",
"lightweight",
"generation",
"model."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2369 | [
"Our",
"method",
"first",
"factors",
"in",
"an",
"instance-dependent",
"prompt,",
"which",
"is",
"robust",
"to",
"data",
"variance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2370 | [
"We",
"have",
"introduced",
"IDPG,",
"an",
"instance-dependent",
"prompt",
"generation",
"model",
"that",
"generalizes",
"better",
"than",
"the",
"existing",
"prompt",
"tuning",
"methods."
] | [
0,
0,
0,
1,
0,
1,
2,
2,
0,
0,
0,
0,
0,
0,
0,
11,
12,
0
] |
2371 | [
"6",
"Conclusion",
"and",
"Discussion"
] | [
0,
0,
0,
0
] |
2372 | [
"Our",
"proposed",
"method,",
"IDPG,",
"is",
"the",
"first",
"prompt",
"generator",
"that",
"is",
"not",
"only",
"taskspecific",
"but",
"also",
"instance-specific."
] | [
0,
0,
1,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2373 | [
"All",
"these",
"methods",
"focus",
"on",
"task-specific",
"prompt",
"optimization."
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2374 | [
"Two",
"contemporaneous",
"works",
"–",
"prompt",
"tuning",
"(Lester",
"et",
"al.,",
"2021)",
"and",
"P-tuning",
"(Liu",
"et",
"al.,",
"2021b),",
"interleave",
"the",
"training",
"parameters",
"in",
"the",
"input",
"embedding",
"layer",
"instead",
"of",
"each",
"transformer",
"layer."
] | [
0,
0,
0,
0,
1,
2,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2375 | [
"Prefix",
"tuning",
"(Li",
"and",
"Liang,",
"2021)",
"and",
"P-tuningv2",
"(Liu",
"et",
"al.,",
"2021a)",
"prepend",
"a",
"sequence",
"of",
"trainable",
"embeddings",
"at",
"each",
"transformer",
"layer",
"and",
"optimizes",
"them."
] | [
1,
2,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2376 | [
"Recently,",
"several",
"researchers",
"have",
"proposed",
"continuous",
"prompts",
"training",
"to",
"overcome",
"the",
"challenges",
"in",
"discrete",
"prompt",
"searching."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2377 | [
"Existing",
"works",
"including",
"LMBFF",
"(Gao",
"et",
"al.,",
"2021a;",
"Wang",
"et",
"al.,",
"2021)",
"explored",
"the",
"prompt",
"searching",
"in",
"a",
"few-shot",
"setting."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2378 | [
"Prompting:",
"Hand-crafted",
"prompts",
"were",
"shown",
"to",
"be",
"helpful",
"to",
"adapt",
"generation",
"in",
"GPT-3",
"(Brown",
"et",
"al.,",
"2020)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2379 | [
"et",
"al.,",
"2021)",
"substitutes",
"the",
"down-projector",
"and",
"upprojector",
"matrices",
"by",
"a",
"sum",
"of",
"Kronecker",
"products,",
"reducing",
"the",
"parameters",
"by",
"a",
"large",
"margin",
"while",
"maintaining",
"the",
"overall",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2380 | [
"On",
"the",
"GLUE",
"benchmark,",
"adapters",
"attain",
"within",
"0.4%",
"of",
"the",
"performance",
"of",
"full",
"fine-tuning",
"by",
"only",
"training",
"3.6%",
"parameters",
"per",
"task."
] | [
0,
0,
13,
0,
0,
0,
0,
9,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2381 | [
"Pfeiffer",
"et",
"al.,",
"2020),",
"in",
"which",
"adapter",
"layers",
"–",
"small",
"bottleneck",
"layers",
"–",
"are",
"inserted",
"and",
"trained",
"between",
"frozen",
"pre-trained",
"transformer",
"layers."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2382 | [
"Adapter",
"tuning",
"has",
"emerged",
"as",
"a",
"novel",
"parameter-efficient",
"transfer",
"learning",
"paradigm",
"(Houlsby",
"et",
"al.,",
"2019;"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2383 | [
"Adapter",
"Tuning:"
] | [
0,
0
] |
2384 | [
"More",
"recently,",
"EFL",
"(Wang",
"et",
"al.,",
"2021)",
"proposed",
"a",
"task",
"transformation",
"paradigm,",
"improving",
"single",
"sentence",
"tasks",
"with",
"less",
"labels",
"using",
"rich",
"sentence-pair",
"datasets."
] | [
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2385 | [
"All",
"of",
"them",
"applied",
"pre-fine",
"tuning",
"on",
"NLI",
"datasets."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2386 | [
"A",
"series",
"of",
"work",
"(SentenceBERT",
"(Reimers",
"and",
"Gurevych,",
"2019),",
"BERT-flow",
"(Li",
"et",
"al.,",
"2020),",
"SimCSE",
"(Gao",
"et",
"al.,",
"2021b))",
"explored",
"intermediate",
"training",
"to",
"improve",
"STS",
"tasks."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0
] |
2387 | [
"Existing",
"works",
"(Phang",
"et",
"al.,",
"2018;",
"Liu",
"et",
"al.,",
"2019)",
"have",
"observed",
"that",
"starting",
"from",
"the",
"fine-tuned",
"MNLI",
"model",
"results",
"in",
"a",
"better",
"performance",
"than",
"directly",
"from",
"the",
"vanilla",
"pre-trained",
"models",
"for",
"RTE,",
"STS,",
"and",
"MRPC",
"tasks."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
13,
0,
13,
0
] |
2388 | [
"Supplementary",
"Training:"
] | [
0,
0
] |
2389 | [
"5",
"Related",
"Work"
] | [
0,
0,
0
] |
2390 | [
"In",
"general,",
"the",
"performance",
"of",
"IDPG",
"in",
"downstream",
"tasks",
"improves",
"gradually",
"when",
"using",
"a",
"larger",
"prompt",
"length",
"(Cf.",
"Appendix",
"A.7)."
] | [
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
0,
0,
0
] |
2391 | [
"4.5.6",
"IDPG",
"Scalability",
"We",
"study",
"our",
"proposed",
"model’s",
"scalability",
"in",
"this",
"section."
] | [
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2392 | [
"As",
"shown",
"in",
"Figure",
"4,",
"prompts",
"bring",
"the",
"similar",
"sentences",
"closer",
"while",
"pushing",
"the",
"dissimilar",
"ones",
"apart."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2393 | [
"Both",
"models",
"are",
"finetuned",
"on",
"STS-B",
"training",
"set."
] | [
0,
0,
0,
0,
0,
13,
0,
0
] |
2394 | [
"We",
"compare",
"a",
"vanilla",
"model",
"without",
"any",
"prompts",
"with",
"M-IDPG-PHM."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
1
] |
2395 | [
"More",
"results",
"can",
"be",
"found",
"in",
"Appendix",
"A.6."
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2396 | [
"We",
"sort",
"all",
"sentence",
"pairs",
"in",
"STS-B",
"dev",
"set",
"in",
"descending",
"order",
"by",
"the",
"cosine"
] | [
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0
] |
2397 | [
"Given",
"two",
"sentences,",
"we",
"encode",
"each",
"of",
"them",
"by",
"one",
"of",
"the",
"comparison",
"models",
"and",
"compute",
"the",
"cosine",
"similarity."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2398 | [
"4.5.5",
"How",
"Prompts",
"Help?"
] | [
0,
0,
0,
0
] |
2399 | [
"An",
"interesting",
"finding",
"is",
"that",
"the",
"impact",
"of",
"supplementary",
"training",
"on",
"S-IDPG",
"is",
"high",
"while",
"it",
"is",
"limited",
"for",
"M-IDPG."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
1
] |