id
stringlengths 1
4
| tokens
sequence | ner_tags
sequence |
---|---|---|
2900 | [
"3.2",
"Model",
"Overview"
] | [
0,
0,
0
] |
2901 | [
"We",
"also",
"experimented",
"with",
"a",
"sentiment",
"classification",
"dataset",
"which",
"provides",
"sentiment",
"labels",
"consisting",
"of",
"positive,",
"negative",
"and",
"neutral."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2902 | [
"Emotions",
"are",
"labeled",
"as",
"one",
"of",
"the",
"predefined",
"classes",
"depending",
"on",
"the",
"dataset,",
"and",
"the",
"emotions",
"we",
"experimented",
"with",
"are",
"either",
"6",
"or",
"7."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2903 | [
"The",
"ERC",
"is",
"a",
"task",
"of",
"predicting",
"the",
"emotion",
"et",
"of",
"ut,",
"the",
"utterance",
"of",
"the",
"t-th",
"turn,",
"given",
"the",
"previous",
"utterances",
"ht",
"=",
"."
] | [
0,
11,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2904 | [
"While",
"pui",
"and",
"puj",
"(i",
"=",
"j)",
"can",
"be",
"the",
"same",
"speaker,",
"the",
"minimum",
"number",
"of",
"the",
"unique",
"conversation",
"participants",
"should",
"be",
"2",
"or",
"more."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2905 | [
"ui",
"is",
"the",
"utterance",
"which",
"the",
"speaker",
"pui",
"uttered,",
"where",
"pui",
"is",
"one",
"of",
"the",
"conversation",
"participants."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2906 | [
"In",
"a",
"conversation,",
"M",
"sequential",
"utterances",
"are",
"given",
"as",
"[(u1,",
"pu1),",
"(u2,",
"pu2),",
"...,",
"(uM",
",",
"puM",
")]."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2907 | [
"3.1",
"Problem",
"Statement"
] | [
0,
0,
0
] |
2908 | [
"3",
"Approach"
] | [
0,
0
] |
2909 | [
"EDA",
"(Bothe",
"et",
"al.,",
"2020)",
"expands",
"the",
"multi-modal",
"emotion",
"datasets",
"by",
"extracting",
"dialog",
"acts",
"from",
"MELD",
"and",
"IEMOCAP",
"and",
"finds",
"out",
"that",
"there",
"is",
"a",
"correlation",
"between",
"dialogue",
"acts",
"and",
"emotion",
"labels."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2910 | [
"Liu",
"et",
"al.",
"(2020)",
"reports",
"that",
"the",
"lack",
"of",
"external",
"knowledge",
"makes",
"it",
"difficult",
"to",
"classify",
"implicit",
"emotions",
"from",
"the",
"conversation",
"history."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2911 | [
"Zhou",
"et",
"al.",
"(2018);",
"Zhang",
"et",
"al.",
"(2018a)",
"shows",
"that",
"commonsense",
"knowledge",
"is",
"important",
"for",
"understanding",
"conversations",
"and",
"generating",
"appropriate",
"responses."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2912 | [
"dyadic-party",
"dialogues,",
"therefore",
"have",
"more",
"conditions",
"to",
"consider",
"and",
"result",
"in",
"poor",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2913 | [
"In",
"general,",
"the",
"multi-party",
"conversations",
"have",
"higher",
"speaker",
"dependency",
"than",
"the"
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2914 | [
"However,",
"as",
"the",
"multi-party",
"conversation",
"datasets",
"including",
"MELD",
"and",
"EmoryNLP",
"have",
"become",
"available,",
"a",
"lot",
"of",
"recent",
"research",
"is",
"being",
"conducted",
"on",
"multi-party",
"dialogues",
"such",
"as",
"Zhang",
"et",
"al."
] | [
0,
0,
0,
0,
0,
0,
0,
13,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2915 | [
";",
"Zadeh",
"et",
"al.",
"(2017);",
"Majumder",
"et",
"al.",
"(2019),",
"most",
"works",
"focused",
"on",
"dyadic-party",
"conversation."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2916 | [
"In",
"this",
"work,",
"we",
"design",
"and",
"introduce",
"a",
"text-based",
"emotion",
"recognition",
"system",
"using",
"neural",
"networks."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
0,
0,
0,
0
] |
2917 | [
"Multimodal",
"information",
"is",
"not",
"always",
"given",
"in",
"most",
"social",
"media,",
"especially",
"in",
"chatbot",
"systems",
"where",
"they",
"are",
"mainly",
"composed",
"of",
"text-based",
"systems."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2918 | [
"MELD",
"and",
"ICON",
"(Hazarika",
"et",
"al.,",
"2018a)",
"show",
"that",
"the",
"more",
"multi-modal",
"information",
"is",
"used,",
"the",
"better",
"the",
"performance",
"and",
"the",
"text",
"information",
"plays",
"the",
"most",
"important",
"role."
] | [
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2919 | [
"Datcu",
"and",
"Rothkrantz",
"(2014)",
"uses",
"speech",
"and",
"visual",
"information",
"to",
"recognize",
"emotions,",
"and",
"(Alm",
"et",
"al.,",
"2005)",
"attempts",
"to",
"recognize",
"emotions",
"based",
"on",
"text",
"information."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2920 | [
"The",
"multi-modal",
"data",
"such",
"as",
"MELD",
"and",
"IEMOCAP",
"are",
"some",
"of",
"the",
"available",
"standard",
"datasets",
"for",
"emotion",
"recognition",
"and",
"they",
"are",
"composed",
"of",
"text,",
"speech",
"and",
"vision-based",
"data."
] | [
0,
0,
0,
0,
0,
13,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2921 | [
"In",
"addition,",
"Ekman",
"explains",
"that",
"a",
"multimodal",
"view",
"is",
"important",
"for",
"multiple",
"emotions",
"recognition."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2922 | [
"Ekman",
"(Ekman,",
"1992)",
"constructs",
"taxonomy",
"of",
"six",
"common",
"emotions",
"(Joy,",
"Sadness,",
"Fear,",
"Anger,",
"Surprise,",
"and",
"Disgust)",
"from",
"human",
"facial",
"expressions."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2923 | [
"ToDKAT",
"(Zhu",
"et",
"al.,",
"2021)",
"improves",
"performance",
"by",
"combining",
"commonsense",
"knowledge",
"using",
"COMET",
"and",
"topic",
"discovery",
"using",
"VHRED",
"(Serban",
"et",
"al.,",
"2017)",
"to",
"the",
"model."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2924 | [
"ATOMIC",
"has",
"9",
"sentence",
"relation",
"types",
"with",
"inferential",
"if-then",
"commonsense",
"knowledge",
"expressed",
"in",
"text."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2925 | [
"Commonsense",
"knowledge",
"feature",
"is",
"extracted",
"and",
"leveraged",
"with",
"COMET",
"(Bosselut",
"et",
"al.,",
"2019)",
"trained",
"with",
"ATOMIC",
"(The",
"Atlas",
"of",
"Machine",
"Commonsense)",
"(Sap",
"et",
"al.,",
"2019)."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2926 | [
"COSMIC",
"(Ghosal",
"et",
"al.,",
"2020)",
"and",
"Psychological",
"(Li",
"et",
"al.,",
"2021)",
"improve",
"the",
"performance",
"of",
"emotion",
"recognition",
"by",
"extracting",
"commonsense",
"knowledge",
"of",
"the",
"previous",
"utterances."
] | [
1,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
0,
0,
0,
0,
0,
0,
0,
0
] |
2927 | [
"NRC_VAD",
"Lexicon",
"has",
"human",
"ratings",
"of",
"valence,",
"arousal,",
"and",
"dominance",
"for",
"more",
"than",
"20,000",
"English",
"words."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2928 | [
"ConceptNet",
"is",
"a",
"knowledge",
"graph",
"that",
"connects",
"words",
"and",
"phrases",
"in",
"natural",
"language",
"using",
"labeled",
"edges."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2929 | [
"KET",
"(Zhong",
"et",
"al.,",
"2019)",
"is",
"used",
"as",
"external",
"knowledge",
"based",
"on",
"ConceptNet",
"(Speer",
"et",
"al.,",
"2017)",
"and",
"emotion",
"lex"
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2930 | [
"Many",
"recent",
"studies",
"use",
"external",
"knowledge",
"to",
"improve",
"the",
"ERC",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
0
] |
2931 | [
"2",
"Related",
"Work"
] | [
0,
0,
0
] |
2932 | [
"Further",
"experiments",
"also",
"show",
"that",
"our",
"approach",
"can",
"be",
"used",
"in",
"other",
"languages",
"and",
"show",
"the",
"performance",
"of",
"CoMPM",
"when",
"the",
"number",
"of",
"data",
"is",
"limited."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0
] |
2933 | [
"We",
"perform",
"an",
"ablation",
"study",
"on",
"each",
"module",
"to",
"show",
"that",
"the",
"proposed",
"approach",
"is",
"effective."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2934 | [
"CoMPM",
"achieves",
"the",
"first",
"or",
"second",
"performance",
"according",
"to",
"the",
"evaluation",
"metric",
"compared",
"to",
"all",
"previous",
"systems."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2935 | [
"Multi-party",
"datasets",
"are",
"MELD",
"(Poria",
"et",
"al.,",
"2019)",
"and",
"EmoryNLP",
"(Zahiri",
"and",
"Choi,",
"2018),",
"and",
"dyadic",
"datasets",
"are",
"IEMOCAP",
"(Busso",
"et",
"al.,",
"2008)",
"and",
"DailyDialog",
"(Li",
"et",
"al.,",
"2017)."
] | [
0,
0,
0,
13,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0
] |
2936 | [
"We",
"experiment",
"on",
"4",
"different",
"English",
"ERC",
"datasets."
] | [
0,
0,
0,
0,
0,
0,
11,
0
] |
2937 | [
"We",
"use",
"the",
"PM",
"to",
"help",
"predict",
"the",
"emotion",
"of",
"the",
"speaker",
"by",
"taking",
"into",
"account",
"the",
"speaker’s",
"linguistic",
"preferences",
"and",
"characteristics."
] | [
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2938 | [
"We",
"use",
"the",
"output",
"of",
"the",
"pre-trained",
"language",
"model",
"as",
"the",
"memory",
"embedding",
"where",
"the",
"utterances",
"are",
"passed",
"into",
"the",
"language",
"model."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2939 | [
"The",
"second",
"is",
"a",
"pre-trained",
"memory",
"module",
"(PM)",
"that",
"extracts",
"memory",
"from",
"utterances."
] | [
0,
0,
0,
0,
1,
2,
2,
1,
0,
0,
0,
0,
0
] |
2940 | [
"(1)",
"The",
"first",
"is",
"a",
"context",
"embedding",
"module",
"(CoM)",
"that",
"reflects",
"all",
"previous",
"utterances",
"as",
"context."
] | [
0,
0,
0,
0,
0,
1,
2,
2,
1,
0,
0,
0,
0,
0,
0,
0
] |
2941 | [
"CoM",
"is",
"an",
"auto-regressive",
"model",
"that",
"predicts",
"the",
"current",
"emotion",
"through",
"attention",
"between",
"the",
"previous",
"utterances",
"of",
"the",
"conversation",
"and",
"the",
"current",
"utterance."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2942 | [
"CoMPM,",
"introduced",
"in",
"this",
"paper,",
"is",
"composed",
"of",
"two",
"modules",
"that",
"take",
"into",
"account",
"previous",
"utterances",
"in",
"dialogue."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2943 | [
"Based",
"on",
"these",
"studies,",
"we",
"eliminate",
"the",
"dependence",
"on",
"structured",
"external",
"data",
"used",
"in",
"cutting-edge",
"systems",
"and",
"use",
"a",
"pre-trained",
"language",
"model",
"as",
"a",
"feature",
"extractor",
"of",
"knowledge."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2944 | [
"Petroni",
"et",
"al.",
"(2019)",
"introduces",
"that",
"these",
"language",
"models",
"can",
"be",
"used",
"as",
"knowledge",
"bases",
"and",
"have",
"many",
"advantages",
"over",
"the",
"structured",
"knowledge",
"bases."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2945 | [
"Since",
"pre-trained",
"language",
"models",
"are",
"trained",
"by",
"unsupervised",
"learning,",
"these",
"models",
"are",
"relatively",
"usable",
"approaches",
"regardless",
"of",
"language",
"types."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2946 | [
"In",
"recent",
"NLP",
"studies,",
"due",
"to",
"the",
"effectiveness",
"of",
"the",
"pre-trained",
"language",
"model,",
"it",
"has",
"already",
"been",
"developed",
"in",
"many",
"countries."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2947 | [
"In",
"order",
"to",
"utilize",
"the",
"previous",
"methods",
"in",
"languages",
"of",
"other",
"countries,",
"it",
"is",
"expensive",
"and",
"difficult",
"to",
"utilize",
"because",
"external",
"knowledge",
"data",
"must",
"be",
"newly",
"constructed."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2948 | [
"Many",
"recent",
"studies",
"use",
"external",
"knowledge",
"to",
"improve",
"the",
"ERC",
"performance."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
0
] |
2949 | [
"AGHMN",
"(Jiao",
"et",
"al.,",
"2020)",
"considers",
"the",
"previous",
"utterances",
"through",
"memory",
"summarizing",
"using",
"GRU",
"with",
"attention."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2950 | [
"Representatively,",
"DialogueRNN",
"(Majumder",
"et",
"al.,",
"2019)",
"recognizes",
"the",
"present",
"emotion",
"by",
"tracking",
"context",
"from",
"the",
"previous",
"utterances",
"and",
"the",
"speaker’s",
"emotion."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2951 | [
"Therefore,",
"recent",
"studies",
"are",
"attempting",
"to",
"recognize",
"emotions",
"while",
"taking",
"into",
"account",
"the",
"previous",
"utterances."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2952 | [
"Like",
"the",
"previous",
"studies",
"(Ghosal",
"et",
"al.,",
"2020),",
"we",
"show",
"that",
"the",
"utterance-level",
"emotion",
"recognition,",
"which",
"does",
"not",
"consider",
"the",
"previous",
"utterance,",
"have",
"limitations",
"and",
"experiments",
"result",
"in",
"poor",
"performances."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2953 | [
"If",
"the",
"system",
"does",
"not",
"take",
"into",
"account",
"previous",
"utterances,",
"it",
"is",
"difficult",
"to",
"properly",
"recognize",
"emotions."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2954 | [
"The",
"emotion",
"of",
"speaker",
"B’s",
"utterance",
"(\"How’d",
"you",
"get",
"to",
"that?\")",
"is",
"angry."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2955 | [
"Figure",
"1",
"is",
"an",
"example",
"of",
"a",
"conversation",
"in",
"which",
"two",
"speakers",
"are",
"angry",
"at",
"each",
"other."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2956 | [
"As",
"these",
"interactive",
"machines",
"increase,",
"the",
"ERC",
"module",
"plays",
"an",
"increasingly",
"important",
"role."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2957 | [
"In",
"addition,",
"emotion",
"recognition",
"can",
"be",
"effectively",
"used",
"for",
"opinion",
"mining,",
"recommender",
"systems,",
"and",
"healthcare",
"systems",
"where",
"it",
"can",
"improve",
"the",
"service",
"qualities",
"by",
"providing",
"personalized",
"results."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2958 | [
"As",
"the",
"number",
"of",
"applications",
"such",
"as",
"interactive",
"chatbots",
"or",
"social",
"media",
"that",
"are",
"used",
"by",
"many",
"users",
"has",
"recently",
"increased",
"dramatically,",
"Emotion",
"Recognition",
"in",
"Conversation",
"(ERC)",
"plays",
"a",
"more",
"important",
"role",
"in",
"natural",
"language",
"processing,",
"and",
"as",
"a",
"proof,",
"a",
"lot",
"of",
"research",
"(Poria",
"et",
"al.,",
"2019;",
"Zhang",
"et",
"al.,",
"2019;",
"Ghosal",
"et",
"al.,",
"2020;",
"Jiao",
"et",
"al.,",
"2020)",
"has",
"been",
"conducted",
"on",
"the",
"task."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
12,
12,
11,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2959 | [
"Introduction"
] | [
0
] |
2960 | [
"Our",
"code",
"is",
"available",
"on",
"github",
"1."
] | [
0,
0,
0,
0,
0,
0,
0
] |
2961 | [
"In",
"addition,",
"our",
"method",
"shows",
"that",
"it",
"can",
"be",
"extended",
"to",
"other",
"languages",
"because",
"structured",
"knowledge",
"is",
"not",
"required,",
"unlike",
"previous",
"methods."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2962 | [
"CoMPM",
"achieves",
"the",
"first",
"or",
"second",
"performance",
"on",
"all",
"data",
"and",
"is",
"state-of-the-art",
"among",
"systems",
"that",
"do",
"not",
"leverage",
"structured",
"data."
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2963 | [
"We",
"introduce",
"CoMPM,",
"which",
"combines",
"the",
"speaker’s",
"pre-trained",
"memory",
"with",
"the",
"context",
"model,",
"and",
"find",
"that",
"the",
"pre-trained",
"memory",
"significantly",
"improves",
"the",
"performance",
"of",
"the",
"context",
"model."
] | [
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2964 | [
"Therefore,",
"we",
"extract",
"the",
"pre-trained",
"memory",
"using",
"the",
"pre-trained",
"language",
"model",
"as",
"an",
"extractor",
"of",
"external",
"knowledge."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2965 | [
"However,",
"structured",
"data",
"is",
"difficult",
"to",
"access",
"in",
"non-English",
"languages,",
"making",
"it",
"difficult",
"to",
"extend",
"to",
"other",
"languages."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2966 | [
"Many",
"recent",
"approaches",
"show",
"performance",
"improvement",
"by",
"combining",
"knowledge",
"into",
"modules",
"learned",
"from",
"external",
"structured",
"data."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2967 | [
"Since",
"emotion",
"recognition",
"in",
"conversation",
"is",
"inaccurate",
"if",
"the",
"previous",
"utterances",
"are",
"not",
"taken",
"into",
"account,",
"many",
"studies",
"reflect",
"the",
"dialogue",
"context",
"to",
"improve",
"the",
"performances."
] | [
0,
11,
12,
12,
12,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2968 | [
"If",
"the",
"machine-generated",
"sentences",
"reflect",
"emotion,",
"more",
"human-like",
"sympathetic",
"conversations",
"are",
"possible."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2969 | [
"As",
"the",
"use",
"of",
"interactive",
"machines",
"grow,",
"the",
"task",
"of",
"Emotion",
"Recognition",
"in",
"Conversation",
"(ERC)",
"became",
"more",
"important."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
12,
12,
11,
0,
0,
0
] |
2970 | [
"Abstract"
] | [
0
] |
2971 | [
"CoMPM:",
"Context",
"Modeling",
"with",
"Speaker’s",
"Pre-trained",
"Memory",
"Tracking",
"for",
"Emotion",
"Recognition",
"in",
"Conversation"
] | [
1,
0,
0,
0,
0,
0,
0,
0,
0,
11,
12,
12,
12
] |
2972 | [
"Figure",
"4b",
"demonstrates",
"a",
"termination",
"of",
"stock",
"sale"
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2973 | [
"Herein,",
"we",
"visualize",
"the",
"heatmaps",
"in",
"some",
"cases",
"to",
"investigate",
"how",
"to",
"assign",
"attention",
"weights",
"to",
"text."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2974 | [
"A.4",
"Case",
"Study:",
"Visualization",
"of",
"Attention"
] | [
0,
0,
0,
0,
0,
0
] |
2975 | [
"We",
"summarize",
"the",
"hyperparameters",
"in",
"Table",
"5."
] | [
0,
0,
0,
0,
0,
0,
0
] |
2976 | [
"In",
"addition,",
"the",
"coefficient",
"λ",
"of",
"the",
"regularization",
"term",
"was",
"set",
"as",
"0.0001."
] | [
0,
0,
0,
3,
3,
0,
0,
0,
0,
0,
0,
0,
5
] |
2977 | [
"We",
"used",
"the",
"Adam",
"optimizer",
"with",
"learning",
"rates",
"of",
"0.1",
"and",
"0.001",
"in",
"the",
"inner",
"and",
"outer",
"updates,",
"that",
"is,",
"α",
"and",
"β",
"in",
",",
"respectively."
] | [
0,
0,
0,
0,
0,
0,
3,
4,
0,
5,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
3,
0,
3,
0,
0,
0
] |
2978 | [
"After",
"that,",
"the",
"performance",
"of",
"the",
"adapted",
"task",
"model",
"is",
"obtained",
"using",
"queries."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2979 | [
"In",
"validation",
"and",
"test",
"process,",
"we",
"val",
"and",
"sampled",
"30",
"tasks",
"with",
"15",
"queries",
"from",
"test,",
"and",
"only",
"performed",
"task",
"adaptation",
"using",
"S",
"K-shots."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2980 | [
"During",
"the",
"meta-training",
"process,",
"we",
"sampled",
"four",
"tasks",
"with",
"15",
"queries",
"tr,",
"and",
"it",
"leads",
"to",
"performing",
"task",
"adaptation",
"from",
"four",
"times",
"per",
"each",
"meta-optimization",
"update",
"and",
"early",
"stop",
"when",
"the",
"validation",
"loss",
"fails",
"to",
"improve",
"for",
"20",
"steps."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2981 | [
"In",
"our",
"work,",
"we",
"train",
"all",
"experiments",
"on",
"a",
"single",
"NVIDIA",
"A100",
"32G",
"GPU."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2982 | [
"A.3",
"Training",
"Details",
"and",
"Hyperparameter"
] | [
0,
0,
0,
0,
0
] |
2983 | [
"Table",
"3",
"summarizes",
"the",
"above",
"model",
"parameters."
] | [
0,
0,
0,
0,
0,
0,
0
] |
2984 | [
"We",
"set",
"the",
"size",
"to",
"150",
"for",
"the",
"meta-attention-aspects",
"dictionary,",
"and",
"importantly",
"fixed",
"the",
"number",
"of",
"top-k",
"attention",
"aspects",
"to",
"20."
] | [
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
4,
4,
4,
0,
5
] |
2985 | [
"For",
"each",
"task",
"classifier,",
"that",
"is,",
"fθ′τi",
",",
"it",
"is",
"designed",
"as",
"single-layer",
"fully",
"connected",
"neural",
"network."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2986 | [
"The",
"gating",
"network,",
"Wg",
"is",
"linear",
"transformation",
"and",
"its",
"size",
"is",
"the",
"number",
"of",
"meta-attention-aspects."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
4
] |
2987 | [
"As",
"shown",
"in",
"these",
"cases,",
"LEA",
"properly",
"captures",
"important",
"words",
"under",
"a",
"certain",
"topic",
"and",
"assigns",
"attention",
"weights",
"to",
"a",
"given",
"text."
] | [
0,
0,
0,
0,
0,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2988 | [
"Figure",
"4d",
"talks",
"about",
"the",
"authority",
"of",
"platinum",
"and",
"gold",
"coins",
"under",
"the",
"Economics",
"topic,",
"and",
"the",
"words",
"“US,”",
"“Mint,”",
"“authority,”",
"“gold,”",
"“platinum,”",
"and",
"“coin”",
"are",
"hence",
"highlighted."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2989 | [
"Figure",
"4c",
"shows",
"that",
"the",
"Turkish",
"market",
"was",
"closed",
"related",
"to",
"the",
"Market",
"topic,",
"and",
"its",
"relevant",
"words",
"such",
"as",
"“Turkish,”",
"“markets,”",
"and",
"“closed”",
"are",
"highly",
"attended",
"as",
"expected."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2990 | [
"These",
"were",
"extracted",
"under",
"the",
"Corporate",
"and",
"Industrial",
"topic",
"in",
"the",
"RCV-1",
"dataset",
"and",
"some",
"seminal",
"words",
"such",
"as",
"“agreement”,",
"“contractual”",
"and",
"“receivership”",
"are",
"highlighted",
"to",
"appear",
"in",
"the",
"topic."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2991 | [
"The",
"texts",
"in",
"(a)",
"and",
"(b)",
"are",
"different",
"samples",
"under",
"a",
"topic",
"Corporate",
"and",
"Industrial."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |
2992 | [
"The",
"relation",
"network,",
"fθr",
"is",
"composed",
"of",
"two-layers",
"neural",
"network",
"with",
"ReLU",
"activation",
"and",
"input",
"size",
"is",
"two",
"times",
"of",
"encoder",
"outputs",
"and",
"the",
"size",
"of",
"output",
"is",
"the",
"number",
"of",
"meta-attention-aspects,",
"i.e.,",
"150."
] | [
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
3,
4,
4,
0,
0,
3,
4,
4,
0,
5
] |
2993 | [
"In",
"the",
"encoding",
"process",
"of",
"our",
"experiments,",
"the",
"768-dimensional",
"[CLS]",
"vector,",
"which",
"is",
"of",
"the",
"same",
"size",
"of",
"the",
"output",
"of",
"the",
"pre-trained",
"BERTbase-uncased,",
"is",
"linearly",
"transformed",
"through",
"fθe",
"into",
"a",
"300-dimensional",
"vector."
] | [
0,
0,
0,
0,
0,
0,
0,
0,
5,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
0
] |
2994 | [
"We",
"share",
"the",
"breakdown",
"of",
"LEA’s",
"implementation."
] | [
0,
0,
0,
0,
0,
1,
0
] |
2995 | [
"A.2",
"Implementation",
"Details"
] | [
0,
0,
0
] |
2996 | [
"We",
"summarize",
"the",
"above",
"information",
"in",
"Table",
"4."
] | [
0,
0,
0,
0,
0,
0,
0,
0
] |
2997 | [
"For",
"the",
"20",
"Newsgroup",
"dataset,",
"we",
"randomly",
"selected",
"20",
"topic",
"classes,",
"and",
"the",
"metatraining",
"set,",
"meta-validation",
"set,",
"and",
"meta-test",
"set",
"contained",
"10,",
"5,",
"and",
"5",
"disjoint",
"classes,",
"respectively."
] | [
0,
0,
13,
14,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
5,
6,
6,
6,
0,
0,
0
] |
2998 | [
"As",
"a",
"result,",
"Amazon",
"product",
"data",
"is",
"divided",
"into",
"15,",
"and",
"9",
"disjoint",
"classes",
"for",
"meta-training",
"and",
"test",
"sets",
"and",
"meta-validation",
"set",
"is",
"not",
"used",
"in",
"Amazon",
"product",
"data."
] | [
0,
0,
0,
13,
14,
14,
0,
0,
0,
5,
6,
6,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
13,
14,
14
] |
2999 | [
"In",
"Amazon",
"product",
"data,",
"we",
"split",
"the",
"data",
"using",
"rules",
"in",
"(Bao",
"et",
"al.,",
"2019),",
"and",
"its",
"training",
"and"
] | [
0,
13,
14,
14,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0
] |