chkla commited on
Commit
f386c05
1 Parent(s): 1d9d8c2

Upload 6 files

Browse files
app.py CHANGED
@@ -6,7 +6,7 @@ import json
6
 
7
  def load_prompts():
8
  prompts = []
9
- prompt_dir = "prompt"
10
 
11
  for filename in os.listdir(prompt_dir):
12
  if filename.endswith(".md"):
@@ -16,6 +16,7 @@ def load_prompts():
16
  "id": prompt.get("id"),
17
  "title": prompt.get("title"),
18
  "author": prompt.get("author"),
 
19
  "date": prompt.get("date"),
20
  "language": prompt.get("language"),
21
  "task": prompt.get("task"),
@@ -28,12 +29,16 @@ def load_prompts():
28
  return prompts
29
 
30
  def render_prompt_details(prompt):
31
- st.markdown(f"# {prompt['title']}")
32
  st.write("---")
33
  st.write(f"**Author:** {prompt['author']}")
 
34
  st.write(f"**Date:** {prompt['date']}")
35
  st.write(f"**Language:** {prompt['language']}")
36
  st.write(f"**Task:** {prompt['task']}")
 
 
 
37
  st.write("---")
38
  st.write(prompt["content"])
39
 
@@ -80,10 +85,13 @@ def render_prompt_card(prompt):
80
 
81
  with col1:
82
  st.write(f"### **Card:** {prompt['title']}")
 
 
 
83
  st.write(f"**Prompt Language:** {prompt['language']}")
84
  st.write(f"**Annotation Task:** {prompt['task']}")
85
- st.write(f"**Version:** {prompt['version']}")
86
  st.write(f"**Keywords:** {prompt['keywords']}")
 
87
 
88
  with col2:
89
  if st.button("View details", key=f"view-details-{prompt['id']}"):
@@ -94,49 +102,53 @@ def render_prompt_card(prompt):
94
 
95
  card_container.container().write("---")
96
 
97
- if "selected_prompt_id" not in st.session_state:
98
- st.session_state.selected_prompt_id = None
 
99
 
100
- st.title("🏷️ Annotation PromptCards Archive")
101
- st.write("Welcome to the Prompt Archive! Click on the 'View details' button on each prompt card to see more information about the annotation prompt.")
102
- st.write("---")
 
 
 
 
103
 
104
- prompts = load_prompts()
105
- language_list = list(set([prompt['language'] for prompt in prompts]))
106
- task_list = list(set([prompt['task'] for prompt in prompts]))
107
 
108
- st.sidebar.header("**Annotation PromptCards Archive**")
109
 
110
- st.sidebar.write("A collection of prompts for annotation tasks in NLP. This is a work in progress. Please contribute your prompts via GitHub [[Upload]](https://github.com/chkla/PromptCards).")
 
 
 
111
 
112
- # add a link to the GitHub repository
113
- st.sidebar.write("---")
114
- st.sidebar.write(f"**Total number of prompts:** {len(prompts)}")
115
- st.sidebar.write("---")
116
 
117
- st.sidebar.header("🧑🏽‍🚀 Explore:")
118
- search = st.sidebar.text_input("Search by title")
119
- language_filter = st.sidebar.selectbox("Filter by Language", ["All"] + language_list)
120
- task_filter = st.sidebar.selectbox("Filter by Task", ["All"] + task_list)
121
 
122
- if st.sidebar.button("Back to list"):
123
- st.session_state.selected_prompt_id = None
124
- st.experimental_rerun()
 
 
 
 
125
 
126
- if st.session_state.selected_prompt_id is None:
127
- filtered_prompts = [
128
- prompt for prompt in prompts
129
- if (search.lower() in prompt['title'].lower() or not search)
130
- and (prompt['language'] == language_filter or language_filter == "All")
131
- and (prompt['task'] == task_filter or task_filter == "All")
132
- ]
133
 
134
- for prompt in filtered_prompts:
135
- render_prompt_card(prompt)
136
- else:
137
- prompt = next((p for p in prompts if p["id"] == st.session_state.selected_prompt_id), None)
138
- if prompt:
139
- render_prompt_details(prompt)
140
 
141
- st.sidebar.write("---")
142
- st.sidebar.write("Made with ❤️ and 🤖 by [Chkla](https://chkla.github.io/).")
 
6
 
7
  def load_prompts():
8
  prompts = []
9
+ prompt_dir = "prompts"
10
 
11
  for filename in os.listdir(prompt_dir):
12
  if filename.endswith(".md"):
 
16
  "id": prompt.get("id"),
17
  "title": prompt.get("title"),
18
  "author": prompt.get("author"),
19
+ "paper": prompt.get("paperlink"),
20
  "date": prompt.get("date"),
21
  "language": prompt.get("language"),
22
  "task": prompt.get("task"),
 
29
  return prompts
30
 
31
  def render_prompt_details(prompt):
32
+ st.markdown(f"# Card: {prompt['title']}")
33
  st.write("---")
34
  st.write(f"**Author:** {prompt['author']}")
35
+ st.write(f"**Paper:** {prompt['paper']}")
36
  st.write(f"**Date:** {prompt['date']}")
37
  st.write(f"**Language:** {prompt['language']}")
38
  st.write(f"**Task:** {prompt['task']}")
39
+ st.write(f"**Keywords:** {prompt['keywords']}")
40
+ st.write(f"**Version:** {prompt['version']}")
41
+ st.write(f"_Added By: {prompt['addedby']}_")
42
  st.write("---")
43
  st.write(prompt["content"])
44
 
 
85
 
86
  with col1:
87
  st.write(f"### **Card:** {prompt['title']}")
88
+ st.write(f"**Author:** {prompt['author']}")
89
+ st.write(f"**Paper:** {prompt['paper']}")
90
+ st.write(f"**Date:** {prompt['date']}")
91
  st.write(f"**Prompt Language:** {prompt['language']}")
92
  st.write(f"**Annotation Task:** {prompt['task']}")
 
93
  st.write(f"**Keywords:** {prompt['keywords']}")
94
+ st.write(f"_Added by: {prompt['addedby']}_")
95
 
96
  with col2:
97
  if st.button("View details", key=f"view-details-{prompt['id']}"):
 
102
 
103
  card_container.container().write("---")
104
 
105
+ def main():
106
+ if "selected_prompt_id" not in st.session_state:
107
+ st.session_state.selected_prompt_id = None
108
 
109
+ st.title("Annotation PromptCards 🏷️ 📝 🤖")
110
+ st.write("Welcome to the Prompt Archive! Click on the 'View details' button on each prompt card to see more information about the annotation prompt.")
111
+ st.write("---")
112
+
113
+ prompts = load_prompts()
114
+ language_list = list(set([prompt['language'] for prompt in prompts]))
115
+ task_list = list(set([prompt['task'] for prompt in prompts]))
116
 
117
+ st.sidebar.header("**Annotation PromptCards**")
 
 
118
 
119
+ st.sidebar.write("A collection of prompts for annotation tasks in NLP. This is a work in progress. Please contribute your prompts via GitHub [[Upload]](https://github.com/chkla/PromptCards) 🙏🏽.")
120
 
121
+ # add a link to the GitHub repository
122
+ st.sidebar.write("---")
123
+ st.sidebar.write(f"**Total number of prompts:** {len(prompts)}")
124
+ st.sidebar.write("---")
125
 
126
+ st.sidebar.header("🧑🏽‍🚀 Explore:")
127
+ search = st.sidebar.text_input("Search by title")
128
+ language_filter = st.sidebar.selectbox("Filter by Language", ["All"] + language_list)
129
+ task_filter = st.sidebar.selectbox("Filter by Task", ["All"] + task_list)
130
 
131
+ if st.sidebar.button("Back to list"):
132
+ st.session_state.selected_prompt_id = None
133
+ st.experimental_rerun()
 
134
 
135
+ if st.session_state.selected_prompt_id is None:
136
+ filtered_prompts = [
137
+ prompt for prompt in prompts
138
+ if (search.lower() in prompt['title'].lower() or not search)
139
+ and (prompt['language'] == language_filter or language_filter == "All")
140
+ and (prompt['task'] == task_filter or task_filter == "All")
141
+ ]
142
 
143
+ for prompt in filtered_prompts:
144
+ render_prompt_card(prompt)
145
+ else:
146
+ prompt = next((p for p in prompts if p["id"] == st.session_state.selected_prompt_id), None)
147
+ if prompt:
148
+ render_prompt_details(prompt)
 
149
 
150
+ st.sidebar.write("---")
151
+ st.sidebar.write("Made with ❤️ and 🤖 by [chkla](klamm.ai).")
 
 
 
 
152
 
153
+ if __name__ == "__main__":
154
+ main()
prompts/gilardi-frame-problem-solution-2023.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ id: problem-solution-en-gilardi2023
3
+ title: Problem/ Solution Frames
4
+ author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
5
+ paperlink: https://arxiv.org/abs/2303.15056
6
+ date: 27.3.2023
7
+ language: en
8
+ task: frames
9
+ version: 1.0
10
+ addedby: chkla
11
+ keywords: frames, content moderation
12
+ ---
13
+
14
+ ## Prompt Description
15
+
16
+ [Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
17
+
18
+ ## Prompt Text
19
+
20
+ Content moderation can be seen from two different perspectives:<br>
21
+ • Content moderation can be seen as a PROBLEM; for example, as a restriction of free speech<br>
22
+ • Content moderation can be seen as a SOLUTION; for example, as a protection from harmful speech<br>
23
+ <br>
24
+ For each tweet in the sample, follow these instructions:<br>
25
+ 1. Carefully read the text of the tweet, paying close attention to details.<br>
26
+ 2. Classify the tweet as describing content moderation as a problem, as a solution, or neither.<br>
27
+ <br>
28
+ Tweets should be classified as describing content moderation as a PROBLEM if they emphasize negative effects of content moderation, such as restrictions to free speech, or the biases that can emerge from decisions regarding what users are allowed to post.<br>
29
+ <br>
30
+ Tweets should be classified as describing content moderation as a SOLUTION if they emphasize positive effects of content moderation, such as protecting users from various kinds of harmful content, including hate speech, misinformation, illegal adult content, or spam.<br>
31
+ <br>
32
+ Tweets should be classified as describing content moderation as NEUTRAL if they do not emphasize possible negative or positive effects of content moderation, for example if they simply report on the content moderation activity of social media platforms without linking them to potential advantages or disadvantages for users or stakeholders.
33
+
34
+ ## Language
35
+
36
+ - Prompt Language: [Specify the language of the prompt, e.g., English]
37
+ - Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
38
+
39
+ ## NLP Task
40
+
41
+ - Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
42
+ - Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
43
+
44
+ ## Example Input and Output
45
+
46
+ - Example 1
47
+ - Input: [Provide an example input for the prompt]
48
+ - Output: [Provide an example output for the given input]
49
+ - Example 2
50
+ - Input: [Provide another example input for the prompt]
51
+ - Output: [Provide another example output for the given input]
52
+
53
+ ## Parameters and Constraints
54
+
55
+ - Parameter 1: [Specify any parameters, such as temperature or token count]
56
+ - Parameter 2: [Specify additional parameters or constraints if applicable]
57
+
58
+ ## Evaluation Metrics
59
+
60
+ [List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
61
+
62
+ ## Use Cases
63
+
64
+ [List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
65
+
66
+ ## Limitations and Potential Biases
67
+
68
+ [Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
69
+
70
+ ## Related Research and References
71
+
72
+ [List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
73
+
74
+ ## Cite
75
+
76
+ >> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
77
+
prompts/gilardi-moderation-relevance-2023.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ id: moderation-relevance-en-gilardi2023
3
+ title: Content Moderation Relevance
4
+ author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
5
+ paperlink: https://arxiv.org/abs/2303.15056
6
+ date: 27.3.2023
7
+ language: en
8
+ task: relevance
9
+ version: 1.0
10
+ addedby: chkla
11
+ keywords: content moderation, relevance
12
+ ---
13
+
14
+ ## Prompt Description
15
+
16
+ [Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
17
+
18
+ ## Prompt Text
19
+ For each tweet in the sample, follow these instructions:<br>
20
+ 1. Carefully read the text of the tweet, paying close attention to details.<br>
21
+ 2. Carefully read the text of the tweet, paying close attention to details.<br>
22
+ <br>
23
+ Tweets should be coded as RELEVANT when they directly relate to content moderation, as defined above. This includes tweets that discuss: social media platforms’ content moderation rules and practices, governments’ regulation of online content moderation, and/or mild forms of content moderation like flagging.<br>
24
+ <br>
25
+ Tweets should be coded as IRRELEVANT if they do not refer to content moderation, as defined above, or if they are themselves examples of moderated content. This would include, for example, a Tweet by Donald Trump that Twitter has labeled as “disputed”, a tweet claiming that something is false, or a tweet containing sensitive content. Such tweets might be subject to content moderation, but are not discussing content moderation. Therefore, they should be coded as irrelevant for our purposes.
26
+
27
+ ## Language
28
+
29
+ - Prompt Language: [Specify the language of the prompt, e.g., English]
30
+ - Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
31
+
32
+ ## NLP Task
33
+
34
+ - Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
35
+ - Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
36
+
37
+ ## Example Input and Output
38
+
39
+ - Example 1
40
+ - Input: [Provide an example input for the prompt]
41
+ - Output: [Provide an example output for the given input]
42
+ - Example 2
43
+ - Input: [Provide another example input for the prompt]
44
+ - Output: [Provide another example output for the given input]
45
+
46
+ ## Parameters and Constraints
47
+
48
+ - Parameter 1: [Specify any parameters, such as temperature or token count]
49
+ - Parameter 2: [Specify additional parameters or constraints if applicable]
50
+
51
+ ## Evaluation Metrics
52
+
53
+ [List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
54
+
55
+ ## Use Cases
56
+
57
+ [List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
58
+
59
+ ## Limitations and Potential Biases
60
+
61
+ [Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
62
+
63
+ ## Related Research and References
64
+
65
+ [List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
66
+
67
+ ## Cite
68
+
69
+ >> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
prompts/gilardi-policy-frames-2023.md ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ id: policy-frames-en-gilardi2023
3
+ title: Policy Frames Classification
4
+ author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
5
+ paperlink: https://arxiv.org/abs/2303.15056
6
+ date: 27.3.2023
7
+ language: en
8
+ task: frames
9
+ version: 1.0
10
+ addedby: chkla
11
+ keywords: policy frames, content moderation
12
+ ---
13
+
14
+ ## Prompt Description
15
+
16
+ [Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
17
+
18
+ ## Prompt Text
19
+
20
+ Content moderation, as described above, can be linked to various other topics, such as health, crime, or equality.
21
+ For each tweet in the sample, follow these instructions:<br>
22
+ 1. Carefully read the text of the tweet, paying close attention to details. 2. Classify the tweet into one of the topics defined below.<br>
23
+ <br>
24
+ The topics are defined as follows:<br>
25
+ * ECONOMY: The costs, benefits, or monetary/financial implications of the issue (to an individual, family, community, or to the economy as a whole).<br>
26
+ * Capacity and resources: The lack of or availability of physical, geographical, spatial, human, and financial resources, or the capacity of existing systems and resources to implement or carry out policy goals.<br>
27
+ * MORALITY: Any perspective—or policy objective or action (including proposed action)that is compelled by religious doctrine or interpretation, duty, honor, righ- teousness or any other sense of ethics or social responsibility.<br>
28
+ * FAIRNESS AND EQUALITY: Equality or inequality with which laws, punishment, rewards, and resources are applied or distributed among individuals or groups. Also the balance between the rights or interests of one individual or group compared to another individual or group.<br>
29
+ * CONSTITUTIONALITY AND JURISPRUDENCE: The constraints imposed on or freedoms granted to individuals, government, and corporations via the Constitution, Bill of Rights and other amendments, or judicial interpretation. This deals specifically with the authority of government to regulate, and the authority of in- dividuals/corporations to act independently of government.<br>
30
+ * POLICY PRESCRIPTION AND EVALUATION: Particular policies proposed for addressing an identified problem, and figuring out if certain policies will work, or if existing policies are effective.<br>
31
+ * LAW AND ORDER, CRIME AND JUSTICE: Specific policies in practice and their enforcement, incentives, and implications. Includes stories about enforcement and interpretation of laws by individuals and law enforcement, breaking laws, loopholes, fines, sentencing and punishment. Increases or reductions in crime.<br>
32
+ * SECURITY AND DEFENSE: Security, threats to security, and protection of one’s person, family, in-group, nation, etc. Generally an action or a call to action that can be taken to protect the welfare of a person, group, nation sometimes from a not yet manifested threat.<br>
33
+ * HEALTH AND SAFETY: Health care access and effectiveness, illness, disease, sanitation, obesity, mental health effects, prevention of or perpetuation of gun violence, infrastructure and building safety.<br>
34
+ * QUALITY OF LIFE: The effects of a policy on individuals’ wealth, mobility, access to resources, happiness, social structures, ease of day-to-day routines, quality of community life, etc.<br>
35
+ * CULTURAL IDENTITY: The social norms, trends, values and customs constituting culture(s), as they relate to a specific policy issue.<br>
36
+ * PUBLIC OPINION: References to general social attitudes, polling and demographic information, as well as implied or actual consequences of diverging from or “getting ahead of” public opinion or polls.<br>
37
+ * POLITICAL: Any political considerations surrounding an issue. Issue actions or efforts or stances that are political, such as partisan filibusters, lobbyist involvement, bipartisan efforts, deal-making and vote trading, appealing to one’s base, mentions of political maneuvering. Explicit statements that a policy issue is good or bad for a particular political party.<br>
38
+ * EXTERNAL REGULATION AND REPUTATION: The United States’ external relations with another nation; the external relations of one state with another; or relations between groups. This includes trade agreements and outcomes, comparisons of policy outcomes or desired policy outcomes.<br>
39
+ * OTHER: Any topic that does not fit into the above categories.<br>
40
+
41
+ ## Language
42
+
43
+ - Prompt Language: [Specify the language of the prompt, e.g., English]
44
+ - Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
45
+
46
+ ## NLP Task
47
+
48
+ - Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
49
+ - Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
50
+
51
+ ## Example Input and Output
52
+
53
+ - Example 1
54
+ - Input: [Provide an example input for the prompt]
55
+ - Output: [Provide an example output for the given input]
56
+ - Example 2
57
+ - Input: [Provide another example input for the prompt]
58
+ - Output: [Provide another example output for the given input]
59
+
60
+ ## Parameters and Constraints
61
+
62
+ - Parameter 1: [Specify any parameters, such as temperature or token count]
63
+ - Parameter 2: [Specify additional parameters or constraints if applicable]
64
+
65
+ ## Evaluation Metrics
66
+
67
+ [List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
68
+
69
+ ## Use Cases
70
+
71
+ [List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
72
+
73
+ ## Limitations and Potential Biases
74
+
75
+ [Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
76
+
77
+ ## Related Research and References
78
+
79
+ [List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
80
+
81
+ ## Cite
82
+
83
+ >> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
prompts/gilardi-stance-detection-2023.md ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ id: stance-detectionen-gilardi2023
3
+ title: Stance Detection
4
+ author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
5
+ paperlink: https://arxiv.org/abs/2303.15056
6
+ date: 27.3.2023
7
+ language: en
8
+ task: stance
9
+ version: 1.0
10
+ addedby: chkla
11
+ keywords: stance, content moderation
12
+ ---
13
+
14
+ ## Prompt Description
15
+
16
+ [Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
17
+
18
+ ## Prompt Text
19
+
20
+ In the context of content moderation, Section 230 is a law in the United States that protects websites and other online platforms from being held legally responsible for the content posted by their users. This means that if someone posts something illegal or harmful on a website, the website itself cannot be sued for allowing it to be posted. However, websites can still choose to moderate content and remove anything that violates their own policies.<br>
21
+ <br>
22
+ For each tweet in the sample, follow these instructions:<br>
23
+ 1. Carefully read the text of the tweet, paying close attention to details.<br>
24
+ 2. Classify the tweet as having a positive stance towards Section 230, a negative stance, or a neutral stance.<br>
25
+
26
+ ## Language
27
+
28
+ - Prompt Language: [Specify the language of the prompt, e.g., English]
29
+ - Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
30
+
31
+ ## NLP Task
32
+
33
+ - Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
34
+ - Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
35
+
36
+ ## Example Input and Output
37
+
38
+ - Example 1
39
+ - Input: [Provide an example input for the prompt]
40
+ - Output: [Provide an example output for the given input]
41
+ - Example 2
42
+ - Input: [Provide another example input for the prompt]
43
+ - Output: [Provide another example output for the given input]
44
+
45
+ ## Parameters and Constraints
46
+
47
+ - Parameter 1: [Specify any parameters, such as temperature or token count]
48
+ - Parameter 2: [Specify additional parameters or constraints if applicable]
49
+
50
+ ## Evaluation Metrics
51
+
52
+ [List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
53
+
54
+ ## Use Cases
55
+
56
+ [List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
57
+
58
+ ## Limitations and Potential Biases
59
+
60
+ [Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
61
+
62
+ ## Related Research and References
63
+
64
+ [List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
65
+
66
+ ## Cite
67
+
68
+ >> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
prompts/gilardi-topic-detection-2023.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ id: topic-detection-gilardi2023
3
+ title: Topic Detection
4
+ author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
5
+ paperlink: https://arxiv.org/abs/2303.15056
6
+ date: 27.3.2023
7
+ language: en
8
+ task: topic detection
9
+ version: 1.0
10
+ addedby: chkla
11
+ keywords: content moderation, topic detection
12
+ ---
13
+
14
+ ## Prompt Description
15
+
16
+ [Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
17
+
18
+ ## Prompt Text
19
+
20
+ Tweets about content moderation may also discuss other related topics, such as:<br>
21
+ 1. Section 230, which is a law in the United States that protects websites and other online platforms from being held legally responsible for the content posted by their users (SECTION 230).<br>
22
+ 2. The decision by many social media platforms, such as Twitter and Facebook, to suspend Donald Trump’s account (TRUMP BAN).
23
+ 3. Requests directed to Twitter’s support account or help center (TWITTER SUPPORT).<br>
24
+ 4. Social media platforms’ policies and practices, such as community guidelines or terms of service (PLATFORM POLICIES).
25
+ 5. Complaints about platform’s policy and practices in deplatforming and content moderation or suggestions to suspend particular accounts, or complaints about accounts being suspended or reported (COMPLAINTS).<br>
26
+ 6. If a text is not about the SECTION 230, COMPLAINTS, TRUMP BAN, TWIT- TER SUPPORT, and PLATFORM POLICIES, then it should be classified in OTHER class (OTHER).<br>
27
+ <br>
28
+ For each tweet in the sample, follow these instructions:<br>
29
+ 1. Carefully read the text of the tweet, paying close attention to details.<br>
30
+ 2. Please classify the following text according to topic (defined by function of the text, author’s purpose and form of the text). You can choose from the following classes: SECTION 230, TRUMP BAN, COMPLAINTS, TWITTER SUPPORT, PLATFORM POLICIES, and OTHER<br>
31
+
32
+ ## Language
33
+
34
+ - Prompt Language: [Specify the language of the prompt, e.g., English]
35
+ - Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
36
+
37
+ ## NLP Task
38
+
39
+ - Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
40
+ - Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
41
+
42
+ ## Example Input and Output
43
+
44
+ - Example 1
45
+ - Input: [Provide an example input for the prompt]
46
+ - Output: [Provide an example output for the given input]
47
+ - Example 2
48
+ - Input: [Provide another example input for the prompt]
49
+ - Output: [Provide another example output for the given input]
50
+
51
+ ## Parameters and Constraints
52
+
53
+ - Parameter 1: [Specify any parameters, such as temperature or token count]
54
+ - Parameter 2: [Specify additional parameters or constraints if applicable]
55
+
56
+ ## Evaluation Metrics
57
+
58
+ [List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
59
+
60
+ ## Use Cases
61
+
62
+ [List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
63
+
64
+ ## Limitations and Potential Biases
65
+
66
+ [Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
67
+
68
+ ## Related Research and References
69
+
70
+ [List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
71
+
72
+ ## Cite
73
+
74
+ >> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)