Spaces:
Sleeping
Sleeping
Upload 6 files
Browse files
app.py
CHANGED
@@ -6,7 +6,7 @@ import json
|
|
6 |
|
7 |
def load_prompts():
|
8 |
prompts = []
|
9 |
-
prompt_dir = "
|
10 |
|
11 |
for filename in os.listdir(prompt_dir):
|
12 |
if filename.endswith(".md"):
|
@@ -16,6 +16,7 @@ def load_prompts():
|
|
16 |
"id": prompt.get("id"),
|
17 |
"title": prompt.get("title"),
|
18 |
"author": prompt.get("author"),
|
|
|
19 |
"date": prompt.get("date"),
|
20 |
"language": prompt.get("language"),
|
21 |
"task": prompt.get("task"),
|
@@ -28,12 +29,16 @@ def load_prompts():
|
|
28 |
return prompts
|
29 |
|
30 |
def render_prompt_details(prompt):
|
31 |
-
st.markdown(f"# {prompt['title']}")
|
32 |
st.write("---")
|
33 |
st.write(f"**Author:** {prompt['author']}")
|
|
|
34 |
st.write(f"**Date:** {prompt['date']}")
|
35 |
st.write(f"**Language:** {prompt['language']}")
|
36 |
st.write(f"**Task:** {prompt['task']}")
|
|
|
|
|
|
|
37 |
st.write("---")
|
38 |
st.write(prompt["content"])
|
39 |
|
@@ -80,10 +85,13 @@ def render_prompt_card(prompt):
|
|
80 |
|
81 |
with col1:
|
82 |
st.write(f"### **Card:** {prompt['title']}")
|
|
|
|
|
|
|
83 |
st.write(f"**Prompt Language:** {prompt['language']}")
|
84 |
st.write(f"**Annotation Task:** {prompt['task']}")
|
85 |
-
st.write(f"**Version:** {prompt['version']}")
|
86 |
st.write(f"**Keywords:** {prompt['keywords']}")
|
|
|
87 |
|
88 |
with col2:
|
89 |
if st.button("View details", key=f"view-details-{prompt['id']}"):
|
@@ -94,49 +102,53 @@ def render_prompt_card(prompt):
|
|
94 |
|
95 |
card_container.container().write("---")
|
96 |
|
97 |
-
|
98 |
-
st.session_state
|
|
|
99 |
|
100 |
-
st.title("
|
101 |
-
st.write("Welcome to the Prompt Archive! Click on the 'View details' button on each prompt card to see more information about the annotation prompt.")
|
102 |
-
st.write("---")
|
|
|
|
|
|
|
|
|
103 |
|
104 |
-
|
105 |
-
language_list = list(set([prompt['language'] for prompt in prompts]))
|
106 |
-
task_list = list(set([prompt['task'] for prompt in prompts]))
|
107 |
|
108 |
-
st.sidebar.
|
109 |
|
110 |
-
|
|
|
|
|
|
|
111 |
|
112 |
-
|
113 |
-
st.sidebar.
|
114 |
-
st.sidebar.
|
115 |
-
st.sidebar.
|
116 |
|
117 |
-
st.sidebar.
|
118 |
-
|
119 |
-
|
120 |
-
task_filter = st.sidebar.selectbox("Filter by Task", ["All"] + task_list)
|
121 |
|
122 |
-
if st.
|
123 |
-
|
124 |
-
|
|
|
|
|
|
|
|
|
125 |
|
126 |
-
|
127 |
-
|
128 |
-
|
129 |
-
|
130 |
-
|
131 |
-
|
132 |
-
]
|
133 |
|
134 |
-
|
135 |
-
|
136 |
-
else:
|
137 |
-
prompt = next((p for p in prompts if p["id"] == st.session_state.selected_prompt_id), None)
|
138 |
-
if prompt:
|
139 |
-
render_prompt_details(prompt)
|
140 |
|
141 |
-
|
142 |
-
|
|
|
6 |
|
7 |
def load_prompts():
|
8 |
prompts = []
|
9 |
+
prompt_dir = "prompts"
|
10 |
|
11 |
for filename in os.listdir(prompt_dir):
|
12 |
if filename.endswith(".md"):
|
|
|
16 |
"id": prompt.get("id"),
|
17 |
"title": prompt.get("title"),
|
18 |
"author": prompt.get("author"),
|
19 |
+
"paper": prompt.get("paperlink"),
|
20 |
"date": prompt.get("date"),
|
21 |
"language": prompt.get("language"),
|
22 |
"task": prompt.get("task"),
|
|
|
29 |
return prompts
|
30 |
|
31 |
def render_prompt_details(prompt):
|
32 |
+
st.markdown(f"# Card: {prompt['title']}")
|
33 |
st.write("---")
|
34 |
st.write(f"**Author:** {prompt['author']}")
|
35 |
+
st.write(f"**Paper:** {prompt['paper']}")
|
36 |
st.write(f"**Date:** {prompt['date']}")
|
37 |
st.write(f"**Language:** {prompt['language']}")
|
38 |
st.write(f"**Task:** {prompt['task']}")
|
39 |
+
st.write(f"**Keywords:** {prompt['keywords']}")
|
40 |
+
st.write(f"**Version:** {prompt['version']}")
|
41 |
+
st.write(f"_Added By: {prompt['addedby']}_")
|
42 |
st.write("---")
|
43 |
st.write(prompt["content"])
|
44 |
|
|
|
85 |
|
86 |
with col1:
|
87 |
st.write(f"### **Card:** {prompt['title']}")
|
88 |
+
st.write(f"**Author:** {prompt['author']}")
|
89 |
+
st.write(f"**Paper:** {prompt['paper']}")
|
90 |
+
st.write(f"**Date:** {prompt['date']}")
|
91 |
st.write(f"**Prompt Language:** {prompt['language']}")
|
92 |
st.write(f"**Annotation Task:** {prompt['task']}")
|
|
|
93 |
st.write(f"**Keywords:** {prompt['keywords']}")
|
94 |
+
st.write(f"_Added by: {prompt['addedby']}_")
|
95 |
|
96 |
with col2:
|
97 |
if st.button("View details", key=f"view-details-{prompt['id']}"):
|
|
|
102 |
|
103 |
card_container.container().write("---")
|
104 |
|
105 |
+
def main():
|
106 |
+
if "selected_prompt_id" not in st.session_state:
|
107 |
+
st.session_state.selected_prompt_id = None
|
108 |
|
109 |
+
st.title("Annotation PromptCards 🏷️ 📝 🤖")
|
110 |
+
st.write("Welcome to the Prompt Archive! Click on the 'View details' button on each prompt card to see more information about the annotation prompt.")
|
111 |
+
st.write("---")
|
112 |
+
|
113 |
+
prompts = load_prompts()
|
114 |
+
language_list = list(set([prompt['language'] for prompt in prompts]))
|
115 |
+
task_list = list(set([prompt['task'] for prompt in prompts]))
|
116 |
|
117 |
+
st.sidebar.header("**Annotation PromptCards**")
|
|
|
|
|
118 |
|
119 |
+
st.sidebar.write("A collection of prompts for annotation tasks in NLP. This is a work in progress. Please contribute your prompts via GitHub [[Upload]](https://github.com/chkla/PromptCards) 🙏🏽.")
|
120 |
|
121 |
+
# add a link to the GitHub repository
|
122 |
+
st.sidebar.write("---")
|
123 |
+
st.sidebar.write(f"**Total number of prompts:** {len(prompts)}")
|
124 |
+
st.sidebar.write("---")
|
125 |
|
126 |
+
st.sidebar.header("🧑🏽🚀 Explore:")
|
127 |
+
search = st.sidebar.text_input("Search by title")
|
128 |
+
language_filter = st.sidebar.selectbox("Filter by Language", ["All"] + language_list)
|
129 |
+
task_filter = st.sidebar.selectbox("Filter by Task", ["All"] + task_list)
|
130 |
|
131 |
+
if st.sidebar.button("Back to list"):
|
132 |
+
st.session_state.selected_prompt_id = None
|
133 |
+
st.experimental_rerun()
|
|
|
134 |
|
135 |
+
if st.session_state.selected_prompt_id is None:
|
136 |
+
filtered_prompts = [
|
137 |
+
prompt for prompt in prompts
|
138 |
+
if (search.lower() in prompt['title'].lower() or not search)
|
139 |
+
and (prompt['language'] == language_filter or language_filter == "All")
|
140 |
+
and (prompt['task'] == task_filter or task_filter == "All")
|
141 |
+
]
|
142 |
|
143 |
+
for prompt in filtered_prompts:
|
144 |
+
render_prompt_card(prompt)
|
145 |
+
else:
|
146 |
+
prompt = next((p for p in prompts if p["id"] == st.session_state.selected_prompt_id), None)
|
147 |
+
if prompt:
|
148 |
+
render_prompt_details(prompt)
|
|
|
149 |
|
150 |
+
st.sidebar.write("---")
|
151 |
+
st.sidebar.write("Made with ❤️ and 🤖 by [chkla](klamm.ai).")
|
|
|
|
|
|
|
|
|
152 |
|
153 |
+
if __name__ == "__main__":
|
154 |
+
main()
|
prompts/gilardi-frame-problem-solution-2023.md
ADDED
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
id: problem-solution-en-gilardi2023
|
3 |
+
title: Problem/ Solution Frames
|
4 |
+
author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
|
5 |
+
paperlink: https://arxiv.org/abs/2303.15056
|
6 |
+
date: 27.3.2023
|
7 |
+
language: en
|
8 |
+
task: frames
|
9 |
+
version: 1.0
|
10 |
+
addedby: chkla
|
11 |
+
keywords: frames, content moderation
|
12 |
+
---
|
13 |
+
|
14 |
+
## Prompt Description
|
15 |
+
|
16 |
+
[Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
|
17 |
+
|
18 |
+
## Prompt Text
|
19 |
+
|
20 |
+
Content moderation can be seen from two different perspectives:<br>
|
21 |
+
• Content moderation can be seen as a PROBLEM; for example, as a restriction of free speech<br>
|
22 |
+
• Content moderation can be seen as a SOLUTION; for example, as a protection from harmful speech<br>
|
23 |
+
<br>
|
24 |
+
For each tweet in the sample, follow these instructions:<br>
|
25 |
+
1. Carefully read the text of the tweet, paying close attention to details.<br>
|
26 |
+
2. Classify the tweet as describing content moderation as a problem, as a solution, or neither.<br>
|
27 |
+
<br>
|
28 |
+
Tweets should be classified as describing content moderation as a PROBLEM if they emphasize negative effects of content moderation, such as restrictions to free speech, or the biases that can emerge from decisions regarding what users are allowed to post.<br>
|
29 |
+
<br>
|
30 |
+
Tweets should be classified as describing content moderation as a SOLUTION if they emphasize positive effects of content moderation, such as protecting users from various kinds of harmful content, including hate speech, misinformation, illegal adult content, or spam.<br>
|
31 |
+
<br>
|
32 |
+
Tweets should be classified as describing content moderation as NEUTRAL if they do not emphasize possible negative or positive effects of content moderation, for example if they simply report on the content moderation activity of social media platforms without linking them to potential advantages or disadvantages for users or stakeholders.
|
33 |
+
|
34 |
+
## Language
|
35 |
+
|
36 |
+
- Prompt Language: [Specify the language of the prompt, e.g., English]
|
37 |
+
- Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
|
38 |
+
|
39 |
+
## NLP Task
|
40 |
+
|
41 |
+
- Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
|
42 |
+
- Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
|
43 |
+
|
44 |
+
## Example Input and Output
|
45 |
+
|
46 |
+
- Example 1
|
47 |
+
- Input: [Provide an example input for the prompt]
|
48 |
+
- Output: [Provide an example output for the given input]
|
49 |
+
- Example 2
|
50 |
+
- Input: [Provide another example input for the prompt]
|
51 |
+
- Output: [Provide another example output for the given input]
|
52 |
+
|
53 |
+
## Parameters and Constraints
|
54 |
+
|
55 |
+
- Parameter 1: [Specify any parameters, such as temperature or token count]
|
56 |
+
- Parameter 2: [Specify additional parameters or constraints if applicable]
|
57 |
+
|
58 |
+
## Evaluation Metrics
|
59 |
+
|
60 |
+
[List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
|
61 |
+
|
62 |
+
## Use Cases
|
63 |
+
|
64 |
+
[List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
|
65 |
+
|
66 |
+
## Limitations and Potential Biases
|
67 |
+
|
68 |
+
[Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
|
69 |
+
|
70 |
+
## Related Research and References
|
71 |
+
|
72 |
+
[List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
|
73 |
+
|
74 |
+
## Cite
|
75 |
+
|
76 |
+
>> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
|
77 |
+
|
prompts/gilardi-moderation-relevance-2023.md
ADDED
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
id: moderation-relevance-en-gilardi2023
|
3 |
+
title: Content Moderation Relevance
|
4 |
+
author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
|
5 |
+
paperlink: https://arxiv.org/abs/2303.15056
|
6 |
+
date: 27.3.2023
|
7 |
+
language: en
|
8 |
+
task: relevance
|
9 |
+
version: 1.0
|
10 |
+
addedby: chkla
|
11 |
+
keywords: content moderation, relevance
|
12 |
+
---
|
13 |
+
|
14 |
+
## Prompt Description
|
15 |
+
|
16 |
+
[Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
|
17 |
+
|
18 |
+
## Prompt Text
|
19 |
+
For each tweet in the sample, follow these instructions:<br>
|
20 |
+
1. Carefully read the text of the tweet, paying close attention to details.<br>
|
21 |
+
2. Carefully read the text of the tweet, paying close attention to details.<br>
|
22 |
+
<br>
|
23 |
+
Tweets should be coded as RELEVANT when they directly relate to content moderation, as defined above. This includes tweets that discuss: social media platforms’ content moderation rules and practices, governments’ regulation of online content moderation, and/or mild forms of content moderation like flagging.<br>
|
24 |
+
<br>
|
25 |
+
Tweets should be coded as IRRELEVANT if they do not refer to content moderation, as defined above, or if they are themselves examples of moderated content. This would include, for example, a Tweet by Donald Trump that Twitter has labeled as “disputed”, a tweet claiming that something is false, or a tweet containing sensitive content. Such tweets might be subject to content moderation, but are not discussing content moderation. Therefore, they should be coded as irrelevant for our purposes.
|
26 |
+
|
27 |
+
## Language
|
28 |
+
|
29 |
+
- Prompt Language: [Specify the language of the prompt, e.g., English]
|
30 |
+
- Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
|
31 |
+
|
32 |
+
## NLP Task
|
33 |
+
|
34 |
+
- Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
|
35 |
+
- Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
|
36 |
+
|
37 |
+
## Example Input and Output
|
38 |
+
|
39 |
+
- Example 1
|
40 |
+
- Input: [Provide an example input for the prompt]
|
41 |
+
- Output: [Provide an example output for the given input]
|
42 |
+
- Example 2
|
43 |
+
- Input: [Provide another example input for the prompt]
|
44 |
+
- Output: [Provide another example output for the given input]
|
45 |
+
|
46 |
+
## Parameters and Constraints
|
47 |
+
|
48 |
+
- Parameter 1: [Specify any parameters, such as temperature or token count]
|
49 |
+
- Parameter 2: [Specify additional parameters or constraints if applicable]
|
50 |
+
|
51 |
+
## Evaluation Metrics
|
52 |
+
|
53 |
+
[List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
|
54 |
+
|
55 |
+
## Use Cases
|
56 |
+
|
57 |
+
[List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
|
58 |
+
|
59 |
+
## Limitations and Potential Biases
|
60 |
+
|
61 |
+
[Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
|
62 |
+
|
63 |
+
## Related Research and References
|
64 |
+
|
65 |
+
[List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
|
66 |
+
|
67 |
+
## Cite
|
68 |
+
|
69 |
+
>> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
|
prompts/gilardi-policy-frames-2023.md
ADDED
@@ -0,0 +1,83 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
id: policy-frames-en-gilardi2023
|
3 |
+
title: Policy Frames Classification
|
4 |
+
author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
|
5 |
+
paperlink: https://arxiv.org/abs/2303.15056
|
6 |
+
date: 27.3.2023
|
7 |
+
language: en
|
8 |
+
task: frames
|
9 |
+
version: 1.0
|
10 |
+
addedby: chkla
|
11 |
+
keywords: policy frames, content moderation
|
12 |
+
---
|
13 |
+
|
14 |
+
## Prompt Description
|
15 |
+
|
16 |
+
[Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
|
17 |
+
|
18 |
+
## Prompt Text
|
19 |
+
|
20 |
+
Content moderation, as described above, can be linked to various other topics, such as health, crime, or equality.
|
21 |
+
For each tweet in the sample, follow these instructions:<br>
|
22 |
+
1. Carefully read the text of the tweet, paying close attention to details. 2. Classify the tweet into one of the topics defined below.<br>
|
23 |
+
<br>
|
24 |
+
The topics are defined as follows:<br>
|
25 |
+
* ECONOMY: The costs, benefits, or monetary/financial implications of the issue (to an individual, family, community, or to the economy as a whole).<br>
|
26 |
+
* Capacity and resources: The lack of or availability of physical, geographical, spatial, human, and financial resources, or the capacity of existing systems and resources to implement or carry out policy goals.<br>
|
27 |
+
* MORALITY: Any perspective—or policy objective or action (including proposed action)that is compelled by religious doctrine or interpretation, duty, honor, righ- teousness or any other sense of ethics or social responsibility.<br>
|
28 |
+
* FAIRNESS AND EQUALITY: Equality or inequality with which laws, punishment, rewards, and resources are applied or distributed among individuals or groups. Also the balance between the rights or interests of one individual or group compared to another individual or group.<br>
|
29 |
+
* CONSTITUTIONALITY AND JURISPRUDENCE: The constraints imposed on or freedoms granted to individuals, government, and corporations via the Constitution, Bill of Rights and other amendments, or judicial interpretation. This deals specifically with the authority of government to regulate, and the authority of in- dividuals/corporations to act independently of government.<br>
|
30 |
+
* POLICY PRESCRIPTION AND EVALUATION: Particular policies proposed for addressing an identified problem, and figuring out if certain policies will work, or if existing policies are effective.<br>
|
31 |
+
* LAW AND ORDER, CRIME AND JUSTICE: Specific policies in practice and their enforcement, incentives, and implications. Includes stories about enforcement and interpretation of laws by individuals and law enforcement, breaking laws, loopholes, fines, sentencing and punishment. Increases or reductions in crime.<br>
|
32 |
+
* SECURITY AND DEFENSE: Security, threats to security, and protection of one’s person, family, in-group, nation, etc. Generally an action or a call to action that can be taken to protect the welfare of a person, group, nation sometimes from a not yet manifested threat.<br>
|
33 |
+
* HEALTH AND SAFETY: Health care access and effectiveness, illness, disease, sanitation, obesity, mental health effects, prevention of or perpetuation of gun violence, infrastructure and building safety.<br>
|
34 |
+
* QUALITY OF LIFE: The effects of a policy on individuals’ wealth, mobility, access to resources, happiness, social structures, ease of day-to-day routines, quality of community life, etc.<br>
|
35 |
+
* CULTURAL IDENTITY: The social norms, trends, values and customs constituting culture(s), as they relate to a specific policy issue.<br>
|
36 |
+
* PUBLIC OPINION: References to general social attitudes, polling and demographic information, as well as implied or actual consequences of diverging from or “getting ahead of” public opinion or polls.<br>
|
37 |
+
* POLITICAL: Any political considerations surrounding an issue. Issue actions or efforts or stances that are political, such as partisan filibusters, lobbyist involvement, bipartisan efforts, deal-making and vote trading, appealing to one’s base, mentions of political maneuvering. Explicit statements that a policy issue is good or bad for a particular political party.<br>
|
38 |
+
* EXTERNAL REGULATION AND REPUTATION: The United States’ external relations with another nation; the external relations of one state with another; or relations between groups. This includes trade agreements and outcomes, comparisons of policy outcomes or desired policy outcomes.<br>
|
39 |
+
* OTHER: Any topic that does not fit into the above categories.<br>
|
40 |
+
|
41 |
+
## Language
|
42 |
+
|
43 |
+
- Prompt Language: [Specify the language of the prompt, e.g., English]
|
44 |
+
- Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
|
45 |
+
|
46 |
+
## NLP Task
|
47 |
+
|
48 |
+
- Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
|
49 |
+
- Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
|
50 |
+
|
51 |
+
## Example Input and Output
|
52 |
+
|
53 |
+
- Example 1
|
54 |
+
- Input: [Provide an example input for the prompt]
|
55 |
+
- Output: [Provide an example output for the given input]
|
56 |
+
- Example 2
|
57 |
+
- Input: [Provide another example input for the prompt]
|
58 |
+
- Output: [Provide another example output for the given input]
|
59 |
+
|
60 |
+
## Parameters and Constraints
|
61 |
+
|
62 |
+
- Parameter 1: [Specify any parameters, such as temperature or token count]
|
63 |
+
- Parameter 2: [Specify additional parameters or constraints if applicable]
|
64 |
+
|
65 |
+
## Evaluation Metrics
|
66 |
+
|
67 |
+
[List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
|
68 |
+
|
69 |
+
## Use Cases
|
70 |
+
|
71 |
+
[List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
|
72 |
+
|
73 |
+
## Limitations and Potential Biases
|
74 |
+
|
75 |
+
[Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
|
76 |
+
|
77 |
+
## Related Research and References
|
78 |
+
|
79 |
+
[List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
|
80 |
+
|
81 |
+
## Cite
|
82 |
+
|
83 |
+
>> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
|
prompts/gilardi-stance-detection-2023.md
ADDED
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
id: stance-detectionen-gilardi2023
|
3 |
+
title: Stance Detection
|
4 |
+
author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
|
5 |
+
paperlink: https://arxiv.org/abs/2303.15056
|
6 |
+
date: 27.3.2023
|
7 |
+
language: en
|
8 |
+
task: stance
|
9 |
+
version: 1.0
|
10 |
+
addedby: chkla
|
11 |
+
keywords: stance, content moderation
|
12 |
+
---
|
13 |
+
|
14 |
+
## Prompt Description
|
15 |
+
|
16 |
+
[Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
|
17 |
+
|
18 |
+
## Prompt Text
|
19 |
+
|
20 |
+
In the context of content moderation, Section 230 is a law in the United States that protects websites and other online platforms from being held legally responsible for the content posted by their users. This means that if someone posts something illegal or harmful on a website, the website itself cannot be sued for allowing it to be posted. However, websites can still choose to moderate content and remove anything that violates their own policies.<br>
|
21 |
+
<br>
|
22 |
+
For each tweet in the sample, follow these instructions:<br>
|
23 |
+
1. Carefully read the text of the tweet, paying close attention to details.<br>
|
24 |
+
2. Classify the tweet as having a positive stance towards Section 230, a negative stance, or a neutral stance.<br>
|
25 |
+
|
26 |
+
## Language
|
27 |
+
|
28 |
+
- Prompt Language: [Specify the language of the prompt, e.g., English]
|
29 |
+
- Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
|
30 |
+
|
31 |
+
## NLP Task
|
32 |
+
|
33 |
+
- Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
|
34 |
+
- Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
|
35 |
+
|
36 |
+
## Example Input and Output
|
37 |
+
|
38 |
+
- Example 1
|
39 |
+
- Input: [Provide an example input for the prompt]
|
40 |
+
- Output: [Provide an example output for the given input]
|
41 |
+
- Example 2
|
42 |
+
- Input: [Provide another example input for the prompt]
|
43 |
+
- Output: [Provide another example output for the given input]
|
44 |
+
|
45 |
+
## Parameters and Constraints
|
46 |
+
|
47 |
+
- Parameter 1: [Specify any parameters, such as temperature or token count]
|
48 |
+
- Parameter 2: [Specify additional parameters or constraints if applicable]
|
49 |
+
|
50 |
+
## Evaluation Metrics
|
51 |
+
|
52 |
+
[List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
|
53 |
+
|
54 |
+
## Use Cases
|
55 |
+
|
56 |
+
[List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
|
57 |
+
|
58 |
+
## Limitations and Potential Biases
|
59 |
+
|
60 |
+
[Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
|
61 |
+
|
62 |
+
## Related Research and References
|
63 |
+
|
64 |
+
[List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
|
65 |
+
|
66 |
+
## Cite
|
67 |
+
|
68 |
+
>> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
|
prompts/gilardi-topic-detection-2023.md
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
id: topic-detection-gilardi2023
|
3 |
+
title: Topic Detection
|
4 |
+
author: Fabrizio Gilardi, Meysam Alizadeh and Maël Kubli
|
5 |
+
paperlink: https://arxiv.org/abs/2303.15056
|
6 |
+
date: 27.3.2023
|
7 |
+
language: en
|
8 |
+
task: topic detection
|
9 |
+
version: 1.0
|
10 |
+
addedby: chkla
|
11 |
+
keywords: content moderation, topic detection
|
12 |
+
---
|
13 |
+
|
14 |
+
## Prompt Description
|
15 |
+
|
16 |
+
[Briefly describe the purpose of the prompt and the context in which it is intended to be used, especially in the context of artificial annotation with generative models.]
|
17 |
+
|
18 |
+
## Prompt Text
|
19 |
+
|
20 |
+
Tweets about content moderation may also discuss other related topics, such as:<br>
|
21 |
+
1. Section 230, which is a law in the United States that protects websites and other online platforms from being held legally responsible for the content posted by their users (SECTION 230).<br>
|
22 |
+
2. The decision by many social media platforms, such as Twitter and Facebook, to suspend Donald Trump’s account (TRUMP BAN).
|
23 |
+
3. Requests directed to Twitter’s support account or help center (TWITTER SUPPORT).<br>
|
24 |
+
4. Social media platforms’ policies and practices, such as community guidelines or terms of service (PLATFORM POLICIES).
|
25 |
+
5. Complaints about platform’s policy and practices in deplatforming and content moderation or suggestions to suspend particular accounts, or complaints about accounts being suspended or reported (COMPLAINTS).<br>
|
26 |
+
6. If a text is not about the SECTION 230, COMPLAINTS, TRUMP BAN, TWIT- TER SUPPORT, and PLATFORM POLICIES, then it should be classified in OTHER class (OTHER).<br>
|
27 |
+
<br>
|
28 |
+
For each tweet in the sample, follow these instructions:<br>
|
29 |
+
1. Carefully read the text of the tweet, paying close attention to details.<br>
|
30 |
+
2. Please classify the following text according to topic (defined by function of the text, author’s purpose and form of the text). You can choose from the following classes: SECTION 230, TRUMP BAN, COMPLAINTS, TWITTER SUPPORT, PLATFORM POLICIES, and OTHER<br>
|
31 |
+
|
32 |
+
## Language
|
33 |
+
|
34 |
+
- Prompt Language: [Specify the language of the prompt, e.g., English]
|
35 |
+
- Dataset Language: [Specify the language of the dataset to which the prompt is applied, e.g., English]
|
36 |
+
|
37 |
+
## NLP Task
|
38 |
+
|
39 |
+
- Task: [Specify the NLP task in more detail, e.g., sentiment analysis, named entity recognition, summarization]
|
40 |
+
- Subtask: [If applicable, provide any subtask or variation related to the main NLP task, e.g., binary sentiment classification, multi-class sentiment classification]
|
41 |
+
|
42 |
+
## Example Input and Output
|
43 |
+
|
44 |
+
- Example 1
|
45 |
+
- Input: [Provide an example input for the prompt]
|
46 |
+
- Output: [Provide an example output for the given input]
|
47 |
+
- Example 2
|
48 |
+
- Input: [Provide another example input for the prompt]
|
49 |
+
- Output: [Provide another example output for the given input]
|
50 |
+
|
51 |
+
## Parameters and Constraints
|
52 |
+
|
53 |
+
- Parameter 1: [Specify any parameters, such as temperature or token count]
|
54 |
+
- Parameter 2: [Specify additional parameters or constraints if applicable]
|
55 |
+
|
56 |
+
## Evaluation Metrics
|
57 |
+
|
58 |
+
[List the evaluation metrics used to assess the quality of the generated artificial annotations, such as accuracy, F1 score, or BLEU score.]
|
59 |
+
|
60 |
+
## Use Cases
|
61 |
+
|
62 |
+
[List any specific use cases or applications for the prompt in artificial annotation, such as data annotation, semi-supervised learning, or active learning.]
|
63 |
+
|
64 |
+
## Limitations and Potential Biases
|
65 |
+
|
66 |
+
[Briefly discuss any limitations or potential biases associated with the prompt, as well as any steps taken to mitigate them, in the context of artificial annotation with generative models.]
|
67 |
+
|
68 |
+
## Related Research and References
|
69 |
+
|
70 |
+
[List any relevant research papers, articles, or resources that informed the creation of the prompt or are closely related to it, especially in the area of artificial annotation with generative models. Include proper citations where applicable.]
|
71 |
+
|
72 |
+
## Cite
|
73 |
+
|
74 |
+
>> Fabrizio Gilardi, Meysam Alizadeh, Maël Kubli (2023) "ChatGPT Outperforms Crowd-Workers for Text-Annotation Tasks" [[Paper]](https://arxiv.org/abs/2303.15056)
|