evijit HF staff commited on
Commit
96c888b
1 Parent(s): 463659e

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +38 -39
index.html CHANGED
@@ -37,10 +37,16 @@
37
  <div class="container is-max-desktop">
38
  <div class="columns is-centered">
39
  <div class="column has-text-centered">
40
- <h1 class="title is-1 publication-title">Evaluating Evaluations</h1>
41
  <h2 class="subtitle is-3 publication-subtitle">Examining Best Practices for Measuring Broader Impacts of Generative AI</h2>
42
  <div class="is-size-5 publication-authors">
43
- <span class="author-block">A NeurIPS 2024 Workshop</span>
 
 
 
 
 
 
44
  </div>
45
  </div>
46
  </div>
@@ -58,10 +64,10 @@
58
  Generative AI systems are becoming increasingly prevalent in society, producing content such as text, images, audio, and video with far-reaching implications. While the NeurIPS Broader Impact statement has notably shifted norms for AI publications to consider negative societal impact, no standard exists for how to approach these impact assessments. This workshop aims to address this critical gap by bringing together experts on evaluation science and practitioners who develop and analyze technical systems.
59
  </p>
60
  <p>
61
- Building upon our previous initiatives, including the FAccT 2023 CRAFT session "Assessing the Impacts of Generative AI Systems Across Modalities and Society" and our initial "Evaluating the Social Impact of Generative AI Systems" report, we have made significant strides in this area. Through these efforts, we collaboratively developed an evaluation framework and guidance for assessing generative systems across modalities. We have since crowdsourced evaluations and analyzed gaps in literature and systemic issues around how evaluations are designed and selected.
62
  </p>
63
  <p>
64
- The goal of this workshop is to share our existing findings with the NeurIPS community and collectively develop future directions for effective community-built evaluations. By fostering collaboration between experts and practitioners, we aim to create more comprehensive evaluations and develop urgently needed policy recommendations for governments and AI safety organizations.
65
  </p>
66
  </div>
67
  </div>
@@ -71,20 +77,28 @@
71
 
72
  <section class="section">
73
  <div class="container is-max-desktop">
74
- <h2 class="title is-3">Call for Papers (CFP)</h2>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
75
  <div class="content has-text-justified">
76
  <p>We are soliciting tiny papers (up to 2 pages long) in the following formats:</p>
77
  <ol>
78
  <li>Extended Abstracts: Short but complete research papers presenting original or interesting results around social impact evaluation for generative AI.</li>
79
  <li>"Provocations": Novel perspectives or challenges to conventional wisdom around social impact evaluation for generative AI.</li>
80
  </ol>
81
- <h3 class="title is-4">Submission Guidelines</h3>
82
- <ul>
83
- <li>Paper Length: Maximum 2 pages, including references</li>
84
- <li>Format: PDF file, using the NeurIPS conference format</li>
85
- <li>Submission Portal: [Insert submission portal link here]</li>
86
- <li>Anonymity: Submissions should be anonymous for blind review</li>
87
- </ul>
88
  <h3 class="title is-4">Themes for Submissions</h3>
89
  <p>We welcome submissions addressing, but not limited to, the following themes:</p>
90
  <ol>
@@ -108,6 +122,15 @@
108
  <li>Comparative analyses of existing evaluation frameworks and their effectiveness</li>
109
  <li>Case studies of social impact evaluations in real-world AI applications</li>
110
  </ol>
 
 
 
 
 
 
 
 
 
111
  <h3 class="title is-4">Important Dates</h3>
112
  <ul>
113
  <li>Submission Deadline: August 1, 2024</li>
@@ -163,7 +186,7 @@
163
  <td>Talks + Provocations</td>
164
  <td>
165
  <ul>
166
- <li>Invited speakers present on current technical evaluations for base models across all modalities</li>
167
  <li>Key social impact categories covered: Bias and stereotyping, Cultural values, Performance disparities, Privacy, Financial and environmental costs, Data moderator labor</li>
168
  <li>Presentations of accepted provocations</li>
169
  </ul>
@@ -228,7 +251,7 @@
228
  <strong>Nitarshan Rajkumar</strong>
229
  <ul>
230
  <li>Cofounder of UK AI Safety Institute</li>
231
- <li>Adviser to the Secretary of State of UK Department for Science, Innovation and Technology</li>
232
  </ul>
233
  </li>
234
  <li>
@@ -287,28 +310,4 @@
287
  </div>
288
  </section>
289
 
290
- <section class="section">
291
- <div class="container is-max-desktop">
292
- <h2 class="title is-3">Contact Information</h2>
293
- <div class="content has-text-justified">
294
- <p>For any queries regarding the workshop or submission process, please contact:</p>
295
- <p>[Insert contact information for workshop organizers]</p>
296
- </div>
297
- </div>
298
- </section>
299
-
300
- <footer class="footer">
301
- <div class="container">
302
- <div class="content has-text-centered">
303
- <p>
304
- Workshop on Evaluating Evaluations: Examining Best Practices for Measuring Broader Impacts of Generative AI
305
- </p>
306
- <p>
307
- Website template borrowed from the <a href="https://github.com/nerfies/nerfies.github.io">nerfies</a> project page.
308
- </p>
309
- </div>
310
- </div>
311
- </footer>
312
-
313
- </body>
314
- </html>
 
37
  <div class="container is-max-desktop">
38
  <div class="columns is-centered">
39
  <div class="column has-text-centered">
40
+ <h1 class="title is-1 publication-title">Evaluating Evaluations (2024)</h1>
41
  <h2 class="subtitle is-3 publication-subtitle">Examining Best Practices for Measuring Broader Impacts of Generative AI</h2>
42
  <div class="is-size-5 publication-authors">
43
+ <span class="author-block">A NeurIPS Workshop</span>
44
+ </div>
45
+ <div class="is-size-5 publication-authors">
46
+ <span class="author-block">co-located with <a href="https://neurips.cc/" target="_blank">NeurIPS 2024</a></span>
47
+ </div>
48
+ <div class="is-size-5 publication-authors">
49
+ <span class="author-block">[Date + maybe zoom link]</span>
50
  </div>
51
  </div>
52
  </div>
 
64
  Generative AI systems are becoming increasingly prevalent in society, producing content such as text, images, audio, and video with far-reaching implications. While the NeurIPS Broader Impact statement has notably shifted norms for AI publications to consider negative societal impact, no standard exists for how to approach these impact assessments. This workshop aims to address this critical gap by bringing together experts on evaluation science and practitioners who develop and analyze technical systems.
65
  </p>
66
  <p>
67
+ Building upon our previous initiatives, including the FAccT 2023 CRAFT session "Assessing the Impacts of Generative AI Systems Across Modalities and Society" and our initial "Evaluating the Social Impact of Generative AI Systems" report, we have made significant strides in this area. Through these efforts, we collaboratively developed an evaluation framework and guidance for assessing generative systems across modalities. We have since crowdsourced evaluations and analyzed gaps in the literature and systemic issues around how evaluations are designed and selected, resulting in a more comprehensive second edition of the paper.
68
  </p>
69
  <p>
70
+ The goal of this workshop is to share existing findings with the NeurIPS community and collectively develop future directions for effective community-driven evaluations. A key focus is participatory AI: Wide benefits can be gained from the scope and involving all participants, not just domain experts. By encouraging collaboration among experts, practitioners, and the wider community, the workshop aims to create more comprehensive evaluations and develop urgent policy recommendations for governments and AI safety organizations.
71
  </p>
72
  </div>
73
  </div>
 
77
 
78
  <section class="section">
79
  <div class="container is-max-desktop">
80
+ <h2 class="title is-3">Workshop Objectives</h2>
81
+ <div class="content">
82
+ <ol>
83
+ <li>Share existing findings and methodologies with the NeurIPS community</li>
84
+ <li>Collectively develop future directions for effective community-built evaluations</li>
85
+ <li>Address barriers to broader adoption of social impact evaluation of Generative AI systems</li>
86
+ <li>Develop policy recommendations for investment in future directions for social impact evaluations</li>
87
+ <li>Create a framework for documenting and standardizing evaluation practices</li>
88
+ </ol>
89
+ </div>
90
+ </div>
91
+ </section>
92
+
93
+ <section class="section">
94
+ <div class="container is-max-desktop">
95
+ <h2 class="title is-3">Call for Tiny Papers</h2>
96
  <div class="content has-text-justified">
97
  <p>We are soliciting tiny papers (up to 2 pages long) in the following formats:</p>
98
  <ol>
99
  <li>Extended Abstracts: Short but complete research papers presenting original or interesting results around social impact evaluation for generative AI.</li>
100
  <li>"Provocations": Novel perspectives or challenges to conventional wisdom around social impact evaluation for generative AI.</li>
101
  </ol>
 
 
 
 
 
 
 
102
  <h3 class="title is-4">Themes for Submissions</h3>
103
  <p>We welcome submissions addressing, but not limited to, the following themes:</p>
104
  <ol>
 
122
  <li>Comparative analyses of existing evaluation frameworks and their effectiveness</li>
123
  <li>Case studies of social impact evaluations in real-world AI applications</li>
124
  </ol>
125
+ <h3 class="title is-4">Submission Guidelines</h3>
126
+ <ul>
127
+ <li>Paper Length: Maximum 2 pages, excluding an unlimited amount of references</li>
128
+ <li>Format: PDF file, using the <a href="#" target="_blank">NeurIPS 2024 LaTeX style file</a></li>
129
+ <li>Submission Portal: [Insert submission portal link here]</li>
130
+ <li>Anonymity: Submissions should be anonymous for two-way anonymized review.</li>
131
+ <li>This is a participatory, in-person event. Accepted Authors are encouraged to present their work and discuss it at the event.</li>
132
+ <li>Broader impact statement and Limitation section are not counted in the paper length.</li>
133
+ </ul>
134
  <h3 class="title is-4">Important Dates</h3>
135
  <ul>
136
  <li>Submission Deadline: August 1, 2024</li>
 
186
  <td>Talks + Provocations</td>
187
  <td>
188
  <ul>
189
+ <li>Invited speakers to present on current technical evaluations for base models across all modalities</li>
190
  <li>Key social impact categories covered: Bias and stereotyping, Cultural values, Performance disparities, Privacy, Financial and environmental costs, Data moderator labor</li>
191
  <li>Presentations of accepted provocations</li>
192
  </ul>
 
251
  <strong>Nitarshan Rajkumar</strong>
252
  <ul>
253
  <li>Cofounder of UK AI Safety Institute</li>
254
+ <li>Adviser to the Secretary of State of the UK Department for Science, Innovation and Technology</li>
255
  </ul>
256
  </li>
257
  <li>
 
310
  </div>
311
  </section>
312
 
313
+ <section class="section">