Affinity Event
|
Tue 14:00
|
Invited Talk 2 by Lama Ahmad (Technical Program Manager, Trustworthy AI at OpenAI): Human and AI Evaluations for Safety and Robustness Testing
Lama Ahmad
|
|
Workshop
|
|
Evaluating Refusal
Shira Abramovich · Anna J. Ma
|
|
Workshop
|
|
Safe and Sound: Evaluating Language Models for Bias Mitigation and Understanding
Shaina Raza · Deval Pandya · Shardul ghuge · Nifemi
|
|
Workshop
|
|
Surveying Surveys: Surveys’ Role in Evaluating AI’s Labor Market Impact
Cassandra Solis
|
|
Workshop
|
Sat 12:00
|
Skilling laws: scaling laws for LLM benchmark performance
Felipe Maia Polo · Seamus Somerstep · Leshem Choshen · Yuekai Sun · Mikhail Yurochkin
|
|
Poster
|
Wed 11:00
|
Weak Supervision Performance Evaluation via Partial Identification
Felipe Maia Polo · Subha Maity · Mikhail Yurochkin · Moulinath Banerjee · Yuekai Sun
|
|
Workshop
|
|
Provocation on Expertise in Social Impact Evaluations for Generative AI (and Beyond)
Zoe Kahn · Nitin Kohli
|
|
Workshop
|
|
Multilingual Hallucination Gaps in Large Language Models
Cléa Chataigner · Afaf Taik · Golnoosh Farnadi
|
|
Tutorial
|
Tue 13:30
|
PrivacyML: Meaningful Privacy-Preserving Machine Learning and How To Evaluate AI Privacy
Mimee Xu · Dmitrii Usynin · Fazl Barez
|
|
Workshop
|
|
Evaluating the role of ‘Constitutions’ for learning from AI feedback
Saskia Redgate · Andrew M. Bean · Adam Mahdi
|
|
Workshop
|
|
AI Sandbagging: Language Models can Selectively Underperform on Evaluations
Teun van der Weij · Felix Hofstätter · Oliver Jaffe · Samuel Brown · Francis Ward
|
|
Workshop
|
|
Evaluating Chemistry Prompts for Large-Language Model Fine-Tuning
Carmelo Gonzales · Michael Pieler · Kevin Maik Jablonka · Santiago Miret
|
|