Skip to yearly menu bar Skip to main content


Search All 2024 Events
 

113 Results

<<   <   Page 1 of 10   >   >>
Affinity Event
Tue 14:00 Invited Talk 2 by Lama Ahmad (Technical Program Manager, Trustworthy AI at OpenAI): Human and AI Evaluations for Safety and Robustness Testing
Lama Ahmad
Workshop
Evaluating Refusal
Shira Abramovich · Anna J. Ma
Workshop
Safe and Sound: Evaluating Language Models for Bias Mitigation and Understanding
Shaina Raza · Deval Pandya · Shardul ghuge · Nifemi
Workshop
Surveying Surveys: Surveys’ Role in Evaluating AI’s Labor Market Impact
Cassandra Solis
Workshop
Sat 12:00 Skilling laws: scaling laws for LLM benchmark performance
Felipe Maia Polo · Seamus Somerstep · Leshem Choshen · Yuekai Sun · Mikhail Yurochkin
Poster
Wed 11:00 Weak Supervision Performance Evaluation via Partial Identification
Felipe Maia Polo · Subha Maity · Mikhail Yurochkin · Moulinath Banerjee · Yuekai Sun
Workshop
Provocation on Expertise in Social Impact Evaluations for Generative AI (and Beyond)
Zoe Kahn · Nitin Kohli
Workshop
Multilingual Hallucination Gaps in Large Language Models
Cléa Chataigner · Afaf Taik · Golnoosh Farnadi
Tutorial
Tue 13:30 PrivacyML: Meaningful Privacy-Preserving Machine Learning and How To Evaluate AI Privacy
Mimee Xu · Dmitrii Usynin · Fazl Barez
Workshop
Evaluating the role of ‘Constitutions’ for learning from AI feedback
Saskia Redgate · Andrew M. Bean · Adam Mahdi
Workshop
AI Sandbagging: Language Models can Selectively Underperform on Evaluations
Teun van der Weij · Felix Hofstätter · Oliver Jaffe · Samuel Brown · Francis Ward
Workshop
Evaluating Chemistry Prompts for Large-Language Model Fine-Tuning
Carmelo Gonzales · Michael Pieler · Kevin Maik Jablonka · Santiago Miret