Workshop
|
Sat 15:45
|
Auto-Evaluation with Few Labels through Post-hoc Regression
Benjamin Eyre · David Madras
|
|
Poster
|
|
SG-Bench: Evaluating LLM Safety Generalization Across Diverse Tasks and Prompt Types
Yutao Mou · Shikun Zhang · Wei Ye
|
|
Affinity Event
|
|
Evaluating Privacy Risks in Synthetic Clinical Text Generation in Spanish
Luis Miranda · Jocelyn Dunstan · Matías Toro · Federico Olmedo · Felix Melo
|
|
Affinity Event
|
|
Evaluating the Usefulness of Large Language Models for Synthetic Samples Generation via Few-shot Learning
Maynara Souza · Flávio Santos · Cleber Zanchettin
|
|
Session
|
Thu 16:30
|
Dialogue with the Machine and Dialogue with the Art World: A Method for Evaluating AI as a Tool for Creativity
Remi Denton · Farbod Mehr · Aroussiak Gabriellan · Rida Qadri · Huma Gupta · Pamela Karimi · Piotr Mirowski
|
|
Affinity Event
|
|
Implement and evaluate adversarial attack and defense in Deep Learning using MNIST Dataset
Henriette Flore KENNE
|
|
Affinity Event
|
|
Ontology Extraction and Evaluation for the Blue Amazon
Vivian Magri Alcaldi Soares · Renata Wassermann
|
|
Affinity Event
|
|
LLM Unlearning EKG: Evaluations using Knowledge Graphs
Rushali Mohbe · Samuel Scarpino
|
|
Affinity Event
|
|
Evaluation of Bagging Ensembles on Multimodal Data for Breast Cancer Diagnosis
Abdulganiyu Jimoh
|
|
Affinity Event
|
|
Evaluating Simplicity over Complexity: Autoencoder Architectures for Real-Time Vehicle Damage Detection
Sara Khan
|
|
Affinity Event
|
|
Reasoning-Driven Jury System for LLM Evaluation
Ayda Sultan
|
|
Affinity Event
|
|
Evaluating Generative AI for Scenario Variation in Automated Driving Validation
Manasa Mariam Mammen · Zafer Kayatas · Eva Zimmermann · Pavel Nedvědický
|
|