Workshop
Adaptive Experimental Design and Active Learning in the Real World
Willie Neiswanger · Mojmir Mutny · Ilija Bogunovic · Ava Amini · Zi Wang · Stefano Ermon · Andreas Krause
Room 208 - 210
Join us for an insightful workshop on adaptive experimental design and active learning. Dive into their use in fields like computational biology, materials discovery, chip design, and more.
Schedule
Sat 6:15 a.m. - 6:30 a.m.
|
Doors Open and Workshop/Poster Setup
(
Setup
)
>
|
🔗 |
Sat 6:30 a.m. - 6:40 a.m.
|
Opening Remarks
(
Remarks
)
>
SlidesLive Video |
🔗 |
Sat 6:40 a.m. - 7:20 a.m.
|
High-throughput Protein Design and Evolution for Synthetic Biology - Erika DeBenedictis
(
Invited Talk
)
>
SlidesLive Video Natural proteins evolved over billions of years, with numerous populations using chance and selection to create useful proteins. This talk will describe PRANCE, an high-throughput protein engineering method that mimics natural evolution. By combining liquid handling robotics and molecular engineering techniques, PRANCE allows us to overcome features of natural evolution that make it challenging to use as an engineering technique, like stochasticity and extinction. These advances make directed evolution more interpretable, accessible, and reproducible. We show that high-throughput directed evolution creates new engineering approaches for large-scale biological engineering challenges like genetic code expansion. |
🔗 |
Sat 7:20 a.m. - 8:00 a.m.
|
From Theory to Impact: Unlocking the Power of Bayesian Optimization on Real-World Science and Engineering Systems - Joel Paulson
(
Invited Talk
)
>
SlidesLive Video Bayesian optimization (BO) is a powerful tool for optimizing non-convex black-box (also known as derivative-free) functions that are expensive or time-consuming to evaluate and subject to noise in their evaluations. Many important problems can be formulated in this manner, such as optimizing outcomes of high-fidelity computer simulations, automated hyperparameter tuning in machine and deep learning algorithms, A/B testing for website design, policy-based reinforcement learning, and material and drug discovery. In this presentation, three key concepts are introduced, which we argue are critical for enabling and/or improving the practical performance of BO on real-world science and engineering systems. Specifically, one must: (1) leverage prior physics-based knowledge to perform highly efficient (targeted) exploration of the solution space; (2) explicitly incorporate safety constraints during interaction with physical systems to avoid unsafe, unethical, and/or undesirable outcomes; and (3) account for external sources of uncertainty during the search process to ensure the best-identified solution is robust/flexible in practice. We discuss a unified framework for adapting BO to handle these considerations and illustrate how this framework can be deployed in practice on a series of examples ranging from the design of safe cold plasma jet devices to the discovery of high-performance sustainable energy storage materials. We also offer perspectives on key challenges and future opportunities in the realm of applied BO. |
🔗 |
Sat 8:00 a.m. - 8:20 a.m.
|
Coffee Break
(
Coffee Break
)
>
|
🔗 |
Sat 8:20 a.m. - 9:00 a.m.
|
Adaptive Clinical Trials: Reinforcement Learning's Most Formidable Challenge - Mihaela van der Schaar
(
Invited Talk
)
>
SlidesLive Video |
🔗 |
Sat 9:00 a.m. - 9:40 a.m.
|
Near-Optimal Non-Parametric Sequential Tests and Confidence Sequences with Possibly Dependent Observations - Nathan Kallus
(
Invited Talk
)
>
SlidesLive Video Sequential tests and their implied confidence sequences, which are valid at arbitrary stopping times, promise flexible statistical inference and on-the-fly decision making. However, strong guarantees are limited to parametric sequential tests, which suffer high type-I error rates in practice because reality isn't parametric, or to concentration-bound-based sequences, which are overly conservative so we get wide intervals and take too long to detect effects. We consider a classic delayed-start normal-mixture sequential probability ratio test and provide the first asymptotic (in the delay) analysis under general non-parametric data generating processes. We guarantee type-I-error rates approach a user-specified α-level (primarily by leveraging a martingale strong invariance principle). Moreover, we show that the expected time-to-reject approaches the minimum possible among all α-level tests (primarily by leveraging an identity inspired by Itô's lemma). Together, our results establish these (ostensibly parametric) tests as general-purpose, non-parametric, and near-optimal. We illustrate this via numerical experiments and a retrospective re-analysis of A/B tests at Netflix. |
🔗 |
Sat 9:40 a.m. - 11:20 a.m.
|
Poster Session and Lunch
(
Poster Session
)
>
|
🔗 |
Sat 11:20 a.m. - 12:00 p.m.
|
Invited Talk 5: Emma Brunskill
(
Invited Talk
)
>
SlidesLive Video |
🔗 |
Sat 12:00 p.m. - 1:10 p.m.
|
Best Paper Candidate Spotlight Talks
(
Spotlight Talks
)
>
SlidesLive Video |
🔗 |
Sat 1:10 p.m. - 1:30 p.m.
|
Coffee Break
(
Coffee Break
)
>
|
🔗 |
Sat 1:30 p.m. - 2:10 p.m.
|
Adaptive Experimentation without an Objective - Eytan Bakshy
(
Invited Talk
)
>
SlidesLive Video Real-world experimentation often involves making complex tradeoffs between many outcomes or targeting noisy or long-term objectives that are difficult to measure. This problem can be particularly challenging in the context of continuous action spaces where an infinite number of tradeoffs can be achieved. I will discuss these problems in the context of real-world problems faced by experimenters at Meta. I will discuss solutions to the many objective problem via learning from human feedback. I will conclude with strategies for targeting long-term outcomes and open questions. |
🔗 |
Sat 2:10 p.m. - 2:50 p.m.
|
RL in the Real World: From Chip Design to LLMs - Anna Goldie
(
Invited Talk
)
>
SlidesLive Video Reinforcement learning (RL) is famously powerful but difficult to wield, and until recently, had demonstrated impressive results on games, but little real world impact. I will start the talk with a discussion of RL for Large Language Models (LLMs), including scalable supervision techniques to better align models with human preferences (Constitutional AI / RLAIF). Next, I will discuss RL for chip floorplanning, one of the first examples of RL solving a real world engineering problem. This learning-based method can generate placements that are superhuman or comparable on modern accelerator chips in a matter of hours, whereas the strongest baselines require human experts in the loop and can take several weeks. This method was published in Nature and used in production to generate superhuman chip layouts for the last four generations of Google’s flagship AI accelerator (TPU), including the recently announced TPU v5p. |
🔗 |
Sat 2:50 p.m. - 3:00 p.m.
|
Best Paper Award and Closing Remarks
(
Remarks
)
>
SlidesLive Video |
🔗 |
Sat 3:00 p.m. - 3:30 p.m.
|
Poster Session
(
Poster Session
)
>
|
🔗 |
-
|
Model-Free Preference Elicitation
(
Poster
)
>
link
Elicitation of user preferences is an effective way to improve the quality of recommendations, especially when there is little or no user history. In this setting, a recommendation system interacts with the user by asking questions and recording the responses. Various criteria have been proposed for optimizing the sequence of queries in order to improve understanding of user preferences and thereby the quality of downstream recommendations. A compelling approach for preference elicitation is the Expected Value of Information (EVOI), a Bayesian approach which computes the expected gain in user utility for possible queries. Previous work on EVOI has focused on probabilistic models of users for computing posterior utilities. In contrast, in this work we explore model-free variants of EVOI which rely on function approximations in order to avoid strong modeling assumptions. Specifically, we propose to learn a user response model and user utility model from existing data, which is often available in real-world systems, and to use these models in EVOI in place of the probabilistic models. We show promising empirical results on a preference elicitation task using our approach. |
Carlos Martin · Craig Boutilier · Ofer Meshi 🔗 |
-
|
Active Learning for Iterative Offline Reinforcement Learning
(
Poster
)
>
link
Offline Reinforcement Learning (RL) has emerged as a promising approach to addressreal-world challenges where online interactions with the environment are limited, risky,or costly. Although, recent advancements produce high quality policies from offline data,currently, there is no systematic methodology to continue to improve them without resortingto online fine-tuning. This paper proposes to repurpose Offline RL to produce a sequenceof improving policies, namely, Iterative Offline Reinforcement Learning (IORL). To producesuch sequence, IORL has to cope with imbalanced offline datasets and to perform controlledenvironment exploration. Specifically, we introduce ”Return-based Sampling” as meansto selectively prioritize experience from high-return trajectories and active learning driven”Dataset Uncertainty Sampling” to probe state-actions inversely proportional to densityin the dataset.We demonstrate that our proposed approach produces policies that achievemonotonically increasing average returns, from 65.4 to 140.2, in the Atari environment. |
Lan Zhang · Luigi Franco Tedesco · Pankaj Rajak · Youcef Zemmouri · Hakan Brunzell 🔗 |
-
|
Transferable Candidate Proposal with Bounded Uncertainty
(
Poster
)
>
link
From an empirical perspective, the subset chosen through active learning cannot guarantee an advantage over random sampling when transferred to another model. While it underscores the significance of verifying transferability, experimental design from previous works often neglected that the informativeness of a data subset can change over model configurations. To tackle this issue, we introduce a new experimental design, coined as Candidate Proposal, to find transferable data candidates from which active learning algorithms choose the informative subset. Correspondingly, a data selection algorithm is proposed, namely Transferable candidate proposal with Bounded Uncertainty (TBU), which constrains the pool of transferable data candidates by filtering out the presumably redundant data points based on uncertainty estimation. We verified the validity of TBU in image classification benchmarks, including CIFAR-10/100 and SVHN. When transferred to different model configurations, TBU consistency improves performance in existing active learning algorithms. Our code is available at https://github.com/gokyeongryeol/TBU. |
Kyeongryeol Go · Kye-Hyeon Kim 🔗 |
-
|
Sequentially Adaptive Experimentation for Learning Optimal Options subject to Unobserved Contexts
(
Poster
)
>
link
Contextual bandits constitute a classical framework for interactive learning of best decisions subject to context information. In this setting, the goal is to sequentially learn arms of highest reward subject to the contextual information, while the unknown reward parameters of each arm need to be learned by experimenting it. Accordingly, a fundamental problem is that of balancing such experimentation (i.e., pulling different arms to learn the parameters), versus sticking with the best arm learned so far, in order to maximize rewards. To study this problem, the existing literature mostly considers perfectly observed contexts. However, the setting of partially observed contexts remains unexplored to date, despite being theoretically more general and practically more versatile. We study bandit policies for learning to select optimal arms based on observations, which are noisy linear functions of the unobserved context vectors. Our theoretical analysis shows that adaptive experiments based on samples from the posterior distribution efficiently learn optimal arms. Specifically, we establish regret bounds that grow logarithmically with time. Extensive simulations for real-world data are presented as well to illustrate this efficacy. |
Hongju Park · Mohamad Kazem Shirani Faradonbeh 🔗 |
-
|
Decentralized Multi-Agent Active Search and Tracking when Targets Outnumber Agents
(
Poster
)
>
link
Multi-agent multi-target tracking has a wide range of applications, including wildlife patrolling, security surveillance or environment monitoring. Such algorithms often assume that agents are pre-assigned to monitor disjoint partitions of the environment, reducing the burden of exploration. This limits applicability when there are fewer agents than targets, since agents are unable to continuously follow the targets in their fields of view. Multi-agent tracking algorithms additionally assume a central controller and synchronous inter-agent communication. Instead, we focus on the setting of decentralized multi-agent, multi-target, simultaneous active search-and-tracking with asynchronous inter-agent communication. Our proposed algorithm MASTER uses a sequential monte carlo implementation of the probability hypothesis density filter for posterior inference combined with Thompson sampling for decentralized multi-agent decision making. We compare different action selection policies, focusing on scenarios where targets outnumber agents. In simulation, MASTER outperforms baselines in terms of the Optimal Sub-Pattern Assignment (OSPA) metric for different numbers of targets and varying teamsizes. |
Arundhati Banerjee · Jeff Schneider 🔗 |
-
|
Anytime-Valid Inference in Adaptive Experiments
(
Poster
)
>
link
We consider the problem of efficient, anytime-valid statistical inference of the Average Treatment Effect (ATE) in a sequential experiment where the assignment of subjects to treatment or control can be made adaptively over time. We relax assumptions necessary forasymptotic normality of the Adaptive Augmented Inverse-Probability Weighted (A2IPW) estimator introduced by Kato et al. (2021), which is semiparametrically efficient due to the adaptive assignment. With the aim of enabling continuous data analysis as the data is being collected, we then derive both asymptotic and nonasymptotic confidence sequences that are considerably tighter than previous methods. In addition to sharper inference tools, we use propensity score truncation techniques from the recent off-policy estimation literature to reduce finite sample variance of our estimator without affecting asymptotic variance (which is optimal). Empirical results demonstrate that our methods yield narrower confidence sequences that maintain time-uniform error control. |
Thomas Cook · Alan Mishler · Aaditya Ramdas 🔗 |
-
|
Class Balanced Dynamic Acquisition for Domain Adaptive Semantic Segmentation using Active Learning
(
Poster
)
>
link
Domain adaptive active learning is leading the charge in label-efficient training of neural networks. For semantic segmentation, state-of-the-art models jointly use two criteria of uncertainty and diversity to select training labels, combined with a pixel-wise acquisition strategy. However, we show that such methods currently suffer from a class imbalance issue which degrades their performance for larger active learning budgets. We then introduce Class Balanced Dynamic Acquisition (CBDA), a novel active learning method that mitigates this issue, especially in high-budget regimes. The more balanced labels increase minority class performance, which in turn allows the model to outperform the previous baseline by 0.6, 1.7, and 2.4 mIoU for budgets of 5%, 10%, and 20%, respectively. Additionally, the focus on minority classes leads to improvements of the minimum class performance of 0.5, 2.9, and 4.6 IoU respectively.The top-performing model even exceeds the fully supervised baseline, showing that a more balanced label than the entire ground truth can be beneficial. |
Marc Schachtsiek · Simone Rossi · Thomas Hannagan 🔗 |
-
|
Pure Exploration under Mediators’ Feedback
(
Poster
)
>
link
Stochastic multi-armed bandits are a sequential-decision-making framework, where, at each interaction step, the learner selects an arm and observes a stochastic reward. Within the context of best-arm identification (BAI) problems, the goal of the agent lies in finding the optimal arm, i.e., the one with the highest expected reward, as accurately and efficiently as possible. Nevertheless, the sequential interaction protocol of classical BAI problems, where the agent has complete control over the arm being pulled at each round, does not effectively model several decision-making problems of interest (e.g., off-policy learning, human feedback). For this reason, in this work, we propose a novel strict generalization of the classical BAI problem that we refer to as best-arm identification under mediators’ feedback (BAI-MF). More specifically, we consider the scenario in which the learner has access to a set of mediators, each of which selects the arms on the agent’s behalf according to a stochastic and possibly unknown policy. The mediator, then, communicates back to the agent the pulled arm together with the observed reward. In this setting, the agent’s goal lies in sequentially choosing which mediator to query to identify with high probability the optimal arm while minimizing the identification time, i.e., the sample complexity. To this end, we first derive and analyze a statistical lower bound on the sample complexity specific to our general mediator feedback scenario. Then, we propose a sequential decision-making strategy for discovering the best arm; as our theory verifies, this algorithm matches the lower bound both almost surely and in expectation. |
Riccardo Poiani · Alberto Maria Metelli · Marcello Restelli 🔗 |
-
|
DISTRIBUTIONALLY ROBUST MODEL-BASED REINFORCEMENT LEARNING WITH LARGE STATE SPACES
(
Poster
)
>
link
Three major challenges in reinforcement learning are the complex dynamical systems with large state spaces, the costly data acquisition processes, and the deviation of real-world dynamics from the training environment deployment. To overcome these issues, we study distributionally robust Markov decision processes with continuous state spaces under the widely used Kullback–Leibler, chi-square, and total variation uncertainty sets. We propose a model-based approach that utilizes Gaussian Processes and the maximum variance reduction algorithm to efficiently learn multi-output nominal transition dynamics, leveraging access to a generative model (i.e., simulator). We further demonstrate the statistical sample complexity of the proposed method for different uncertainty sets. These complexity bounds are independent of the number of states and extend beyond linear dynamics, ensuring the effectiveness of our approach in identifying near-optimal distributionally-robust policies. The proposed method can be further combined with other model-free distributionally robust reinforcement learning methods to obtain a near-optimal robust policy. Experimental results demonstrate the robustness of our algorithm to distributional shifts and its superior performance in terms of the number of samples needed. |
Shyam Sundhar Ramesh · Pier Giuseppe Sessa · Yifan Hu · Andreas Krause · Ilija Bogunovic 🔗 |
-
|
Sustainable Concrete via Bayesian Optimization
(
Poster
)
>
link
Eight percent of global carbon dioxide emissions can be attributed to the production of cement, the main component of concrete, which is also the dominant source of CO2 emissions in the construction of data centers. The discovery of lower-carbon concrete formulas is therefore of high significance for sustainability. However, experimenting with new concrete formulae is time consuming and labor intensive, as one usually has to wait to record the concrete’s 28-day compressive strength, a quantity whose measurement can by its definition not be accelerated. This provides an opportunity for experimental design methodology like Bayesian Optimization (BO) to accelerate the search for strong and sustainable concrete formulae. Herein, we 1) propose modeling steps that make concrete strength amenable to be predicted accurately by a Gaussian process model with relatively few measurements, 2) formulate the search for sustainable concrete as a multi-objective optimization problem, and 3) leverage the proposed model to carry out multi-objective BO with real-world strength measurements of the algorithmically proposed mixes. Our experimental results show improved trade-offs between the mixtures’ global warming potential (GWP) and their associated compressive strengths, compared to mixes based on current industry practices. |
Sebastian Ament · Andrew Witte · Nishant Garg · Julius Kusuma 🔗 |
-
|
Active Model Selection: A Variance Minimization Approach
(
Poster
)
>
link
The cost of labeling is a significant challenge in practical machine learning.This issue arises not only during the learning phase but also at the model evaluation phase, as there is a need for a substantial amount of labeled test data in addition to the training data.In this study, we address the challenge of active model selection with the goal of minimizing labeling costs for choosing the best-performing model from a set of model candidates.Based on an appropriate test loss estimator, we propose an adaptive labeling strategy that can estimate the difference of test losses with small variance, thereby enabling the estimation of the best model using fewer labeling cost.Experimental results on real-world datasets confirm that our method efficiently selects the best model. |
Mitsuru Matsuura · Satoshi Hara 🔗 |
-
|
Preference-Guided Bayesian Optimization for Control Policy Learning: Application to Personalized Plasma Medicine
(
Poster
)
>
link
This paper investigates the adaptation of control policies for personalized dose delivery in plasma medicine using preference-learning based Bayesian optimization. Preference learning empowers users to incorporate their preferences or domain expertise during the exploration of optimal control policies, which often results in fast attainment of personalized treatment outcomes. We establish that, compared to multi-objective Bayesian optimization (BO), preference-guided BO offers statistically faster convergence and computes solutions that better reflect user preferences. Moreover, it enables users to actively provide feedback during the policy search procedure, which helps to focus the search in sub-regions of the search space likely to contain preferred local optima. Our findings highlight the suitability of preference-learning-based BO for adapting control policies in plasma treatments, where both user preferences and swift convergence are of paramount importance. |
Ketong Shao · Diego Romeres · Ankush Chakrabarty · Ali Mesbah 🔗 |
-
|
Residual Deep Gaussian Processes on Manifolds for Geometry-aware Bayesian Optimization on Hyperspheres
(
Poster
)
>
link
Gaussian processes (GPs) are a widely-used model class for approximating unknown functions, especially useful in tasks such as Bayesian optimisation, where accurate uncertainty estimates are key. Deep Gaussian processes (DGPs) are a multi-layered generalisation of GPs, which promises improved performance at modelling complex functions. Some of the problems where GPs and DGPs may be utilised involve data on manifolds like hyperspheres. Recent work has recognised this, generalising scalar-valued and vector-valued Matérn GPs to a broad class of Riemannian manifolds. Despite that, an appropriate analogue of DGP for Riemannian manifolds is missing. We introduce a new model, residual manifold DGP, and a suitable doubly stochastic variational inference technique that helps train and deploy it on hyperspheres. Through examination on stylised examples, we highlight the usefulness of residual deep manifold GPs on regression tasks and in Bayesian optimisation. |
Kacper Wyrwal · Slava Borovitskiy 🔗 |
-
|
Exploratory Training: When Annotators Learn About Data
(
Poster
)
>
link
ML systems often present examples and solicit labels from users to learn a target model, i.e., active learning. However, due to the complexity of the underlying data, users may not initially have a perfect understanding of the effective model and do not know the accurate labeling. For example, a user who is training a model for detecting noisy or abnormal values may not perfectly know the properties of typical and clean values in the data. Users may improve their knowledge about the data and target model as they observe examples during training. As users gradually learn about the data and model, they may revise their labeling strategies. Current systems assume that users always provide correct labeling with potentially a fixed and small chance of annotation mistakes. Nonetheless, if the trainer revises its belief during training, such mistakes become significant and non-stationarity. Hence, current systems consume incorrect labels and may learn inaccurate models. In this paper, we build theoretical underpinnings and design algorithms to develop systems that collaborate with users to learn the target model accurately and efficiently. At the core of our proposal, a game-theoretic framework models the joint learning of user and system to reach a desirable eventual stable state, where both user and system share the same belief about the target model. We extensively evaluate our system using user studies over various real-world datasets and show that our algorithms lead to accurate results with a smaller number of interactions compared to existing methods. |
Rajesh Shrestha · Omeed Habibelahian · Arash Termehchy · Paolo Papotti 🔗 |
-
|
Graph Neural Bayesian Optimization for Virtual Screening
(
Poster
)
>
link
Virtual screening is an essential component of early-stage drug and materials discovery. This is challenged by the increasingly intractable size of virtual libraries and the high cost of evaluating properties. We propose GNN-SS, a Graph Neural Network (GNN) powered Bayesian Optimization (BO) algorithm. GNN-SS utilizes random sub-sampling to reduce the computational complexity of the BO problem, and diversifies queries for training the model. We further introduce data-independent projections to efficiently model second-order random feature interactions, and improve uncertainty estimates. GNN-SS is computationally light, sample-efficient, and rapidly narrows the search space by leveraging the generalization ability of GNNs. Our algorithm achieves state-of-the-art performance among screening methods for the Practical Molecular Optimization benchmark. |
Miles Wang-Henderson · Bartu Soyuer · Parnian Kassraie · Andreas Krause · Ilija Bogunovic 🔗 |
-
|
Zooming Optimistic Optimization Method to solve the Threshold Estimation Problem
(
Poster
)
>
link
This paper introduces a new global optimization algorithm that solves the threshold estimation problem.In this active learning problem, underlying many empirical neuroscience and psychophysics experiments, the objective is to estimate the input values that would produce the desired output value from an unknown, noisy, non-decreasing response function. Compared to previous approaches, ZOOM (Zooming Optimistic Optimization Method) offers the best of both worlds: ZOOM is model-agnostic, benefits from stronger theoretical guarantees and faster convergence rate, but also quickly jumps between arms, offering strong performance even for small sampling budgets. |
Julien Audiffren 🔗 |
-
|
Hessian-Free Laplace in Bayesian Deep Learning
(
Poster
)
>
link
The Laplace approximation (LA) of the Bayesian posterior is a Gaussian distribution centered at the maximum a posteriori estimate. Its appeal in Bayesian deep learning stems from the ability to quantify uncertainty post-hoc (i.e., after optimizing network parameters), the ease of sampling from the approximate posterior, and the analytic form of model evidence. Uncertainty in turn can direct experimentation. However, an important computational bottleneck of LA is the necessary step of calculating and inverting the Hessian matrix of the log posterior. The Hessian may be approximated in a variety of ways, with quality varying with a number of factors including the network, dataset, and inference task. In this paper, we propose an alternative algorithm that sidesteps Hessian calculation and inversion. The Hessian-free Laplace (HFL) approximation uses curvature of both the log posterior and network prediction to estimate its variance. Two point estimates are required: the standard maximum a posteriori parameters and optimal parameters under a loss regularized by the network prediction. We show that under standard assumptions of LA in Bayesian deep learning, HFL targets the same variance as LA, and this is empirically explored in small-scale simulated experiments. |
James McInerney · Nathan Kallus 🔗 |
-
|
On Complex Network Dynamics of an In-Vitro Neuronal System during Rest and Gameplay
(
Poster
)
>
link
In this study, we characterize complex network dynamics in live in vitro neuronal systems during two distinct activity states: spontaneous rest state and engagement in a real-time (closed-loop) game environment using the DishBrain system. First, we embed the spiking activity of these channels in a lower-dimensional space using various representation learning methods and then extract a subset of representative channels. Next, by analyzing these low-dimensional representations, we explore the patterns of macroscopic neuronal network dynamics during learning. Remarkably, our findings indicate that just using the low-dimensional embedding of representative channels is sufficient to differentiate the neuronal culture during the Rest and Gameplay. Notably, our investigation shows dynamic changes in the connectivity patterns within the same region and across multiple regions on the multi-electrode array only during Gameplay. These findings underscore the plasticity of neuronal networks in response to external stimuli and highlight the potential for modulating connectivity in a controlled environment. The ability to distinguish between neuronal states using reduced-dimensional representations points to the presence of underlying patterns that could be pivotal for real-time monitoring and manipulation of neuronal cultures. Additionally, this provides insight into how biological based information processing systems rapidly adapt and learn and may lead to new improved algorithms. |
Moein Khajehnejad · Forough Habibollahi · Alon Loeffler · Brett J. Kagan · Adeel Razi 🔗 |
-
|
Human-in-the-Loop Out-of-Distribution Detection with False Positive Rate Control
(
Poster
)
>
link
Robustness to Out-of-Distribution (OOD) samples is essential for successful deployment of machine learning models in the open world. Since it is not possible to have a priori access to variety of OOD data before deployment, several recent works have focused on designing scoring functions to quantify OOD uncertainty. These methods often find a threshold that achieves 95% true positive rate (TPR) on the In-Distribution (ID) data used for training and use this threshold for detecting OOD samples. However, this can lead to very high FPR as seen in a comprehensive evaluation in the Open-OOD benchmark, the FPR can range between 60 to 96% on several ID and OOD dataset combinations. In contrast, practical systems deal with a variety of OOD samples on the fly and critical applications, e.g., medical diagnosis, demand guaranteed control of the false positive rate (FPR). To meet these challenges, we propose a mathematically grounded framework for human-in-the-loop OOD detection, wherein expert feedback is used to update the threshold. This allows the system to adapt to variations in the OOD data while adhering to the quality constraints. We propose an algorithm that uses any time valid confidence intervals based on the Law of Iterated Logarithm (LIL). Our theoretical results show that the system meets FPR constraints while minimizing the human feedback for point that are in-distribution. Another key feature of the system is that it can work with any existing post-hoc OOD uncertainty-quantification methods. We evaluate our system empirically on a mixture of benchmark OOD datasets in image classification task on CIFAR-10 and CIFAR-100 as in distribution datasets and show that our method can maintain FPR at most 5% while maximizing TPR. |
Harit Vishwakarma · Heguang Lin · Ramya Korlakai Vinayak 🔗 |
-
|
Cross-Entropy Estimators for Sequential Experiment Design with Reinforcement Learning
(
Poster
)
>
link
Reinforcement learning can learn amortised design policies for designing sequences of experiments. However, current methods rely on contrastive estimators of expected information gain, which require an exponential number of contrastive samples to achieve an unbiased estimation. We propose the use of an alternative lower bound estimator, based on the cross-entropy of the joint model distribution and a flexible proposal distribution. This proposal distribution approximates the true posterior of the model parameters given the experimental history and the design policy. Our method requires no contrastive samples, can achieve more accurate estimates of high information gains, allows learning of superior design policies, and is compatible with implicit probabilistic models. We assess our algorithm's performance in various tasks, including continuous and discrete designs and explicit and implicit likelihoods. |
Tom Blau · Iadine Chades · Amir Dezfouli · Daniel Steinberg · Edwin Bonilla 🔗 |
-
|
Correlated Trajectory Uncertainty for Adaptive Sequential Decision Making
(
Poster
)
>
link
One of the great challenges with decision making tasks on real world systems is the fact that data is sparse and acquiring additional data is expensive. In these cases, it is often crucial to make a model of the environment to assist in making decisions. At the same time, limited data means that learned models are erroneous, making it just as important to equip the model with good predictive uncertainties. In the context of learning sequential decision making policies, these uncertainties can prove useful for informing which data to collect for the greatest improvement in policy performance \citep{mehta2021experimental, mehta2022exploration} or informing the policy about unsure regions of state and action space to avoid during test time \citep{yu2020mopo}. Additionally, assuming that realistic samples of the environment can be drawn, an adaptable policy can be trained that attempts to make optimal decisions for any given possible instance of the environment \citep{ghosh2022offline, chen2021offline}. In this work, we examine the so-called ``probabilistic neural network'' (PNN) model that is ubiquitous in model-based reinforcement learning (MBRL) works. We argue that while PNN models may have good marginal uncertainties, they form a distribution of non-smooth transition functions. Not only are these samples unrealistic and may hamper adaptability, but we also assert that this leads to poor uncertainty estimates when predicting multiple step trajectory estimates. To address this issue, we propose a simple sampling method that can be implemented on top of pre-existing models.We evaluate our sampling technique on a number of environments, including a realistic nuclear fusion task, and find that, not only do smooth transition function samples produce more calibrated uncertainties, but they also lead to better downstream performance for an adaptive policy. |
Ian Char · Youngseog Chung · Rohan Shah · Willie Neiswanger · Jeff Schneider 🔗 |
-
|
Efficient and scalable reinforcement learning via Hypermodel
(
Poster
)
>
link
Data-efficient reinforcement learning(RL) requires deep exploration. Thompson sampling is a principled method for deep exploration in reinforcement learning. However, Thompson sampling need to track the degree of uncertainty by maintaining the posterior distribution of models, which is computationally feasible only in simple environments with restrictive assumptions. A key problem in modern RL is how to develop data and computation efficient algorithm that is scalable to large-scale complex environments. We develop a principled framework, called HyperFQI, to tackle both the computation and data efficiency issues. HyperFQI can be regarded as approximate Thompson sampling for reinforcement learning based on hypermodel. Hypermodel in this context serves as the role for uncertainty estimation of action-value function. HyperFQI demonstrates its ability for efficient and scalable deep exploration in DeepSea benchmark with large state space. HyperFQI also achieves super-human performance in Atari benchmark with low computation costs and low data consumption. We also give a rigorous performance analysis for the proposed method, justifying its computation and data efficiency. To the best of knowledge, this is the first principled RL algorithm that is provably efficient and also practically scalable to complex environments such as Arcade learning environment that requires deep networks for pixel-based control. |
Yingru Li · Jiawei Xu · Zhi-Quan Luo 🔗 |
-
|
ALAS: Active Learning for Autoconversion Rates Prediction from Satellite Data
(
Poster
)
>
link
High-resolution simulations, such as the ICOsahedral Non-hydrostatic Large-Eddy Model (ICON-LEM), provide valuable insights into the complex interactions among aerosols, clouds, and precipitation, which are the major contributors to climate change uncertainty. However, due to its exorbitant computational costs, it can only be employed for a limited period and geographical area. To address this, we propose a more cost-effective method powered by emerging machine learning approach to better understand the intricate dynamics of the climate system. Our approach involves active learning techniques -- by leveraging high-resolution climate simulation as the oracle and an abundant amount of unlabeled data drawn from satellite observations -- to predict autoconversion rates, a crucial step in precipitation formation, while significantly reducing the need for a large number of labeled instances. In this study, we present novel methods: custom query strategy fusion for labeling instances, WiFi and MeFi, along with active feature selection based on SHAP, designed to tackle real-world challenges due to its simplicity and practicality in application, specifically focusing on the prediction of autoconversion rates. |
Maria Carolina Novitasari · Johannes Quaas · Miguel Rodrigues 🔗 |
-
|
Near-equivalence between bounded regret and delay robustness in interactive decision making
(
Poster
)
>
link
Interactive decision making, encompassing bandits, contextual bandits, and reinforcement learning, has recently been of interest to theoretical studies of experimentation design and recommender system algorithm research. Recently, it has been shown that the well-known Graves-Lai constant being zero is a necessary and sufficient condition for achieving bounded (or constant) regret in interactive decision making. As this condition may be a strong requirement for many applications, the practical usefulness of pursuing bounded regret has been questioned. In this paper, we show that the condition of the Graves-Lai constant being zero is also necessary to achieve delay model robustness when reward delays are unknown (i.e., when feedbacks are anonymous). Here, model robustness is measured in terms of $\epsilon$-robustness, one of the most widely used and one of the least adversarial robustness concepts in the robust statistics literature. In particular, we show that $\epsilon$-robustness cannot be achieved for a consistent (i.e., uniformly sub-polynomial regret) algorithm however small the nonzero $\epsilon$ value is when the Grave-Lai constant is not zero. While this is a strongly negative result, we also provide a positive result for linear rewards models (Linear contextual bandits, Reinforcement learning with linear MDP) that the Grave-Lai constant being zero is also sufficient for achieving bounded regret without any knowledge of delay models, i.e., the best of both the efficiency world and the delay robustness world.
|
Enoch H. Kang · P. R. Kumar 🔗 |
-
|
Expert-guided Bayesian Optimisation for Human-in-the-loop Experimental Design of Known Systems
(
Poster
)
>
link
Domain experts often possess valuable physical insights that are overlooked in fully automated decision-making processes such as Bayesian optimisation. In this article we apply high-throughput (batch) Bayesian optimisation alongside anthropological decision theory to enable domain experts to influence the selection of optimal experiments. Our methodology exploits the hypothesis that humans are better at making discrete choices than continuous ones and enables experts to influence critical early decisions. At each iteration we solve an augmented multi-objective optimisation problem across a number of alternate solutions, maximising both the sum of their utility function values and the determinant of their covariance matrix, equivalent to their total variability. By taking the solution at the knee point of the Pareto front, we return a set of alternate solutions at each iteration that have both high utility values and are reasonably distinct, from which the expert selects one for evaluation. We demonstrate that even in the case of an uninformed practitioner, our algorithm recovers the regret of standard Bayesian optimisation. |
Tom Savage · Antonio del Rio Chanona 🔗 |
-
|
Active Testing of Binary Classification Model Using Level Set Estimation
(
Poster
)
>
link
In this study, we propose a method for estimating the test loss in binary classification model with minimal labeling of the test data. The central idea of the proposed method is to reduce the problem of test loss estimation to the problem of level set estimation for the loss function. This reduction allows us to achieve sequential test loss estimation through iterative labeling using active learning methods for level set estimation. Through multiple dataset experiments, we confirmed that the proposed method is effective for evaluating binary classification models and allows for test loss estimation with fewer labeled samples compared to existing methods. |
Takuma Ochiai · Keiichiro Seno · Kota Matsui · Satoshi Hara 🔗 |
-
|
Nonparametric Discrete Choice Experiments with Machine Learning Guided Adaptive Design
(
Poster
)
>
link
Designing products to meet consumers' preferences is essential for a business's success. We propose Gradient-based Survey (GBS), a discrete choice experiment for multiattribute product design. The experiment elicits consumer preferences through a sequence of paired comparisons for partial profiles. GBS adaptively constructs paired comparison questions based on the respondents' previous choices. Unlike the traditional random utility maximization paradigm, GBS is robust to model misspecification by not requiring a parametric utility model. Cross-pollinating the machine learning and experiment design, GBS is scalable to products with hundreds of attributes and can design personalized products for heterogeneous consumers. We demonstrate the advantage of GBS in accuracy and sample efficiency compared to the existing parametric and nonparametric methods in simulations. |
Mingzhang Yin · Ruijiang Gao · Weiran Lin · Steven Shugan 🔗 |
-
|
Active Learning Policies for Solving Inverse Problems
(
Poster
)
>
link
In recent years, solving inverse problems for black-box simulators has become a point of focus for the machine learning community due to their ubiquity in science and engineering scenarios. In such settings, the simulator describes a forward process $f: (\psi, x) \rightarrow y$ from simulator parameters $\psi$ and input data $x$ to observations $y$, and the goal of the inverse problem is to optimise $\psi$ to minimise some observation loss. Simulator gradients are often unavailable or prohibitively expensive to obtain, making optimisation of these simulators particularly challenging. Moreover, in many applications, the goal is to solve a family of related inverse problems. Thus, starting optimisation ab-initio/from-scratch may be infeasible if the forward model is expensive to evaluate. In this paper, we propose a novel method for solving classes of similar inverse problems. We learn an active learning policy that guides the training of a surrogate and use the gradients of this surrogate to optimise the simulator parameters with gradient descent. After training the policy, downstream inverse problem optimisations require up to 90\% fewer forward model evaluations than the baseline.
|
Tim Bakker · Thomas Hehn · Tribhuvanesh Orekondy · Arash Behboodi · Fabio Valerio Massoli 🔗 |
-
|
Physics-Enhanced Multi-fidelity Learning for Optical Surface Imprint
(
Poster
)
>
link
Human fingerprints serve as one unique and powerful characteristic for each person, from which policemen can recognize the identity. Similar to humans, many natural bodies and intrinsic mechanical qualities can also be uniquely identified from surface characteristics. To measure the elasto-plastic properties of one material, one formally sharp indenter is pushed into the measured body under constant force and retracted, leaving a unique residual imprint of the minute size from several micrometers to nanometers. However, one great challenge is how to map the optical image of this residual imprint into the real wanted mechanical properties, i.e., the tensile force curve. In this paper, we propose a novel method to use multi-fidelity neural networks (MFNN) to solve this inverse problem. We first actively train the NN model via pure simulation data, and then bridge the sim-to-real gap via transfer learning. The most innovative part is that we use NN to dig out the unknown physics and also implant the known physics into the transfer learning framework, thus highly improving the model stability and decreasing the data requirement. This work serves as one great example of applying machine learning into the real experimental research, especially under the constraints of data limitation and fidelity variance. |
Yongchao Chen 🔗 |
-
|
Generalized Objectives in Adaptive Experimentation: The Frontier between Within- and Post-Experiment Objectives
(
Poster
)
>
link
This paper formulates a generalized model of multi-armed bandit experiments that accommodates both cumulative regret minimization and best-arm identification objectives. We identify the optimal instance-dependent scaling of the cumulative cost across experimentation and deployment, which is expressed in the familiar form uncovered by Lai and Robbins (1985). We show that the nature of asymptotically efficient algorithms is nearly independent of the cost functions, emphasizing a remarkable universality phenomenon. Balancing various cost considerations is reduced to an appropriate choice of exploitation rate. Additionally, we explore the Pareto frontier between the length of experiment and the cumulative regret across experimentation and deployment. A notable and universal feature is that even a slight reduction in the exploitation rate from one results in a substantial decrease in the experiment's length, accompanied by only a minimal increase in the cumulative regret. |
Chao Qin · Daniel Russo 🔗 |
-
|
ExPT: Scaling Foundation Models for Experimental Design via Synthetic Pretraining
(
Poster
)
>
link
Experimental design is a fundamental problem in many science and engineering fields. In this problem, sample efficiency is crucial due to the time, money, and safety costs of real-world design evaluations. Existing approaches either rely on active data collection or access to large, labeled datasets of past experiments, making them impractical in many real-world scenarios. In this work, we address the more challenging yet realistic setting of few-shot experimental design, where only a few labeled data points of input designs and their corresponding values are available. We approach this problem as a conditional generation task, where a model conditions on a few labeled examples and the desired output to generate an optimal input design. To this end, we present Pretrained Transformers for Experimental Design (ExPT), which uses a novel combination of synthetic pretraining with in-context learning to enable few-shot generalization. In ExPT, we only assume knowledge of a finite collection of unlabelled data points from the input domain and pretrain a transformer neural network to optimize diverse synthetic functions defined over this domain. Unsupervised pretraining allows ExPT to adapt to any design task at test time in an in-context fashion by conditioning on a few labeled data points from the target task and generating the candidate optima. We evaluate ExPT on few-shot experimental design in challenging domains and demonstrate its superior generality and performance compared to existing methods. |
Tung Nguyen · Sudhanshu Agrawal · Aditya Grover 🔗 |
-
|
AutODEx: Automated Optimal Design of Experiments Platform with Data- and Time-Efficient Multi-Objective Optimization
(
Poster
)
>
link
We introduce AutODEx, an automated machine learning platform for optimal design of experiments to expedite solution discovery with optimal objective trade-offs. We implement state-of-the-art multi-objective Bayesian optimization (MOBO) algorithms in a unified and flexible framework for optimal design of experiments, along with efficient asynchronous batch strategies extended to MOBO to harness experiment parallelization. For users with little or no experience with coding or machine learning, we provide an intuitive graphical user interface (GUI) to help quickly visualize and guide the experiment design. For experienced researchers, our modular code structure serves as a testbed to quickly customize, develop, and evaluate their own MOBO algorithms. Extensive benchmark experiments against other MOBO packages demonstrate \platname's competitive and stable performance. Furthermore, we showcase \platname's real-world utility by autonomously guiding hardware experiments with minimal human involvement. |
Yunsheng Tian · Pavle Konakovic · Beichen Li · Ane Zuniga · Michael Foshey · Timothy Erps · Wojciech Matusik · Mina Konakovic Lukovic 🔗 |
-
|
Accelerated High-Entropy Alloys Discovery for Electrocatalysis via Robotic-Aided Active Learning
(
Poster
)
>
link
This work explores the accelerated discovery of High-Entropy Alloys electrocatalysts using a novel carbothermal shock fabrication method, underpinned by an active learning approach. A high-throughput robotic platform, integrating a BoTorch-based active learning module with an Opentrons liquid handling robot and a 7-axis robotic arm, expedites the iterative experimental cycles. The recent integration of large language models leverages ChatGPT’s API, facilitating voice-driven interactions between researchers and the automation setup, further enhancing the autonomous workflow under experimental materials science scenarios. Initial optimization efforts for green hydrogen production catalyst yield promising results, showcasing the efficacy of the active learning framework in navigating the complex materials design space of HEAs. This study also emphasizes the crucial need for consistency and reproducibility in real-world experiments to fully harness the potential of active learning in materials science explorations. |
Zhichu Ren · Zhen Zhang · Yunsheng Tian · Ju Li 🔗 |
-
|
Long-run Behaviour of Multi-fidelity Bayesian Optimisation
(
Poster
)
>
link
Multi-fidelity Bayesian Optimisation (MFBO) has been shown to generally converge faster than single-fidelity Bayesian Optimisation (SFBO) (\cite{poloczek2017multi}). Inspired by recent benchmark papers, we are investigating the long-run behaviour of MFBO, based on observations in the literature that it might under-perform in certain scenarios (\cite{mikkola2023multi}, \cite{eggensperger2021hpobench}). An under-performance of MBFO in the long-run could significantly undermine its application to many research tasks, especially when we are not able to identify when the under-performance begins, and other BO algorithms would have performed better. We create a simple benchmark study, showcase empirical results and discuss scenarios, concluding with inconclusive results. |
Gbetondji Dovonon · Jakob Zeitler 🔗 |
-
|
Understanding Threshold-based Auto-labeling: The Good, the Bad, and the Terra Incognita
(
Poster
)
>
link
Creating large-scale high-quality labeled datasets is a major bottleneck in supervised machine learning workflows. Threshold-based auto-labeling (TBAL), where validation data obtained from humans is used to find a confidence threshold above which the data is machine-labeled, reduces reliance on manual annotation. TBAL is emerging as a widely-used solution in practice. Given the long shelf-life and diverse usage of the resulting datasets, understanding when the data obtained by such auto-labeling systems can be relied on is crucial. This is the first work to analyze TBAL systems and derive sample complexity bounds on the amount of human-labeled validation data required for guaranteeing the quality of machine-labeled data. Our results provide two crucial insights. First, reasonable chunks of unlabeled data can be automatically and accurately labeled by seemingly bad models. Second, a hidden downside of TBAL systems is potentially prohibitive validation data usage. Together, these insights describe the promise and pitfalls of using such systems. We validate our theoretical guarantees with extensive experiments on synthetic and real datasets. |
Harit Vishwakarma · Heguang Lin · Frederic Sala · Ramya Korlakai Vinayak 🔗 |
-
|
Probabilistic Generative Modeling for Procedural Roundabout Generation for Developing Countries
(
Poster
)
>
link
Due to limited resources and fast economic growth, designing optimal transportation road networks with traffic simulation and validation in a cost-effective manner is vital for developing countries, where extensive manual testing is expensive and often infeasible. Current rule-based road design generators lack diversity, a key feature for design robustness. Generative Flow Networks (GFlowNets) learn stochastic policies to sample from an unnormalized reward distribution, thus generating high-quality solutions while preserving their diversity. In this work, we formulate the problem of linking incident roads to the circular junction of a roundabout by a Markov decision process, and we leverage GFlowNets as the Junction-Art road generator. We compare our method with related methods and our empirical results show that our method achieves better diversity while preserving a high validity score. |
Zarif Ikram · Ling Pan · Dianbo Liu 🔗 |
-
|
Efficient Variational Sequential Information Control
(
Poster
)
>
link
We develop a family of fast variational methods for sequential control in dynamical settings where an agent is incentivized to maximize information gain. We consider the case of optimal control in continuous nonlinear dynamical systems that prohibit exact evaluation of the mutual information (MI) reward. Our approach couples efficient message-passing inference with variational bounds on the MI objective under Gaussian projections. We also develop a Gaussian mixture approximation that enables exact MI evaluation under constraints on the component covariances. We validate our methodology in nonlinear systems with superior and faster control compared to standard particle-based methods. We show our approach improves the accuracy and efficiency of one-shot robotic learning with intrinsic MI rewards. |
Jianwei Shen · Jason Pacheco 🔗 |
-
|
Anytime Model Selection in Linear Bandits
(
Poster
)
>
link
Model selection in the context of bandit optimization is a challenging problem, as it requires balancing exploration and exploitation not only for action selection, but also for model selection. One natural approach is to rely on online learning algorithms that treat different models as experts. Existing methods, however, scale poorly ($\mathrm{poly}M$) with the number of models $M$ in terms of their regret. We develop \alexp, an anytime algorithm, which has an exponentially improved ($\log M$) dependence on $M$ for its regret. We neither require knowledge of the horizon $n$, nor rely on an initial purely exploratory stage. Our approach utilizes a novel time-uniform analysis of the Lasso, by defining a self-normalized martingale sequence based on the empirical process error, establishing a new connection between interactive learning and high-dimensional statistics.
|
Parnian Kassraie · Nicolas Emmenegger · Andreas Krause · Aldo Pacchiano 🔗 |
-
|
Multi-Fidelity Active Learning with GFlowNets
(
Poster
)
>
link
Many relevant scientific and engineering problems present challenges where current machine learning methods cannot yet efficiently leverage the available data and resources. For example, certain relevant problems involve exploring very large, structured and high-dimensional spaces, and where querying a high fidelity, black-box objective function is very expensive. Progress in machine learning methods that can efficiently tackle such problems would help accelerate currently crucial areas such as drug and materials discovery. In this paper, we propose a multi-fidelity active learning algorithm with GFlowNets as a sampler, to efficiently discover diverse, high-scoring candidates where multiple approximations of the black-box function are available at lower fidelity and cost. Our evaluation on molecular discovery tasks show that multi-fidelity active learning with GFlowNets can discover high-scoring candidates at a fraction of the budget of its single-fidelity counterpart while maintaining diversity, unlike RL-based alternatives. These results open new avenues for multi-fidelity active learning to accelerate scientific discovery and engineering design. |
Alex Hernandez-Garcia · Nikita Saxena · Moksh Jain · Chenghao Liu · Yoshua Bengio 🔗 |
-
|
Learning in Clinical Trial Settings
(
Poster
)
>
link
This paper presents an approach to active learning that considers the non-independent and identically distributed (non-i.i.d.) structure of a clinical trial setting. There exists two types of clinical trials: retrospective and prospective. Retrospective clinical trials analyze data after treatment has been performed; prospective clinical trials collect data as treatment is ongoing. Traditional active learning approaches are often unrealistic in practice and assume the dataset is i.i.d. when selecting training samples; however, in the case of clinical trials, treatment results in a dependency between the data collected at the current and past visits. Thus, we propose prospective active learning to overcome the limitations present in traditional active learning methods, where we condition on the time data was collected. We compare our proposed method to the traditional active learning paradigm, which we refer to as retrospective in nature, on one clinical trial dataset and one non-clinical trial dataset. We show that in clinical trial settings, our proposed method outperforms retrospective active learning. |
Zoe Fowler · Kiran Kokilepersaud · Mohit Prabhushankar · Ghassan AlRegib 🔗 |
-
|
Preferential Heteroscedastic Bayesian Optimization with Informative Noise Priors
(
Poster
)
>
link
Bayesian optimization is a robust framework for optimizing black-box, expensive-to-evaluatefunctions. It is often the case that black-box functions can only be queried by means ofpairwise comparisons between two candidate solutions, a setting known as PreferentialBO (PBO), for which efficient algorithms have been designed. Nevertheless, for high-dimensional problems, performing a comparison becomes cumbersome, typically for human subjects, and the binary information provided by a preference might become buriedin the noise induced by human answers in such a way that PBO becomes challenging. Tocircumvent this issue, we propose to account for the aleatoric uncertainty using suitableheteroscedastic noise models, based on an informative noise prior built from a user-specifiedset of reliable inputs. We empirically evaluate the proposed approach on a range of synthetic black-box functions, demonstrating a consistent improvement over homoscedasticPBO. |
Marshal Sinaga · Julien Martinelli · Samuel Kaski 🔗 |
-
|
Optimistic Games for Combinatorial Bayesian Optimization with Applications to Protein Design
(
Poster
)
>
link
Bayesian optimization (BO) is a powerful framework to optimize black box expensive-to-evaluate functions via sequential interactions. In several important problems (e.g. drug discovery, circuit design, neural architecture search, etc.), though, such functions are defined over $\textit{combinatorial and unstructured}$ spaces. This makes existing BO algorithms not feasible due to the intractable maximization of the acquisition function to find informative evaluation points. To address this issue, we propose $\textbf{GameOpt}$, a novel game-theoretical approach to combinatorial BO. $\textbf{GameOpt}$ establishes a cooperative game between the different optimization variables and computes informative points to be game $\textit{equilibria}$ of the acquisition function. These are stable configurations from which no variable has an incentive to deviate -- analogous to local optima in continuous domains. Crucially, this allows us to efficiently break down the complexity of the combinatorial domain into individual decision sets, making $\textbf{GameOpt}$ scalable to large combinatorial spaces. We demonstrate the application of $\textbf{GameOpt}$ to the challenging $\textit{protein design}$ problem and validate its performance on two real-world protein datasets. Each protein can take up to $20^{X}$ possible configurations, where $X$ is the length of a protein, making standard BO methods unusable. Instead, our approach iteratively selects informative protein configurations and very quickly discovers highly active protein variants compared to other baselines.
|
Melis Ilayda Bal · Pier Giuseppe Sessa · Mojmir Mutny · Andreas Krause 🔗 |
-
|
NPC-NIS: Navigating Semiconductor Process Corners with Neural Importance Sampling
(
Poster
)
>
link
Traditional corner case analysis in semiconductor circuit design typically involves theuse of predetermined semiconductor process parameters, including Fast, Typical, and Slowcorners for PMOS and NMOS devices, frequently yielding overly conservative designs dueto the utilization of fixed, and potentially non-representative, process parameter valuesfor circuit simulations. Identifying the worst cases of circuit FoMs within typical semiconductor process variation ranges presents a considerable challenge, especially given thecomplexities associated with accurately sampling rare semiconductor events. In response,we introduce NPC-NIS, a model specifically developed for estimating rare cases in semiconductor circuit analysis, leveraging a learnable importance sampling strategy. We modelthe distribution of process parameters that exhibit the worst FoMs within a realistic range.This adaptable framework dynamically identifies and addresses rare semiconductor caseswithin typical process variation ranges, enhancing our circuit design optimization capabilities under realistic conditions. Our empirical results validate the effectiveness of the NeuralImportance Sampling (NIS) approach in identifying and mitigating rare semiconductor scenarios, thereby contributing to the development of more robust and reliable semiconductorcircuit designs and connecting traditional semiconductor corner case analysis with realworld semiconductor applications. |
Hong Chul Nam · Chanwoo Park 🔗 |
-
|
Planning Contextual Adaptive Experiments with Model Predictive Control
(
Poster
)
>
link
Implementing adaptive experimentation methods in the real world often encounters a multitude of operational difficulties, including batched/delayed feedback, non-stationary environments, and constraints on treatment allocations. To improve the flexibility of adaptive experimentation, we propose a Bayesian, optimization-based framework founded on model-predictive control (MPC) for the linear contextual bandit setting. While we focus on simple regret minimization, the framework can flexibly incorporate multiple objectives along with constraints, batches, personalized and non-personalized policies, as well as predictions of future context arrivals. Most importantly, it maintains this flexibility while guaranteeing improvement over non-adaptive A/B testing across all time horizons, and empirically outperforms standard policies such as Thompson Sampling. Overall, this framework offers a way to guide adaptive designs across the varied demands of modern large-scale experiments. |
Ethan Che · Jimmy Wang · Hongseok Namkoong 🔗 |
-
|
Learning relevant contextual variables within Bayesian optimization
(
Poster
)
>
link
Contextual Bayesian Optimization (CBO) efficiently optimizes black-box, expensive-to-evaluate functions with respect to design variables, while simultaneously integrating relevant contextual information regarding the environment, such as experimental conditions.However, the relevance of contextual variables is not necessarily known beforehand. Moreover, contextual variables can sometimes be optimized themselves, an overlooked setting bycurrent CBO algorithms. Optimizing contextual variables may be costly, which raises thequestion of determining a minimal relevant subset. We address this problem using a novelmethod, Sensitivity-Analysis-Driven Contextual BO (SADCBO). We learn the relevance ofcontext variables by sensitivity analysis of the posterior surrogate model, whilst minimizing the cost of optimization by leveraging recent developments on early stopping for BO.We empirically evaluate our proposed SADCBO against alternatives on both synthetic andreal-world experiments, and demonstrate a consistent improvement across examples. |
Julien Martinelli · Ayush Bharti · Armi Tiihonen · Louis Filstroff · ST John · Sabina Sloman · Patrick Rinke · Samuel Kaski 🔗 |
-
|
PINNACLE: PINN Adaptive ColLocation and Experimental points selection
(
Poster
)
>
link
Physics-Informed Neural Networks (PINNs), which incorporate PDEs as soft constraints, train with a composite loss function that contains multiple training point types: different types of collocation points chosen during training to enforce each PDE and initial/boundary conditions, and experimental points which are usually costly to obtain via experiments or simulations. Training PINNs using this loss function is challenging as it typically requires selecting large numbers of points of different types, each with different training dynamics. Unlike past works that focused on the selection of either collocation or experimental points, this work introduces PINN Adaptive ColLocation and Experimental points selection (PINNACLE), the first algorithm that jointly optimizes the selection of all training point types, while automatically adjusting the proportion of collocation point types as training progresses. PINNACLE uses information on the interactions among training point types, which had not been considered before, based on an analysis of PINN training dynamics via the Neural Tangent Kernel (NTK). We theoretically show that the criterion used by PINNACLE is related to the PINN generalization error, and empirically demonstrate that PINNACLE is able to outperform existing point selection methods for forward, inverse, and transfer learning problems. |
Gregory Kang Ruey Lau · Apivich Hemachandra · See-Kiong Ng · Bryan Kian Hsiang Low 🔗 |
-
|
Accelerating Black-Box Molecular Property Optimization by Adaptively Learning Sparse Subspaces
(
Poster
)
>
link
Molecular property optimization (MPO) problems are inherently challenging since they are formulated over discrete, unstructured spaces and the labeling process involves expensive simulations or experiments, which fundamentally limits the amount of available data. Bayesian optimization (BO), which is a powerful and popular framework for efficient optimization of noisy, black-box objective functions (e.g., measured property values), thus is a potentially attractive framework for MPO. To apply BO to MPO problems, one must select a structured molecular representation that enables construction of a probabilistic surrogate model. Many molecular representations have been developed, however, they are all high-dimensional, which introduces important challenges in the BO process – mainly because the curse of dimensionality makes it difficult to define and perform inference over a suitable class of surrogate models. This challenge has been recently addressed by learning a lower-dimensional encoding of a SMILE or graph representation of a molecule in an unsupervised manner and then performing BO in the encoded space. In this work, we show that such methods have a tendency to “get stuck,” which we hypothesize occurs since the mapping from the encoded space to property values is not necessarily well-modeled by a Gaussian process. We argue for an alternative approach that combines numerical molecular descriptors with a sparse axis-aligned Gaussian process model, which is capable of rapidly identifying sparse subspaces that are most relevant to modeling the unknown property function. We demonstrate that our proposed method substantially outperforms existing MPO methods on a variety of benchmark and real-world problems. Specifically, we show that our method can routinely find near-optimal molecules out of a set of more than > 100k alternatives within 100 or fewer expensive queries. |
Farshud Sorourifar · Thomas Banker · Joel Paulson 🔗 |
-
|
Ever Evolving Evaluator (EV3): Towards Flexible and Reliable Meta-Optimization for Knowledge Distillation
(
Poster
)
>
link
We introduce EV3, a novel meta-optimization framework designed to efficiently train scalable machine learning models through an intuitive explore-assess-adapt protocol. In each iteration of EV3, we explore various model parameter updates, assess them using pertinent evaluation methods, and adapt the model based on the optimal updates and previous progress history. EV3 offers substantial flexibility without imposing stringent constraints like differentiability on the key objectives relevant to the tasks of interest. Moreover, this protocol welcomes updates with biased gradients and allows for the use of a diversity of losses and optimizers. Additionally, in scenarios with multiple objectives, it can be used to dynamically prioritize tasks. With inspiration drawn from evolutionary algorithms, meta-learning, and neural architecture search, we investigate an application of EV3 to knowledge distillation. Our experimental results illustrate EV3's capability to safely explore model spaces, while hinting at its potential applicability across numerous domains due to its inherent flexibility and adaptability. |
Li Ding · Masrour Zoghi · Guy Tennenholtz · Maryam Karimzadehgan 🔗 |
-
|
Provably-Convergent Bayesian Source Seeking for Multimodal Fields using Mobile Agents
(
Poster
)
>
link
We consider source seeking tasks, where the goal is to locate a source using a mobile agent that gathers potentially noisy measurements of the signal emitted by the source. This arises, for example, when searching for a radioactive or chemical source using mobile physical sensors that detect particles carried by the wind. In this work, we propose an iterative Bayesian algorithm for source seeking, especially well-suited for challenging environments where the signal intensity is multimodal and observations are noisy. At every step, our algorithm computes a Bayesian posterior distribution over the source location using prior physical knowledge of the observation process along with the observations collected so far. Then, it decides where the agent should move and observe next by following a search strategy that implicitly considers paths to the source's most likely location under the posterior. We show that the trajectory of an agent executing the proposed algorithm converges to the source location asymptotically with probability one. We validate the algorithm's convergence through simulated experiments of an agent seeking a source of a chemical plume in a turbulent environment. |
Vivek Mishra · Raul Astudillo · Peter Frazier · Fumin Zhang 🔗 |
-
|
Compute-Efficient Active Learning
(
Poster
)
>
link
Active learning, a powerful paradigm in machine learning, aims at reducing labeling costs by selecting the most informative samples from an unlabeled dataset. However, traditional active learning process often demands extensive computational resources, hindering scalability and efficiency. In this paper, we address this critical issue by presenting a novel method designed to alleviate the computational burden associated with active learning on massive datasets. To achieve this goal, we introduce a simple, yet effective method-agnostic framework that outlines how to strategically choose and annotate data points, optimizing the process for efficiency while maintaining model performance. Through case studies, we demonstrate the effectiveness of our proposed method in reducing computational costs while maintaining or, in some cases, even surpassing baseline model outcomes. Code is available at https://github.com/aimotive/Compute-Efficient-Active-Learning |
Gábor Németh · Tamás Matuszka 🔗 |
-
|
CircuitVAE: Efficient and Scalable Latent Circuit Optimization
(
Poster
)
>
link
Automatically designing fast and space-efficient digital circuits is challenging because circuits are discrete, must exactly implement the desired logic, and are costly to simulate. We address these challenges with CircuitVAE, a search algorithm that embeds computation graphs in a continuous space and optimizes a learned surrogate of physical simulation by gradient descent. By carefully controlling overfitting of the simulation surrogate and ensuring diverse exploration, our algorithm is highly sample-efficient, yet gracefully scales to large problem instances and high sample budgets. We test CircuitVAE by designing binary adders across a large range of sizes, IO timing constraints, and sample budgets. Our method excels at designing large circuits, where other algorithms struggle: compared to reinforcement learning and genetic algorithms, CircuitVAE typically finds 64-bit adders which are smaller and faster using less than half the sample budget. We also find CircuitVAE can design state-of-the-art adders in a real-world chip, demonstrating that our method can outperform commercial tools in a realistic setting. |
Jialin Song · Aidan Swope · Robert Kirby · Rajarshi Roy · Saad Godil · Jonathan Raiman · Bryan Catanzaro 🔗 |
-
|
Unleashing the Autoconversion Rates Forecasting: Evidential Regression from Satellite Data
(
Poster
)
>
link
High-resolution simulations such as the ICOsahedral Non-hydrostatic Large-Eddy Model (ICON-LEM) can be used to understand the interactions between aerosols, clouds, and precipitation processes that currently represent the largest source of uncertainty involved in determining the radiative forcing of climate change. Nevertheless, due to the exceptionally high computing cost required, this simulation-based approach can only be employed for a short period of time within a limited area. Despite the fact that machine learning can solve this problem, the related model uncertainties may make it less reliable. To address this, we developed a neural network (NN) model powered with evidential learning to assess the data and model uncertainties applied to satellite observation data. Our study focuses on estimating the rate at which small droplets (cloud droplets) collide and coalesce to become larger droplets (raindrops) – autoconversion rates -- since this is one of the key processes in the precipitation formation of liquid clouds, hence crucial to better understanding cloud responses to anthropogenic aerosols. The results of estimating the autoconversion rates demonstrate that the model performs reasonably well, with the inclusion of both aleatoric and epistemic uncertainty estimation, which improves the credibility of the model and provides useful insights for future improvement. |
Maria Carolina Novitasari · Johannes Quaas · Miguel Rodrigues 🔗 |
-
|
Towards Scalable Identification of Brick Kilns from Satellite Imagery with Active Learning
(
Poster
)
>
link
Air pollution is a leading cause of death globally, especially in south-east Asia. Brick production contributes significantly to air pollution. However, unlike other sources such as power plants, brick production is unregulated and thus hard to monitor. Traditional survey-based methods for kiln identification are time and resource-intensive. Similarly, it is time-consuming for air quality experts to annotate satellite imagery manually. Recently, computer vision machine learning models have helped reduce labeling costs, but they need sufficiently large labeled imagery. In this paper, we propose scalable methods using active learning to accurately detect brick kilns with minimal manual labeling effort. Through this work, we have identified more than 700 new brick kilns across the Indo-Gangetic region: a highly populous and polluted region spanning 0.4 million square kilometers in India. In addition, we have deployed our model as a web application for automatically identifying brick kilns given a specific area by the user. |
Aditi Agarwal · Suraj Jaiswal · Madhav Kanda · Dhruv Patel · Rishabh Mondal · Vannsh Jani · Zeel Bharatkumar Patel · Nipun Batra · Sarath Guttikunda 🔗 |
-
|
Improved Bounds for Agnostic Active Learning of Single Index Models
(
Poster
)
>
link
We study methods for actively learning single index models of the form $F({\boldsymbol x}) = f(\langle {\boldsymbol w}, {\boldsymbol x}\rangle)$, where $f:\mathbb{R} \to \mathbb{R}$ and ${\boldsymbol w} \in \mathbb{R}^d$. Such functions are important in scientific computing, where they are used to construct surrogate models for partial differential equations (PDEs) and to approximate high-dimensional Quantities of Interest (QoIs). In these applications, collecting function samples requires solving a partial differential equation, so sample-efficient active learning methods translate to reduced computational cost.Our work provides two main results. First, when $f$ is known and Lipschitz, we show that $\tilde{O}(d)$ samples collected via \emph{statistical leverage score sampling} are sufficient to find an optimal single index model for a given target function, even in the challenging and practically important agnostic (adversarial noise) setting. This result is optimal up to logarithmic factors and improves quadratically on a recent $\tilde{O}(d^{2})$ bound of Gajjar et al. (2023). Second, we show that $\tilde{O}(d^{3/2})$ samples suffice in the more difficult non-parametric setting when $f$ is \emph{unknown}, which is the also best result known in this general setting.
|
Aarshvi Gajjar · Xingyu Xu · Christopher Musco · Chinmay Hegde 🔗 |
-
|
Practical Path-based Bayesian Optimization
(
Poster
)
>
link
There has been a surge in interest in data-driven experimental design with applications to chemical engineering and drug manufacturing. Bayesian optimization (BO) has proven to be adaptable to such cases, since we can model the reactions of interest as expensive black-box functions. Sometimes, the cost of this black-box functions can be separated into two parts: (a) the cost of the experiment itself, and (b) the cost of changing the input parameters. In this short paper, we extend the SnAKe algorithm to deal with both types of costs simultaneously. We further propose extensions to the case of a maximum allowable input change, as well as to the multi-objective setting. |
Jose Pablo Folch · James Odgers · Shiqiang Zhang · Robert Lee · Behrang Shafei · David Walz · Calvin Tsay · Mark van der Wilk · Ruth Misener 🔗 |
-
|
Data-driven Prior Learning for Bayesian Optimisation
(
Poster
)
>
link
Transfer learning for Bayesian optimisation has generally assumed a strong similarity between optimisation tasks, with at least a subset sharing similar optima. This assumption can reduce computational costs, but it is violated in a wide range of optimisation problems where transfer learning may nonetheless be useful. We replace this assumption with a weaker one only requiring the shape of the optimisation landscape to be similar, and analyse the recent method Prior Learning for Bayesian Optimisation — PLeBO — in this setting. By learning priors for the hyperparameters of the Gaussian process surrogate model we can better approximate the underlying function, especially for few function evaluations. We validate the learned priors and compare to a breadth of transfer learning approaches, using synthetic data and a recent air pollution optimisation problem as benchmarks. We show that PLeBO and prior transfer more generally find good inputs in fewer evaluations. |
Sigrid Passano Hellan · Christopher G Lucas · Nigel Goddard 🔗 |
-
|
Model Robustness and Active Learning with Missing-Not-At-Random Outcomes
(
Poster
)
>
link
We consider prediction problems where outcomes in the training data are missing not at random (MNAR). MNAR outcomes can induce arbitrary levels of bias in downstream prediction tasks. To counteract this bias, one may (1) incorporate additional information that is external to the data, or (2) collect additional data under a different policy from the policy that generated the original dataset. For (1), we consider making models robust to MNAR via distributionally robust optimization. For (2), we develop an active learning approach in which the model training procedure and the acquisition function are attuned to the MNAR setting. Experiments demonstrate the benefits of this approach over standard active and passive learning approaches. |
Alan Mishler · Mohsen Ghassemi · Alec Koppel · Sumitra Ganesh 🔗 |
-
|
Robust Fitted-Q-Evaluation and Iteration under Sequentially Exogenous Unobserved Confounders
(
Poster
)
>
link
Offline reinforcement learning is important in domains such as medicine, economics, and e-commerce where online experimentation is costly, dangerous or unethical, and where the true model is unknown. We study robust policy evaluation and policy optimization in the presence of sequentially-exogenous unobserved confounders under a sensitivity model. We propose and analyze orthogonalized robust fitted-Q-iteration that uses closed-form solutions of the robust Bellman operator to derive a loss minimization problem for the robust Q function, and adds a bias-correction to quantile estimation. Our algorithm enjoys the computational ease of fitted-Q-iteration and statistical improvements (reduced dependence on quantile estimation error) from orthogonalization. We provide sample complexity bounds, insights, and show effectiveness both in simulations and on real-world longitudinal healthcare data of treating sepsis. In particular, our model of sequential unobserved confounders yields an online Markov decision process, rather than partially observed Markov decision process: we illustrate how this can enable warm-starting optimistic reinforcement learning algorithms with valid robust bounds from observational data. |
David Bruns-Smith · Angela Zhou 🔗 |
-
|
BoChemian: Large Language Model Embeddings for Bayesian Optimization of Chemical Reactions
(
Poster
)
>
link
This paper explores the integration of Large Language Models (LLM) embeddings with Bayesian Op-timization (BO) in the domain of chemical reaction optimization with the showcase studyon Buchwald-Hartwig reactions. By leveraging llms, we can transform textual chemi-cal procedures into an informative feature space suitable for Bayesian optimization. Our findings show thateven out-of-the-box open-source LLMs can map chemical reactions for optimization tasks,highlighting their latent specialized knowledge. The results motivate the considerationof further model specialization through adaptive fine-tuning within the bo framework foron-the-fly optimization. This work serves as a foundational step toward a unified com-putational framework that synergizes textual chemical descriptions with machine-drivenoptimization, aiming for more efficient and accessible chemical research.The code is available at: https://github.com/schwallergroup/bochemian. |
Bojana Rankovic · Philippe Schwaller 🔗 |
-
|
\textit{Less But Better}\\ Towards better \textit{AQ} Monitoring by learning \\ Inducing Points for Multi-Task Gaussian Processes
(
Poster
)
>
link
Air pollution is a pressing global issue affecting both human health and environmental sustainability. The high financial burden of conventional Air Quality (AQ) monitoring stations and their sparse spatial distribution necessitate advanced inferencing techniques for effective regulation and public health policies. We introduce a comprehensive framework employing Variational Multi-Output Gaussian Processes (VMOGP) with a Spectral Mixture (SM) kernel designed to model and predict multiple AQ indicators, particularly $PM_{2.5}$ and Carbon Monoxide ($CO$). Our method unifies the strengths of Multi-Output Gaussian Processes (MOGPs) and Variational Multi-Task Gaussian Processes (VMTGP) to capture intricate spatio-temporal correlations among air pollutants, thus delivering enhanced robustness and accuracy over Single-Output Gaussian Processes (SOGPs) and state-of-the-art neural attention-based methods. Importantly, by analyzing the variational distribution of auxiliary inducing points, we identify high-information geographical locales for optimized AQ monitoring frameworks. Through extensive empirical evaluations, we demonstrate superior performance in both accuracy and uncertainty quantification. Our methodology promises significant implications for urban planning, adaptive station placement, and public health policy formulation.
|
Progyan Das · Mihir Agarwal 🔗 |
-
|
ActSort: An active-learning accelerated cell sorting algorithm for large-scale calcium imaging datasets
(
Poster
)
>
link
Due to rapid progress in optical imaging technologies, contemporary neural calcium imaging studies can monitor the dynamics of 10,000 or more neurons at once in the brains of awake behaving mammals. After automated extraction of the neurons' putative locations, a typical experiment involves extensive human labor to cull false-positive cells from the data, a process called \emph{cell sorting.} Efforts to automate cell sorting via the use of trained models either employ pre-trained, suboptimal classifiers or require reduced but still substantial human labor to train dataset-specific classifiers. In this workshop paper, we introduce an active-learning accelerated cell-sorting paradigm, termed ActSort, which establishes an online feedback loop between the human annotator and the cell classifier. To test this paradigm, we designed a first-of-a-kind benchmark by curating large-scale calcium imaging datasets from 5 mice, with approximately 40,000 cell candidates in total. Each movie was annotated by 4 (out of 6 total) human annotators, yielding about 160,000 total annotations. With this approach, we tested two active learning strategies, discriminative active learning (DAL) and confidence-based active learning (CAL). To create a baseline representing the traditional strategy, we performed random and first-to-last annotations, in which cells are annotated in either a random order or the order they are received from the cell-extraction algorithm. Our analysis revealed that, even when using the active learning-derived results of $<5$% of the human-annotated cells, CAL surpassed human performance levels in both precision and recall. In comparison, the first-to-last strategy required $80$\% of the cells to be annotated to achieve the same mark. By decreasing the human labor needed from hours to minutes while also enabling more accurate predictions than a typical human annotator, ActSort overcomes a bottleneck in neuroscience research and enables rapid pre-processing of large-scale brain-imaging datasets.
|
Hakki Akengin · Mehmet Aslihak · Yiqi Jiang · Christopher Miranda · Marta Pozo · Yang Li · Oscar Hernandez · Hakan Inan · Fatih Dinc · Mark Schnitzer 🔗 |
-
|
Actively learning a Bayesian matrix fusion model with deep side information
(
Poster
)
>
link
High-dimensional deep neural network representations of images and concepts can bealigned to predict human annotations of diverse stimuli. However, such alignment requiresthe costly collection of behavioral responses, such that, in practice, the deep-featurespaces are only ever sparsely sampled. Here, we propose an active learning approach toadaptively sample experimental stimuli to efficiently learn a Bayesian matrix factorizationmodel with deep side information. We observe a significant efficiency gain over a passivebaseline. Furthermore, with a sequential batched sampling strategy, the algorithm is applicablenot only to small datasets collected from traditional laboratory experiments butalso to settings where large-scale crowdsourced data collection is needed to accurately alignthe high-dimensional deep feature representations derived from pre-trained networks. Thisprovides cost-effective solutions for collecting and generating quality-assured predictions inlarge-scale behavioral and cognitive studies. |
Yangyang Yu · Jordan Suchow 🔗 |
-
|
Agile Modeling: From Concept to Classifier in Minutes
(
Poster
)
>
link
The application of computer vision methods to nuanced, subjective concepts is growing. While crowdsourcing has served the vision community well for most objective tasks (such as labeling a "zebra"), it now falters on tasks where there is substantial subjectivity in the concept (such as identifying "gourmet tuna"). However, empowering any user to develop a classifier for their concept is technically difficult: users are neither machine learning experts nor have the patience to label thousands of examples. In reaction, we introduce the problem of Agile Modeling: the process of turning any subjective visual concept into a computer vision model through real-time user-in-the-loop interactions. We instantiate an Agile Modeling prototype for image classification and show through a user study (N=14) that users can create classifiers with minimal effort in under 30 minutes. We compare this user driven process with the traditional crowdsourcing paradigm and find that the crowd's notion often differs from that of the user's, especially as the concepts become more subjective. Finally, we scale our experiments with simulations of users training classifiers for ImageNet21k categories to further demonstrate the efficacy of the approach. |
Otilia Stretcu · Edward Vendrow · Kenji Hata · Krishnamurthy Viswanathan · Vittorio Ferrari · Sasan Tavakkol · Wenlei Zhou · Aditya Avinash · Enming Luo · Neil Alldrin · Mohammadhossein Bateni · Gabriel Berger · Andrew Bunner · Chun-Ta Lu · Javier Rey · Giulia DeSalvo · Ranjay Krishna · Ariel Fuxman
|
-
|
Learning Models and Evaluating Policies with Offline Off-Policy Data under Partial Observability
(
Poster
)
>
link
Models in reinforcement learning are often estimated from offline data, which in many real-world scenarios is subject to partial observability.In this work, we study the challenges that emerge from using models estimated from partially-observable offline data for policy evaluation.Notably, these models must be defined in conjunction with the data-collecting policy.To address this issue, we introduce a method for model estimation thatincorporates importance weighting in the model learning process.The off-policy samples are reweighted to be reflective of their probabilities under a different policy, such that the resultant model is a consistent estimator of the off-policy model and provides consistent off-policy estimates of the expected return.This is a crucial step towards the reliable and responsible use of models learned under partial observability, particularly in scenarios where inaccurate policy evaluation can have catastrophic consequences.We empirically demonstrate the efficacy of our method and its resilience to common approximations such as weight clipping on a range of domains with diverse types of partial observability. |
Shreyas Chaudhari · Philip Thomas · Bruno C. da Silva 🔗 |
-
|
LabelBench: A Comprehensive Framework for Benchmarking Adaptive Label-Efficient Learning
(
Poster
)
>
link
Labeled data are critical to modern machine learning applications, but obtaining labels can be expensive. To mitigate this cost, machine learning methods, such as transfer learning, semi-supervised learning and active learning, aim to be $\text{\textit{label-efficient}}$: achieving high predictive performance from relatively few labeled examples. While obtaining the best label-efficiency in practice often requires combinations of these techniques, existing benchmark and evaluation frameworks do not capture a concerted combination of all such techniques. This paper addresses this deficiency by introducing LabelBench, a new computationally-efficient framework for joint evaluation of multiple label-efficient learning techniques. As an application of LabelBench, we introduce a novel benchmark of state-of-the-art active learning methods in combination with semi-supervised learning for fine-tuning pretrained vision transformers. Our benchmark demonstrates significantly better label-efficiencies than previously reported in active learning. LabelBench's modular codebase is open-sourced for the broader community to contribute label-efficient learning methods and benchmarks. The repository can be found at: https://github.com/EfficientTraining/LabelBench.
|
Jifan Zhang · Yifang Chen · Gregory Canal · Stephen Mussmann · Yinglun Zhu · Simon Du · Kevin Jamieson · Robert Nowak 🔗 |
-
|
REDUCR: Robust Data Downsampling Using Class Priority Reweighting
(
Poster
)
>
link
Modern machine learning models are becoming increasingly expensive to train for real-world image and text classification tasks, where massive web-scale data is collected in a streaming fashion. To reduce the training cost, online batch selection techniques have been developed to choose the most informative datapoints. However, these techniques can suffer from poor worst-class generalization performance due to class imbalance and distributional shifts. This work introduces REDUCR, a robust and efficient data downsampling method that uses class priority reweighting. REDUCR reduces the training data while preserving worst-class generalization performance. REDUCR assigns priority weights to datapoints in a class-aware manner using an online learning algorithm. We demonstrate the data efficiency and robust performance of REDUCR on vision and text classification tasks. On web-scraped datasets with imbalanced class distributions, REDUCR achieves significant test accuracy boosts for the worst-performing class (but also on average), surpassing state-of-the-art methods by around 14%. |
William Bankes · George Hughes · Ilija Bogunovic · Zi Wang 🔗 |
-
|
Local Acquisition Function for Active Level Set Estimation
(
Poster
)
>
link
In this paper, we propose a new acquisition function based on local search for active super-level set estimation. Conventional acquisition functions for level set estimation problems are considered to struggle with problems where the threshold is high, and many points in the upper-level set have function values close to the threshold. The proposed method addresses this issue by effectively switching between two acquisition functions: one rapidly finds local level set and the other performs global exploration. The effectiveness of the proposed method is evaluated through experiments with synthetic and real-world datasets. |
Yuta Kokubun · Kota Matsui · Kentaro Kutsukake · Wataru Kumagai · Takafumi Kanamori 🔗 |
-
|
Improving class and group imbalanced classification with uncertainty-based active learning
(
Poster
)
>
link
Recent experimental and theoretical analyses have revealed thatuncertainty-based active learning algorithms (U-AL) are often not able toimprove the average accuracy compared to even the simple baseline of passivelearning (PL). However, we show in this work that U-AL is a competitivemethod in problems with severe data imbalance, when instead of the\emph{average} accuracy, the focus is the \emph{worst-subpopulation} accuracy.We show in extensive experiments that U-AL outperforms algorithms thatexplicitly aim to improve worst-subpopulation performance such as reweighting.We provide insights that explain the good performance of U-AL and show atheoretical result that is supported by our experimental observations. |
Alexandru Tifrea · John Hill · Fanny Yang 🔗 |