Periodic Reporting for period 1 - ixAutoML (Interactive and Explainable Human-Centered AutoML)
Reporting period: 2022-12-01 to 2025-05-31
A key methodological innovation is the development of symbolic explanation techniques for hyperparameter optimization based on performance data collected via Bayesian Optimization, making complex AutoML decisions comprehensible for users. In addition, we developed a new variant of Bayesian Optimization (BO) that re-interprets the Bayesian Execution Algorithm (BAX) to obtain high-quality partial dependence plots (PDPs). By interleaving BAX with traditional BO, we can show that the optimization performance does not decrease while obtaining much better PDPs.
We are the first to propose a systematic approach for studying the hyperparameter landscapes of reinforcement learning algorithms dynamically during policy training. Since at each decision point, we could spawn a large number of new policy trainings, we are facing a combinatorial problem. We propose to store checkpoints and continue policy learning from there until convergence. By choosing the best-performing policies out of all hyperparameter configurations, we can go back to a checkpoint and spawn new hyperparameter configuration runs from there. This technique is overall very efficient and allows a deeper understanding of the landscapes.
To improve the efficiency, robustness, and soundness of AutoRL systems, we propose a recipe for efficient RL research, incl. training test splits on different levels, designing configuration spaces, deciding on the correct optimization approach, settling on an important cost metric, running HPO for several seeds and evaluate on new seeds. Although, maybe surprisingly trivial, we were able to show that results in comparing RL and AutoRL algorithms improved in quality substantially.
To enable interaction between users and AutoML packages, we propose to bias hyperparameter optimization towards users' expectations (e.g. preferred learning rates). In contrast to previous approaches, we proposed a simple-to-implement generic approach based on re-weighting the acquisition function w.r.t. the user prior that can outperform complex previous approaches and keeps the theoretical guarantees for BO.
Furthermore, we have established an interactive framework for hyperparameter optimization (HPO), leveraging preference learning to decide between examples of Pareto fronts. The learned preferences can then be used to guide HPO towards preferred shapes of Pareto Fronts. This allows human users to actively guide the HPO process in multi-objective settings, creating a seamless integration of human intuition and algorithmic efficiency.
Static AutoML methods are widely used but often perceived as black-box systems, which can limit their adoption. Our research has made strides in enhancing the transparency and explainability of these approaches. One notable achievement is the development of frameworks such as Enhancing Explainability of Hyperparameter Optimization via Bayesian Algorithm Execution. This work guides the AutoML optimization process so that more reliable and trustworthy explanations can be derived post hoc without investigating further configurations of the ML system.
Additionally, we introduced Symbolic Explanations for Hyperparameter Optimization, a method that generates interpretable, symbolic representations of relationships between hyperparameters and their effects on model performance. This approach empowers users to comprehend the rationale behind AutoML-generated configurations, bridging the gap between complex optimization algorithms and machine learning expertise. We believe that this will make AutoML also more interesting as a tool to derive theories in ML learning processes.
Explaining Dynamic AutoML Policies
Dynamic AutoML systems, including those based on RL, adapt their policies in real time based on evolving data and changing objectives. These systems pose unique challenges regarding explainability, as their decisions are inherently more complex and context-dependent. Our work on AutoRL Hyperparameter Landscapes provides fundamental insights into the structure and behavior of RL hyperparameter spaces, enabling more informed tuning strategies.
In Hyperparameters in Reinforcement Learning and How to Tune Them, we were the first to show how difficult the tuning of RL algorithms is. We synthesized practical guidelines for optimizing RL hyperparameters, offering practitioners a robust framework for improving model performance.
A Human-Centered Interactive AutoML Paradigm
A key aspect of our vision is to empower users to bring their knowledge into the optimization loop. Approaches like πBO: Augmenting Acquisition Functions with User Beliefs for Bayesian Optimization demonstrate how user expertise on promising hyperparameter configurations can be seamlessly integrated into the optimization processes, improving performance and the perceived reliability of the system. This work is complemented by PriorBand: Practical Hyperparameter Optimization in the Age of Deep Learning, introducing a novel approach leveraging prior knowledge to enhance optimization efficiency for expensive deep neural networks.
We also developed Interactive Hyperparameter Optimization in Multi-Objective Problems via Preference Learning, a framework that learns user preferences to guide optimization processes in complex, multi-objective settings. This is particularly important in cases where accuracy is just as important as, e.g. fairness, energy consumption, or latency. Our approach enables users to guide hyperparameter optimization in these multi-objective settings without understanding complex multi-objective metrics by ranking a few different possible Pareto fronts.