www.cs.uwyo.edu
Open in
urlscan Pro
129.72.216.15
Public Scan
Submitted URL: http://www.cs.uwyo.edu///~larsko///?page\=publications
Effective URL: https://www.cs.uwyo.edu///~larsko///?page\=publications
Submission: On October 10 via api from US — Scanned from DE
Effective URL: https://www.cs.uwyo.edu///~larsko///?page\=publications
Submission: On October 10 via api from US — Scanned from DE
Form analysis
0 forms found in the DOMText Content
Home NewsPublicationsTeachingSoftwareAwardsOther LARS KOTTHOFF TEMPLETON ASSOCIATE PROFESSOR DERECHO PROFESSOR PRESIDENTIAL FACULTY FELLOW larsko@uwyo.edu EERB 422b Department of Electrical Engineering and Computer Science School of Computing University of Wyoming 1000 E University Ave Laramie, WY 82071-2000 My research combines artificial intelligence and machine learning to build robust systems with state-of-the-art performance. I develop techniques to induce models of how algorithms for solving computationally difficult problems behave in practice. Such models allow to select the best algorithm and choose the best parameter configuration for solving a given problem. I lead the Meta-Algorithmics, Learning and Large-scale Empirical Testing (MALLET) lab and direct the Artificially Intelligent Manufacturing center (AIM) at the University of Wyoming. More broadly, I am interested in innovative ways of modelling and solving challenging problems and applying such approaches to the real world. Part of this is making cutting edge research available to and usable by non-experts. Machine learning often plays a crucial role in this, and I am also working on making machine learning more accessible and easier to use. Interested in coming to beautiful Wyoming and joining MALLET? Please drop me an email or, if you are already here, come by my office. I also have Master's projects and projects for undergraduates seeking research experience available. NEWS * I was interviewed on Automated Machine Learning by Built In. * Had a great time visiting Colorado State University! Thanks Darrell for hosting me! * Good to be traveling again! This trip: Lorentz Center, TU Eindhoven, University Hannover, University Leipzig. * Two papers accepted at the first AutoML conference, for which I’m also a senior area chair. * Our paper on automating the production of laser-induced graphene was accepted at PAIS 2022. See all Gave an invited presentation on our work on using Bayesian Optimization in Materials Science at the 2022 SIAM conference on Uncertainty Quantification. My interdisciplinary research is featured in the Fall 2021 issue of UWyo Magazine. Congratulations to my colleague Mike Borowczak for being awarded an NSF grant to create Research Experiences for Teachers (award page, UW press release). Oh and I am part of this as well. UW has issued a press release on our work contributing to the AI Index 2021. Congratulations to Austin, who did all the hard work on this. I will be on the panel for the AAAI 2021 workshop on Meta-Learning. One of my students, Damir Pulatov, was highlighted by our research computing center for his computational work. Congratulations Damir! I am mentoring Google Summer of Code student Akshit Achara, who is creating a MiniZinc interface for R. Check out his code here. I am tutorial chair for the CP 2020 conference and co-organizer for the Fourth Workshop on Progress Towards the Holy Grail. We have been awarded a $750,000 grant from NASA EPSCoR for research into manufacturing advanced electronic devices (with Patrick Johnson and DP Aidhy). More in the university press release. The grant was also covered by NPR. I have been awarded funding from Microsoft (with Todd Schoborg and Jay Gatlin in Molecular Biology and Brant Schumaker in Veterinary Sciences) for biomedical and wildlife imaging. See the university press release. Extremely honored to accept the Open Source Machine Learning Award on behalf of the mlr team at ODSC West 2019 (article). Had a great time at the workshop on measurement in AI policy at Stanford. I was interviewed on automated machine learning AutoML on a German AI podcast. Find it here (in German). I visited NASA Ames to talk about our work on applying Bayesian optimization to materials. PDF slides The National Institute for Standards and Technology (NIST) lists our mlr package in the US Leadership in AI plan. I visited LIACS and gave a talk on our work on using Bayesian Optimization to optimize graphene production (PDF slides). Had a great time at COSEAL 2019, where I presented our posters on software features for algorithm selection (PDF), interactive visualizations for ASlib (PDF), and Bayesian Optimization for graphene production (PDF). Congratulations to my colleague Mike Borowczak for being awarded an NSF grant for improving CS education in Wyoming (award page). Oh and I am part of this as well. Attended the Materials Science in Space Workshop at the International Space Station Research and Development Conference 2019 in Atlanta. I’m organizing an introduction to data science and machine learning workshop at the end of September. More information here. I will give a tutorial on AI in Materials Science (T32) at IJCAI 2019. Our paper on applying Bayesian Optimization to graphene production has been accepted at the IJCAI 2019 Workshop on Data Science and Optimisation (PDF slides). I am giving a tutorial on automated parameter tuning techniques and applications in engineering at RMACC 2019 (PDF slides). Our book on automated machine learning is now available on Springer’s website. Mentoring a student for a Google Summer of Code project to create visualizations for mlr3. The project page will be here. I was awarded an REU supplement to my NSF grant #1813537 to employ two undergraduate research assistants over the summer. I gave a talk at the University of Warsaw on algorithm selection and configuration. You can find the slides here. We went to visit the Confirm Centre in Ireland. You can find the slides of my talk here. A bunch of students and I had a good time at the AAAI 2019 conference. See the news item on the college website. I gave a talk at NCAR CISL on algorithm selection and configuration (PDF slides, video). Joeran Beel and I are organizing the First Workshop on Algorithm Selection and Meta-Learning in Information Retrieval. I attended Dagstuhl Seminar 18401: “Automating Data Science”. Giving a talk on automatic machine learning at the AutoML workshop at the Pacific Rim Conference on AI 2018 (PDF slides). Our research center on Artificially Intelligent Manufacturing (AIM) was funded by the Engineering Initiative. I gave a talk at the University of St Andrews on algorithm selection and configuration (PDF slides). I gave the invited talk at the IDIR summer workshop on mlr (PDF slides). I gave a talk at the University of Glasgow on algorithm selection and configuration (PDF slides). I presented our paper on the Temporal Shapley Value at IJCAI 2018 (slides). I gave a talk at LIACS on the Shapley Value and its temporal cousin for analysing algorithm performance (PDF slides). I gave a talk at TU Eindhoven on algorithm selection and configuration (PDF slides). My project proposal for more robust performance models has been funded by the NSF ($412,000). I will be at ISMP 2018 in Bordeaux, giving an invited talk on our work on the Shapley value to evalute the contribution of algorithms (PDF slides). Co-organizing the Second Workshop on Progress Towards the Holy Grail, co-located with CP 2018. I have secured €3,000 funding from Artificial Intelligence Journal for the ACP summer school 2018. Our paper “Quantifying Algorithmic Improvements over Time” has been accepted to the IJCAI/ECAI 2018 special track on the evolution of the contours of AI (23% acceptance rate). Giving a talk at Tech Talk Laramie on April 19th 6pm: https://www.meetup.com/TechTalkLaramie/events/sdksvnyxgbzb/ I was awarded a University of Wyoming Global Engagement Office travel grant worth $2000. I am featured on the University of Wyoming new faculty profile for March. Attended the 2018 CRA career mentoring workshop. I’m organizing the ACP summer school 2018. I’m very honored to have been named outstanding PC member at AAAI 2018. Looking forward to AAAI 2018 in New Orleans! The proceedings for the 2017 Algorithm Selection Challenge are online! The results of the 2017 Algorithm Selection Challenge are in. Congratulations to the winner! I’ll be at the Wyoming Global Technology Summit. Looking forward to the COSEAL meeting 2017 in Brussels! Giving an invited talk “Intelligent Constraint Programming: Algorithm Selection for fun and profit” at the Workshop on Progress Towards the Holy Grail (which I am co-organizing), co-located with CP 2017, ICLP 2017, and SAT 2017. I am also on the programme committee for the joint doctoral consortium. Will be at IJCAI 2017. See you there! PUBLICATIONS For citation numbers, please see my Google Scholar page. 2023 * Kotthoff, Lars. “Towards Machine-Generated Algorithms.” In AAAI 2023 Bridge Constraint Programming and Machine Learning, 2023. bibTeX * Wahab, Hud, Lars Kotthoff, and Patrick Johnson. “Optimization of Laser-Induced Graphene Manufacturing.” In AAAI 2023 Bridge AI for Materials Science, 2023. bibTeX * Shoaib, Mirza, Neelesh Sharma, Lars Kotthoff, Marius Lindauer, and Surya Kant. “AutoML: Advanced Tool for Mining Multivariate Plant Traits.” Trends in Plant Science, 2023. https://doi.org/https://doi.org/10.1016/j.tplants.2023.09.008. preprint PDF bibTeX * Iqbal, Md Shahriar, Jianhai Su, Lars Kotthoff, and Pooyan Jamshidi. “FlexiBO: A Decoupled Cost-Aware Multi-Objective Optimization Approach for Deep Neural Networks.” Journal of Artificial Intelligence Research 77 (June 2023): 645–82. preprint PDF bibTeX abstract The design of machine learning systems often requires trading off different objectives, for example, prediction error and energy consumption for deep neural networks (DNNs). Typically, no single design performs well in all objectives; therefore, finding Pareto-optimal designs is of interest. The search for Pareto-optimal designs involves evaluating designs in an iterative process, and the measurements are used to evaluate an acquisition function that guides the search process. However, measuring different objectives incurs different costs. For example, the cost of measuring the prediction error of DNNs is orders of magnitude higher than that of measuring the energy consumption of a pre-trained DNN as it requires re-training the DNN. Current state-of-the-art methods do not consider this difference in objective evaluation cost, potentially incurring expensive evaluations of objective functions in the optimization process. In this paper, we develop a novel decoupled and cost-aware multi-objective optimization algorithm, which we call Flexible Multi-Objective Bayesian Optimization (FlexiBO) to address this issue. For evaluating each design, FlexiBO selects the objective with higher relative gain by weighting the improvement of the hypervolume of the Pareto region with the measurement cost of each objective. This strategy, therefore, balances the expense of collecting new information with the knowledge gained through objective evaluations, preventing FlexiBO from performing expensive measurements for little to no gain. We evaluate FlexiBO on seven state-of-the-art DNNs for image recognition, natural language processing (NLP), and speech-to-text translation. Our results indicate that, given the same total experimental budget, FlexiBO discovers designs with 4.8\% to 12.4\% lower hypervolume error than the best method in state-of-the-art multi-objective optimization. * Kashgarani, Haniye, and Lars Kotthoff. “Automatic Parallel Portfolio Selection.” In 26th European Conference on Artificial Intelligence, 372:1215–22. Frontiers in Artificial Intelligence and Applications. IOS Press, 2023. preprint PDF bibTeX abstract Algorithms to solve hard combinatorial problems often exhibit complementary performance, i.e. where one algorithm fails, another shines. Algorithm portfolios and algorithm selection take advantage of this by running all algorithms in parallel or choosing the best one to run on a problem instance. In this paper, we show that neither of these approaches gives the best possible performance and propose the happy medium of running a subset of all algorithms in parallel. We propose a method to choose this subset automatically for each problem instance, and demonstrate empirical improvements of up to 19\% in terms of runtime, 81\% in terms of misclassification penalty, and 26\% in terms of penalized averaged runtime on scenarios from the ASlib benchmark library. Unlike all other algorithm selection and scheduling approaches in the literature, our performance measures are based on the actual performance for algorithms running in parallel rather than assuming overhead-free parallelization based on sequential performance. Our approach is easy to apply in practice and does not require to solve hard problems to obtain a schedule, unlike other techniques in the literature, while still delivering superior performance. See all 2022 * Bistarelli, Stefano, Lars Kotthoff, Francesco Santini, and Carlo Taticchi. “Summary Report for the Third International Competition on Computational Models of Argumentation.” AI Magazine 42, no. 3 (2022): 70–73. preprint PDF bibTeX abstract The Third International Competition on Computational Models of Argumentation (ICCMA’19) focused on reasoning tasks in abstract argumentation frameworks. Submitted solvers were tested on a selected collection of benchmark instances, including artificially generated argumentation frameworks and some frameworks formalizing real-world problems. This competition introduced two main novelties over the two previous editions: the first one is the use of the Docker platform for packaging the participating solvers into virtual “light” containers; the second novelty consists of a new track for dynamic frameworks. * Pulatov, Damir, Marie Anastacio, Lars Kotthoff, and Holger H. Hoos. “Opening the Black Box: Automated Software Analysis for Algorithm Selection.” In INFORMS Computing Society Conference, 2022. bibTeX abstract Impressive performance improvements have been achieved in many areas of AI by meta-algorithmic techniques, such as automated algorithm selection and configuration. However, existing techniques treat the target algorithms they are applied to as black boxes -- nothing is known about their inner workings. This allows metaalgorithmic techniques to be used broadly, but leaves untapped potential performance improvements enabled by information gained from a deeper analysis of the target algorithms. In this paper, we open the black box without sacrificing universal applicability of meta-algorithmic techniques by automatically analyzing algorithms. We show how to use this information to perform algorithm selection, and demonstrate improved performance compared to previous approaches that treat algorithms as black boxes. * Pulatov, Damir, Marie Anastacio, Lars Kotthoff, and Holger Hoos. “Opening the Black Box: Automated Software Analysis for Algorithm Selection.” In First Conference on Automated Machine Learning (Main Track), 2022. preprint PDF bibTeX abstract Impressive performance improvements have been achieved in many areas of AI by metaalgorithmic techniques, such as automated algorithm selection and configuration. However, existing techniques treat the target algorithms they are applied to as black boxes – nothing is known about their inner workings. This allows meta-algorithmic techniques to be used broadly, but leaves untapped potential performance improvements enabled by information gained from a deeper analysis of the target algorithms. In this paper, we open the black box without sacrificing universal applicability of meta-algorithmic techniques by automatically analyzing algorithms. We show how to use this information to perform algorithm selection, and demonstrate improved performance compared to previous approaches that treat algorithms as black boxes. * Iqbal, Md Shahriar, Jianhai Su, Lars Kotthoff, and Pooyan Jamshidi. “Getting the Best Bang For Your Buck: Choosing What to Evaluate for Faster Bayesian Optimization.” In First Conference on Automated Machine Learning (Late-Breaking Workshop Track), 2022. preprint PDF bibTeX abstract Machine learning system design frequently necessitates balancing multiple objectives, such as prediction error and energy consumption, for deep neural networks (DNNs). Typically, no single design performs well across all objectives; thus, finding Pareto-optimal designs is of interest. Measuring different objectives frequently incurs different costs; for example, measuring the prediction error of DNNs is significantly more expensive than measuring the energy consumption of a pre-trained DNN because it requires re-training the DNN. Current state-of-the-art methods do not account for this difference in objective evaluation cost, potentially wasting costly evaluations of objective functions for little information gain. To address this issue, we propose a novel cost-aware decoupled approach that weights the improvement of the hypervolume of the Pareto region by the measurement cost of each objective. To evaluate our approach, we perform experiments on several machine learning systems deployed on energy constraints environments. * Kotthoff, Lars, Sourin Dey, Jake Heil, Vivek Jain, Todd Muller, Alexander Tyrrell, Hud Wahab, and Patrick Johnson. “Optimizing Laser-Induced Graphene Production.” In 11th Conference on Prestigious Applications of Artificial Intelligence, 31–44, 2022. preprint PDF bibTeX abstract A lot of technological advances depend on next-generation materials, such as graphene, which enables better electronics, to name but one example. Manufacturing such materials is often difficult, in particular, producing graphene at scale is an open problem. We apply state-of-the-art machine learning to optimize the production of laser-induced graphene, an established manufacturing method that has shown great promise. We demonstrate improvements over previous results in terms of the quality of the produced graphene from a variety of different precursor materials. We use Bayesian model-based optimization to quickly improve outcomes based on little initial data and show the robustness of our approach to different experimental conditions, tackling a small-data problem in contrast to the more common big-data applications of machine learning. We analyze the learned surrogate models with respect to the quality of their predictions and learned relationships that may be of interest to domain experts and improve our understanding of the processes governing laser-induced graphene production. * Moosbauer, Julia, Martin Binder, Lennart Schneider, Florian Pfisterer, Marc Becker, Michel Lang, Lars Kotthoff, and Bernd Bischl. “Automated Benchmark-Driven Design and Explanation of Hyperparameter Optimizers.” IEEE Transactions on Evolutionary Computation Special Issue on Benchmarking Sampling-Based Optimization Heuristics: Methodology and Software (BENCH) 26, no. 6 (December 2022): 1336–50. preprint PDF bibTeX abstract Automated hyperparameter optimization (HPO) has gained great popularity and is an important component of most automated machine learning frameworks. However, the process of designing HPO algorithms is still an unsystematic and manual process: New algorithms are often built on top of prior work, where limitations are identified and improvements are proposed. Even though this approach is guided by expert knowledge, it is still somewhat arbitrary. The process rarely allows for gaining a holistic understanding of which algorithmic components drive performance and carries the risk of overlooking good algorithmic design choices. We present a principled approach to automated benchmark-driven algorithm design applied to multi-fidelity HPO (MF-HPO). First, we formalize a rich space of MF-HPO candidates that includes, but is not limited to, common existing HPO algorithms and then present a configurable framework covering this space. To find the best candidate automatically and systematically, we follow a programming-by-optimization approach and search over the space of algorithm candidates via Bayesian optimization. We challenge whether the found design choices are necessary or could be replaced by more naive and simpler ones by performing an ablation analysis. We observe that using a relatively simple configuration (in some ways, simpler than established methods) performs very well as long as some critical configuration parameters are set to the right value. 2021 * Kashgarani, Haniye, and Lars Kotthoff. “Is Algorithm Selection Worth It? Comparing Selecting Single Algorithms and Parallel Execution.” In AAAI Workshop on Meta-Learning and MetaDL Challenge, 140:58–64. Proceedings of Machine Learning Research. PMLR, 2021. preprint PDF bibTeX abstract For many practical problems, there is more than one algorithm or approach to solve them. Such algorithms often have complementary performance – where one fails, another performs well, and vice versa. Per-instance algorithm selection leverages this by employing portfolios of complementary algorithms to solve sets of difficult problems, choosing the most appropriate algorithm for each problem instance. However, this requires complex models to effect this selection and introduces overhead to compute the data needed for those models. On the other hand, even basic hardware is more than capable of running several algorithms in parallel. We investigate the tradeoff between selecting a single algorithm and running multiple in parallel and incurring a slowdown because of contention for shared resources. Our results indicate that algorithm selection is worth it, especially for large portfolios. * Binder, Martin, Florian Pfisterer, Michel Lang, Lennart Schneider, Lars Kotthoff, and Bernd Bischl. “mlr3pipelines - Flexible Machine Learning Pipelines in R.” Journal of Machine Learning Research 22, no. 184 (2021): 1–7. preprint PDF bibTeX abstract Recent years have seen a proliferation of ML frameworks. Such systems make ML accessible to non-experts, especially when combined with powerful parameter tuning and AutoML techniques. Modern, applied ML extends beyond direct learning on clean data, however, and needs an expressive language for the construction of complex ML workflows beyond simple pre- and post-processing. We present mlr3pipelines, an R framework which can be used to define linear and complex non-linear ML workflows as directed acyclic graphs. The framework is part of the mlr3 ecosystem, leveraging convenient resampling, benchmarking, and tuning components. * Kotthoff, Lars, Sourin Dey, Vivek Jain, Alexander Tyrrell, Hud Wahab, and Patrick Johnson. “Modeling and Optimizing Laser-Induced Graphene,” 2021. preprint PDF bibTeX abstract A lot of technological advances depend on next-generation materials, such as graphene, which enables a raft of new applications, for example better electronics. Manufacturing such materials is often difficult; in particular, producing graphene at scale is an open problem. We provide a series of datasets that describe the optimization of the production of laser-induced graphene, an established manufacturing method that has shown great promise. We pose three challenges based on the datasets we provide -- modeling the behavior of laser-induced graphene production with respect to parameters of the production process, transferring models and knowledge between different precursor materials, and optimizing the outcome of the transformation over the space of possible production parameters. We present illustrative results, along with the code used to generate them, as a starting point for interested users. The data we provide represents an important real-world application of machine learning; to the best of our knowledge, no similar datasets are available. * Kotthoff, Lars, Hud Wahab, and Patrick Johnson. “Bayesian Optimization in Materials Science: A Survey,” 2021. preprint PDF bibTeX abstract Bayesian optimization is used in many areas of AI for the optimization of black-box processes and has achieved impressive improvements of the state of the art for a lot of applications. It intelligently explores large and complex design spaces while minimizing the number of evaluations of the expensive underlying process to be optimized. Materials science considers the problem of optimizing materials' properties given a large design space that defines how to synthesize or process them, with evaluations requiring expensive experiments or simulations -- a very similar setting. While Bayesian optimization is also a popular approach to tackle such problems, there is almost no overlap between the two communities that are investigating the same concepts. We present a survey of Bayesian optimization approaches in materials science to increase cross-fertilization and avoid duplication of work. We highlight common challenges and opportunities for joint research efforts. 2020 * Wahab, Hud, Vivek Jain, Alexander Scott Tyrrell, Michael Alan Seas, Lars Kotthoff, and Patrick Alfred Johnson. “Machine-Learning-Assisted Fabrication: Bayesian Optimization of Laser-Induced Graphene Patterning Using in-Situ Raman Analysis.” Carbon 167 (2020): 609–19. https://doi.org/https://doi.org/10.1016/j.carbon.2020.05.087. preprint PDF bibTeX abstract The control of the physical, chemical, and electronic properties of laser-induced graphene (LIG) is crucial in the fabrication of flexible electronic devices. However, the optimization of LIG production is time-consuming and costly. Here, we demonstrate state-of-the-art automated parameter tuning techniques using Bayesian optimization to advance rapid single-step laser patterning and structuring capabilities with a view to fabricate graphene-based electronic devices. In particular, a large search space of parameters for LIG explored efficiently. As a result, high-quality LIG patterns exhibiting high Raman G/D ratios at least a factor of four larger than those found in the literature were achieved within 50 optimization iterations in which the laser power, irradiation time, pressure and type of gas were optimized. Human-interpretable conclusions may be derived from our machine learning model to aid our understanding of the underlying mechanism for substrate-dependent LIG growth, e.g. high-quality graphene patterns are obtained at low and high gas pressures for quartz and polyimide, respectively. Our Bayesian optimization search method allows for an efficient experimental design that is independent of the experience and skills of individual researchers, while reducing experimental time and cost and accelerating materials research. * Bistarelli, Stefano, Lars Kotthoff, Francesco Santini, and Carlo Taticchi. “A First Overview of ICCMA’19.” In Workshop on Advances In Argumentation In Artificial Intelligence 2020 Co-Located with the 19th International Conference of the Italian Association for Artificial Intelligence (AIxIA 2020), Online, November 25-26, 2020, edited by Bettina Fazzinga, Filippo Furfaro, and Francesco Parisi, 2777:90–102. CEUR Workshop Proceedings. CEUR-WS.org, 2020. preprint PDF bibTeX * Pulatov, Damir, and Lars Kotthoff. “Opening the Black Box: Automatically Characterizing Software for Algorithm Selection.” In AAAI Student Abstracts, 2020. bibTeX 2019 * Hutter, Frank, Lars Kotthoff, and Joaquin Vanschoren, eds. Automated Machine Learning: Methods, Systems, Challenges. 1st ed. The Springer Series on Challenges in Machine Learning. Springer, Cham, 2019. preprint PDF bibTeX abstract This open access book presents the first comprehensive overview of general methods in Automated Machine Learning (AutoML), collects descriptions of existing systems based on these methods, and discusses the first series of international challenges of AutoML systems. The recent success of commercial ML applications and the rapid growth of the field has created a high demand for off-the-shelf ML methods that can be used easily and without expert knowledge. However, many of the recent machine learning successes crucially rely on human experts, who manually select appropriate ML architectures (deep learning architectures or more traditional ML workflows) and their hyperparameters. To overcome this problem, the field of AutoML targets a progressive automation of machine learning, based on principles from optimization and machine learning itself. This book serves as a point of entry into this quickly-developing field for researchers and advanced students alike, as well as providing a reference for practitioners aiming to use AutoML in their work. * Schwarz, Hannes, Lars Kotthoff, Holger Hoos, Wolf Fichtner, and Valentin Bertsch. “Improving the Computational Efficiency of Stochastic Programs Using Automated Algorithm Configuration: an Application to Decentralized Energy Systems.” Annals of Operations Research, January 2019. https://doi.org/10.1007/s10479-018-3122-6. preprint PDF bibTeX abstract The optimization of decentralized energy systems is an important practical problem that can be modeled using stochastic programs and solved via their large-scale, deterministic-equivalent formulations. Unfortunately, using this approach, even when leveraging a high degree of parallelism on large high-performance computing systems, finding close-to-optimal solutions still requires substantial computational effort. In this work, we present a procedure to reduce this computational effort substantially, using a state-of-the-art automated algorithm configuration method. We apply this procedure to a well-known example of a residential quarter with photovoltaic systems and storage units, modeled as a two-stage stochastic mixed-integer linear program. We demonstrate that the computing time and costs can be substantially reduced by up to 50\% by use of our procedure. Our methodology can be applied to other, similarly-modeled energy systems. * Lindauer, Marius, Jan N. van Rijn, and Lars Kotthoff. “The Algorithm Selection Competitions 2015 and 2017.” Artificial Intelligence 272 (2019): 86–100. https://doi.org/https://doi.org/10.1016/j.artint.2018.10.004. preprint PDF bibTeX abstract The algorithm selection problem is to choose the most suitable algorithm for solving a given problem instance. It leverages the complementarity between different approaches that is present in many areas of AI. We report on the state of the art in algorithm selection, as defined by the Algorithm Selection competitions in 2015 and 2017. The results of these competitions show how the state of the art improved over the years. We show that although performance in some cases is very good, there is still room for improvement in other cases. Finally, we provide insights into why some scenarios are hard, and pose challenges to the community on how to advance the current state of the art. * Iqbal, Md Shahriar, Lars Kotthoff, and Pooyan Jamshidi. “Transfer Learning for Performance Modeling of Deep Neural Network Systems.” In USENIX Conference on Operational Machine Learning. Santa Clara, CA: USENIX Association, 2019. preprint PDF bibTeX abstract Modern deep neural network (DNN) systems are highly configurable with large a number of options that significantly affect their non-functional behavior, for example inference time and energy consumption. Performance models allow to understand and predict the effects of such configuration options on system behavior, but are costly to build because of large configuration spaces. Performance models from one environment cannot be transferred directly to another; usually models are rebuilt from scratch for different environments, for example different hardware. Recently, transfer learning methods have been applied to reuse knowledge from performance models trained in one environment in another. In this paper, we perform an empirical study to understand the effectiveness of different transfer learning strategies for building performance models of DNN systems. Our results show that transferring information on the most influential configuration options and their interactions is an effective way of reducing the cost to build performance models in new environments. * Beel, Joeran, and Lars Kotthoff. “Proposal for the 1st Interdisciplinary Workshop on Algorithm Selection and Meta-Learning in Information Retrieval (AMIR).” In Advances in Information Retrieval, edited by Leif Azzopardi, Benno Stein, Norbert Fuhr, Philipp Mayr, Claudia Hauff, and Djoerd Hiemstra, 383–88. Cham: Springer International Publishing, 2019. preprint PDF bibTeX abstract The algorithm selection problem describes the challenge of identifying the best algorithm for a given problem space. In many domains, particularly artificial intelligence, the algorithm selection problem is well studied, and various approaches and tools exist to tackle it in practice. Especially through meta-learning impressive performance improvements have been achieved. The information retrieval (IR) community, however, has paid little attention to the algorithm selection problem, although the problem is highly relevant in information retrieval. This workshop will bring together researchers from the fields of algorithm selection and meta-learning as well as information retrieval. We aim to raise the awareness in the IR community of the algorithm selection problem; identify the potential for automatic algorithm selection in information retrieval; and explore possible solutions for this context. In particular, we will explore to what extent existing solutions to the algorithm selection problem from other domains can be applied in information retrieval, and also how techniques from IR can be used for automated algorithm selection and meta-learning. * Kotthoff, Lars, Vivek Jain, Alexander Tyrrell, Hud Wahab, and Patrick Johnson. “AI for Materials Science: Tuning Laser-Induced Graphene Production.” In Data Science Meets Optimisation Workshop at IJCAI 2019, 2019. preprint PDF bibTeX abstract AI has advanced the state of the art in many application domains, including ones not ordinarily associated with computer science. We present an application of automated parameter tuning to materials science, in particular, we use surrogate models for automated parameter tuning to optimize the fabrication of laser-induced graphene. This process allows to create microscopic conductive lines in thin layers of insulating material, enabling the development of next-generation nano-circuits. Optimizing the parameters that control the laser irradiation process is crucial to creating high-quality graphene that is suitable for this purpose. Through the application of state-of-the-art parameter tuning techniques, we are able to achieve improvements of up to a factor of two compared to existing approaches in the literature and to what human experts are able to achieve. Our results are reproducible across different experimental specimen and the deployed application can be used by domain scientists without a background in AI or machine learning. * Lang, Michel, Martin Binder, Jakob Richter, Patrick Schratz, Florian Pfisterer, Stefan Coors, Quay Au, Giuseppe Casalicchio, Lars Kotthoff, and Bernd Bischl. “mlr3: A Modern Object-Oriented Machine Learning Framework in R.” Journal of Open Source Software 4, no. 44 (2019). preprint PDF bibTeX abstract The R (R Core Team, 2019) package mlr3 and its associated ecosystem of extension packages implements a powerful, object-oriented and extensible framework for machine learning (ML) in R. It provides a unified interface to many learning algorithms available on CRAN, augmenting them with model-agnostic general-purpose functionality that is needed in every ML project, for example train-test-evaluation, resampling, preprocessing, hyperparameter tuning, nested resampling, and visualization of results from ML experiments. The package is a complete reimplementation of the mlr (Bischl et al., 2016) package that leverages many years of experience and learned best practices to provide a state-of-the-art system that is powerful, flexible, extensible, and maintainable. We target both practitioners who want to quickly apply ML algorithms to their problems and researchers who want to implement, benchmark, and compare their new methods in a structured environment. mlr3 is suitable for short scripts that test an idea, for complex multi-stage experiments with advanced functionality that use a broad range of ML functionality, as a foundation to implement new ML (meta-)algorithms (for example AutoML systems), and everything in between. Functional correctness is ensured through extensive unit and integration tests. * Wahab, Hud, Alexander Tyrrell, Vivek Jain, Lars Kotthoff, and Patrick Johnson. “Model-Based Optimization of Laser-Reduced Graphene Using in-Situ Raman Analysis.” In Materials Research Society Fall Symposium, 2019. preprint PDF bibTeX * Hankins, Sarah, Lars Kotthoff, and Ray S. Fertig. “Bio-like Composite Microstructure Designs for Enhanced Damage Tolerance via Machine Learning.” In American Society for Composites 34th Annual Technical Conference, 2019. preprint PDF bibTeX abstract Variations of unique and tailored composite microstructures have been observed in nature and have served as templates for the development of new synthetic materials. Microstructures are studied in fish scales for their penetration resistance, in spider webs for energy absorption, and in seashells and bone for their strength and toughness. However, it has proven difficult to reproduce the properties found in natural materials, due to the interaction between the intricate structures at different length scales. Rather than attempting to replicate these materials (biomimetics), the focus of this work is to use a bio-inspired pattern generation algorithm to search for new topologies that outperform traditional composite structures due to their naturelike design. The bio-inspired pattern generation algorithm employed in this research is known as the Gray-Scott model. This model was selected due to its unique ability to manufacture patterns that propagate with time, allowing the reinforcement volume fraction of the composite structure to be controlled. The model is capable of producing Turing patterns, propagating wave fronts, homogeneous oscillations, and chaos. Traditionally, Turing models have been primarily studied for their applications in morphogenesis and pattern development. However, this research extends the application of the Gray-Scott model by investigating the patterns as physical load bearing structures. A methodology was developed by which the patterns can be converted to structures, analyzed for a desired mechanical property, and optimized via Bayesian machine learning algorithms that yield an improvement of the average quality of structures produced by almost a factor of 10. * Pulatov, Damir, and Lars Kotthoff. “Utilizing Software Features for Algorithm Selection.” In COSEAL Workshop 2019 (Poster Presentation), 2019. bibTeX * Chawla, Katherine, and Lars Kotthoff. “Interactive Visualizations for ASlib.net.” In COSEAL Workshop 2019 (Poster Presentation), 2019. bibTeX * Kotthoff, Lars, Vivek Jain, Alexander Tyrrell, Hud Wahab, and Patrick Johnson. “AI for Materials Science: Tuning Laser-Induced Graphene Production.” In COSEAL Workshop 2019 (Spotlight Talk and Poster Presentation), 2019. bibTeX 2018 * Kotthoff, Lars, Alexandre Fréchette, Tomasz P. Michalak, Talal Rahwan, Holger H. Hoos, and Kevin Leyton-Brown. “Quantifying Algorithmic Improvements over Time.” In 27th International Joint Conference on Artificial Intelligence (IJCAI) Special Track on the Evolution of the Contours of AI, 2018. preprint PDF bibTeX abstract Assessing the progress made in AI and contribu- tions to the state of the art is of major concern to the community. Recently, Frechette et al. [2016] advocated performing such analysis via the Shapley value, a concept from coalitional game theory. In this paper, we argue that while this general idea is sound, it unfairly penalizes older algorithms that advanced the state of the art when introduced, but were then outperformed by modern counterparts. Driven by this observation, we introduce the tem- poral Shapley value, a measure that addresses this problem while maintaining the desirable properties of the (classical) Shapley value. We use the tempo- ral Shapley value to analyze the progress made in (i) the different versions of the Quicksort algorithm; (ii) the annual SAT competitions 2007–2014; (iii) an annual competition of Constraint Programming, namely the MiniZinc challenge 2014–2016. Our analysis reveals novel insights into the development made in these important areas of research over time. * Degroote, Hans, Patrick De Causmaecker, Bernd Bischl, and Lars Kotthoff. “A Regression-Based Methodology for Online Algorithm Selection.” In 11th International Symposium on Combinatorial Search (SoCS), 37–45, 2018. preprint PDF bibTeX abstract Algorithm selection approaches have achieved impressive performance improvements in many areas of AI. Most of the literature considers the offline algorithm selection problem, where the initial selection model is never updated after training. However, new data from running algorithms on instances becomes available when algorithms are selected and run. We investigate how this online data can be used to improve the selection model over time. This is especially relevant when insufficient training instances were used, but potentially improves the performance of algorithm selection in all cases. We formally define the online algorithm selection problem and model it as a contextual multi-armed bandit problem, propose a methodology for solving it, and empirically demonstrate performance improvements. We also show that our online algorithm selection method can be used when no training data whatsoever is available, a setting where offline algorithm selection cannot be used. Our experiments indicate that a simple greedy approach achieves the best performance. * Bhuiyan, Faisal H., Lars Kotthoff, and Ray S. Fertig. “A Machine Learning Technique to Predict Static Multi-Axial Failure Envelope of Laminated Composites.” In American Society for Composites 33rd Annual Technical Conference, 2018. preprint PDF bibTeX abstract A machine learning technique was used to predict static, failure envelopes of unidirectional composite laminas under combined normal (longitudinal or transverse) and shear loading at different biaxial ratios. An artificial neural network was chosen for this purpose due to their superior computational efficiency and ability to handle nonlinear relationships between inputs and outputs. Training and test data for the neural network were taken from the experimental composite failure data for glass- and carbon-fiber reinforced epoxies provided by the world-wide failure exercise (WWFE) program. A quadratic, stress interactive Tsai-Wu failure theory was calibrated based on the reported strength values, as well as optimized from the experimental failure data points. The prediction made by the neural network was compared against the Tsai-Wu failure criterion predictions and it was observed that the trained neural network provides a better representation of the experimental data. * Bistarelli, Stefano, Lars Kotthoff, Francesco Santini, and Carlo Taticchi. “Containerisation and Dynamic Frameworks in ICCMA’19.” In Second International Workshop on Systems and Algorithms for Formal Argumentation (SAFA 2018) Co-Located with the 7th International Conference on Computational Models of Argument (COMMA 2018), 2171:4–9. CEUR Workshop Proceedings. CEUR-WS.org, 2018. preprint PDF bibTeX abstract The International Competition on Computational Models of Argumentation (ICCMA) is a successful event dedicated to advancing the state of the art of solvers in Abstract Argumentation. We describe two proposals that will further improve the third and next edition of the competition, i.e. ICCMA 2019. The first novelty concerns the packaging of each solver-application participating in the competition in a virtual “light” container (using Docker): this allows for easy deploy- ment and to (re)running all of the submissions on different architectures (Linux, Windows, macOS, and also in the cloud). The second proposal consists of a new track focused on solvers processing dynamic frameworks, i.e., solvers described in terms of changes w.r.t. previous ones: a solver can reuse the solution obtained previously to be faster on the same framework modulo a new argument/attack. * Bessière, Christian, Luc de Raedt, Tias Guns, Lars Kotthoff, Mirco Nanni, Siegfried Nijssen, Barry O’Sullivan, Anastasia Paparrizou, Dino Pedreschi, and Helmut Simonis. “The Inductive Constraint Programming Loop.” IEEE Intelligent Systems, 2018. https://doi.org/10.1109/MIS.2017.265115706. preprint PDF bibTeX abstract Constraint programming is used for a variety of real-world optimization problems, such as planning, scheduling and resource allocation problems. At the same time, one continuously gathers vast amounts of data about these problems. Current constraint programming software does not exploit such data to update schedules, resources and plans. We propose a new framework, which we call the inductive constraint programming loop. In this approach data is gathered and analyzed systematically in order to dynamically revise and adapt constraints and optimization criteria. Inductive Constraint Programming aims at bridging the gap between the areas of data mining and machine learning on the one hand, and constraint programming on the other. * Kerschke, Pascal, Lars Kotthoff, Jakob Bossek, Holger H. Hoos, and Heike Trautmann. “Leveraging TSP Solver Complementarity through Machine Learning.” Evolutionary Computation 26, no. 4 (2018): 597–620. https://doi.org/10.1162/evco_a_00215. preprint PDF bibTeX abstract The Travelling Salesperson Problem (TSP) is one of the best-studied NP-hard problems. Over the years, many different solution approaches and solvers have been developed. For the first time, we directly compare five state-of-the-art inexact solvers—namely, LKH, EAX, restart variants of those, and MAOS—on a large set of well-known benchmark instances and demonstrate complementary performance, in that different instances may be solved most effectively by different algorithms. We leverage this complementarity to build an algorithm selector, which selects the best TSP solver on a per-instance basis and thus achieves significantly improved performance compared to the single best solver, representing an advance in the state of the art in solving the Euclidean TSP. Our in-depth analysis of the selectors provides insight into what drives this performance improvement. 2017 * Kotthoff, Lars, Barry Hurley, and Barry O’Sullivan. “The ICON Challenge on Algorithm Selection.” AI Magazine 38, no. 2 (2017): 91–93. preprint PDF bibTeX * Kotthoff, Lars, Chris Thornton, Holger H. Hoos, Frank Hutter, and Kevin Leyton-Brown. “Auto-WEKA 2.0: Automatic Model Selection and Hyperparameter Optimization in WEKA.” Journal of Machine Learning Research 18, no. 25 (2017): 1–5. preprint PDF bibTeX abstract WEKA is a widely used, open-source machine learning platform. Due to its intuitive interface, it is particularly popular with novice users. However, such users often find it hard to identify the best approach for their particular dataset among the many available. We describe the new version of Auto-WEKA, a system designed to help such users by automatically searching through the joint space of WEKA's learning algorithms and their respective hyperparameter settings to maximize performance, using a state-of-the-art Bayesian optimization method. Our new package is tightly integrated with WEKA, making it just as accessible to end users as any other learning algorithm. * Lindauer, Marius, Jan N. van Rijn, and Lars Kotthoff, eds. Proceedings of the Open Algorithm Selection Challenge. Vol. 79. Proceedings of Machine Learning Research. PMLR, 2017. bibTeX * ———. “Open Algorithm Selection Challenge 2017: Setup and Scenarios.” In Proceedings of the Open Algorithm Selection Challenge, 79:1–7. Proceedings of Machine Learning Research. Brussels, Belgium: PMLR, 2017. preprint PDF bibTeX abstract The 2017 algorithm selection challenge provided a snapshot of the state of the art in algorithm selection and garnered submissions from four teams. In this chapter, we describe the setup of the challenge and the algorithm scenarios that were used. * Fawcett, Chris, Lars Kotthoff, and Holger H. Hoos. “Hot-Rodding the Browser Engine: Automatic Configuration of JavaScript Compilers.” CoRR abs/1707.04245 (2017). preprint PDF bibTeX 2016 * Fréchette, Alexandre, Lars Kotthoff, Talal Rahwan, Holger H. Hoos, Kevin Leyton-Brown, and Tomasz P. Michalak. “Using the Shapley Value to Analyze Algorithm Portfolios.” In 30th AAAI Conference on Artificial Intelligence, 3397–3403, 2016. preprint PDF bibTeX abstract Algorithms for NP-complete problems often have different strengths and weaknesses, and thus algorithm portfolios often outperform individual algorithms. It is surprisingly difficult to quantify an component algorithm's contribution to such a portfolio. Reporting a component's standalone performance wrongly rewards near-clones while penalizing algorithms that have small but distinct areas of strength. Measuring a component's marginal contribution to an existing portfolio is better, but penalizes sets of strongly correlated algorithms, thereby obscuring situations in which it is essential to have at least one algorithm from such a set. This paper argues for analyzing component algorithm contributions via a measure drawn from coalitional game theory---the Shapley value---and yields insight into a research community's progress over time. We conclude with an application of the analysis we advocate to SAT competitions, yielding novel insights into the behaviour of algorithm portfolios, their components, and the state of SAT solving technology. * Bischl, Bernd, Pascal Kerschke, Lars Kotthoff, Marius Lindauer, Yuri Malitsky, Alexandre Fréchette, Holger H. Hoos, et al. “ASlib: A Benchmark Library for Algorithm Selection.” Artificial Intelligence Journal 237 (2016): 41–58. preprint PDF bibTeX abstract The task of algorithm selection involves choosing an algorithm from a set of algorithms on a per-instance basis in order to exploit the varying performance of algorithms over a set of instances. The algorithm selection problem is attracting increasing attention from researchers and practitioners in AI. Years of fruitful applications in a number of domains have resulted in a large amount of data, but the community lacks a standard format or repository for this data. This situation makes it difficult to share and compare different approaches effectively, as is done in other, more established fields. It also unnecessarily hinders new researchers who want to work in this area. To address this problem, we introduce a standardized format for representing algorithm selection scenarios and a repository that contains a growing number of data sets from the literature. Our format has been designed to be able to express a wide variety of different scenarios. To demonstrate the breadth and power of our platform, we describe a study that builds and evaluates algorithm selection models through a common interface. The results display the potential of algorithm selection to achieve significant performance improvements across a broad range of problems and algorithms. * Kotthoff, Lars, Ciaran McCreesh, and Christine Solnon. “Portfolios of Subgraph Isomorphism Algorithms.” In LION 10, 2016. preprint PDF bibTeX abstract Subgraph isomorphism is a computationally challenging problem with important practical applications, for example in computer vision, biochemistry, and model checking. There are a number of state-of-the-art algorithms for solving the problem, each of which has its own performance characteristics. As with many other hard problems, the single best choice of algorithm overall is rarely the best algorithm on an instance-by-instance. We develop an algorithm selection approach which leverages novel features to characterise subgraph isomorphism problems and dynamically decides which algorithm to use on a per-instance basis. We demonstrate substantial performance improvements on a large set of hard benchmark problems. In addition, we show how algorithm selection models can be leveraged to gain new insights into what affects the performance of an algorithm. * Bischl, Bernd, Michel Lang, Lars Kotthoff, Julia Schiffner, Jakob Richter, Erich Studerus, Giuseppe Casalicchio, and Zachary M. Jones. “Mlr: Machine Learning in R.” Journal of Machine Learning Research 17, no. 170 (2016): 1–5. preprint PDF bibTeX abstract The mlr package provides a generic, object- oriented, and extensible framework for classification, regression, survival analysis and clustering for the R language. It provides a unified interface to more than 160 basic learners and includes meta-algorithms and model selection techniques to improve and extend the functionality of basic learners with, e.g., hyperparameter tuning, feature selection, and ensemble construction. Parallel high-performance computing is natively supported. The package targets practitioners who want to quickly apply machine learning algorithms, as well as researchers who want to implement, benchmark, and compare their new methods in a structured environment. * Degroote, Hans, Bernd Bischl, Lars Kotthoff, and Patrick de Causmacker. “Reinforcement Learning for Automatic Online Algorithm Selection - an Empirical Study.” In ITAT, 1649:93–101. CEUR Workshop Proceedings, 2016. preprint PDF bibTeX abstract In this paper a reinforcement learning methodology for automatic online algorithm selection is introduced and empirically tested. It is applicable to automatic algorithm selection methods that predict the performance of each available algorithm and then pick the best one. The experiments confirm the usefulness of the methodology: using online data results in better performance. As in many online learning settings an exploration vs. exploitation trade-off, synonymously learning vs. earning trade-off, is incurred. Empirically investigating the quality of classic solution strategies for handling this trade-off in the automatic online algorithm selection setting is the secondary goal of this paper. The automatic online algorithm selection problem can be modelled as a contextual multi-armed bandit problem. Two classic strategies for solving this problem are tested in the context of automatic online algorithm selection: epsilon-greedy and lower confidence bound. The experiments show that a simple purely exploitative greedy strategy outperforms strategies explicitly performing exploration. * Bessière, Christian, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, eds. Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach. 1st ed. Vol. 10101. Lecture Notes in Artificial Intelligence. Springer, 2016. preprint PDF bibTeX abstract A successful integration of constraint programming and data mining has the potential to lead to a new ICT paradigm with far reaching implications. It could change the face of data mining and machine learning, as well as constraint programming technology. It would not only allow one to use data mining techniques in constraint programming to identify and update constraints and optimization criteria, but also to employ constraints and criteria in data mining and machine learning in order to discover models compatible with prior knowledge. This book reports on some key results obtained on this integrated and cross- disciplinary approach within the European FP7 FET Open project no. 284715 on “Inductive Constraint Programming” and a number of associated workshops and Dagstuhl seminars. The book is structured in five parts: background; learning to model; learning to solve; constraint programming for data mining; and showcases. * Kotthoff, Lars. “Algorithm Selection for Combinatorial Search Problems: A Survey.” In Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach, edited by Christian Bessiere, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, 149–90. Cham: Springer International Publishing, 2016. https://doi.org/10.1007/978-3-319-50137-6_7. bibTeX abstract The Algorithm Selection Problem is concerned with selecting the best algorithm to solve a given problem on a case-by-case basis. It has become especially relevant in the last decade, as researchers are increasingly investigating how to identify the most suitable existing algorithm for solving a problem instead of developing new algorithms. This survey presents an overview of this work focusing on the contributions made in the area of combinatorial search problems, where Algorithm Selection techniques have achieved significant performance improvements. We unify and organise the vast literature according to criteria that determine Algorithm Selection systems in practice. The comprehensive classification of approaches identifies and analyses the different directions from which Algorithm Selection has been approached. This chapter contrasts and compares different methods for solving the problem as well as ways of using these solutions. * Hurley, Barry, Lars Kotthoff, Barry O’Sullivan, and Helmut Simonis. “ICON Loop Health Show Case.” In Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach, edited by Christian Bessiere, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, 325–33. Cham: Springer International Publishing, 2016. https://doi.org/10.1007/978-3-319-50137-6_14. bibTeX abstract In this document we describe the health show case for the ICON project. This corresponds to Task 6.2 in WP 6 of the Description of Work for the project. The description provides a high-level abstraction, detailed description of the interfaces between modules, and a description of sample data. * Nanni, Mirco, Lars Kotthoff, Riccardo Guidotti, Barry O’Sullivan, and Dino Pedreschi. “ICON Loop Carpooling Show Case.” In Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach, edited by Christian Bessiere, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, 310–24. Cham: Springer International Publishing, 2016. https://doi.org/10.1007/978-3-319-50137-6_13. bibTeX abstract In this chapter we describe a proactive carpooling service that combines induction and optimization mechanisms to maximize the impact of carpooling within a community. The approach autonomously infers the mobility demand of the users through the analysis of their mobility traces (i.e. Data Mining of GPS trajectories) and builds the network of all possible ride sharing opportunities among the users. Then, the maximal set of carpooling matches that satisfy some standard requirements (maximal capacity of vehicles, etc.) is computed through Constraint Programming models, and the resulting matches are proactively proposed to the users. Finally, in order to maximize the expected impact of the service, the probability that each carpooling match is accepted by the users involved is inferred through Machine Learning mechanisms and put in the CP model. The whole process is reiterated at regular intervals, thus forming an instance of the general ICON loop. * Bessière, Christian, Luc De Raedt, Tias Guns, Lars Kotthoff, Mirco Nanni, Siegfried Nijssen, Barry O’Sullivan, Anastasia Paparrizou, Dino Pedreschi, and Helmut Simonis. “The Inductive Constraint Programming Loop.” In Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach, edited by Christian Bessiere, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, 303–9. Cham: Springer International Publishing, 2016. https://doi.org/10.1007/978-3-319-50137-6_12. bibTeX abstract Constraint programming is used for a variety of real-world optimization problems, such as planning, scheduling and resource allocation problems. At the same time, one continuously gathers vast amounts of data about these problems. Current constraint programming software does not exploit such data to update schedules, resources and plans. We propose a new framework, that we call the Inductive Constraint Programming (ICON) loop. In this approach data is gathered and analyzed systematically in order to dynamically revise and adapt constraints and optimization criteria. Inductive Constraint Programming aims at bridging the gap between the areas of data mining and machine learning on the one hand, and constraint programming on the other hand. * Hurley, Barry, Lars Kotthoff, Yuri Malitsky, Deepak Mehta, and Barry O’Sullivan. “Advanced Portfolio Techniques.” In Data Mining and Constraint Programming: Foundations of a Cross-Disciplinary Approach, edited by Christian Bessiere, Luc De Raedt, Lars Kotthoff, Siegfried Nijssen, Barry O’Sullivan, and Dino Pedreschi, 191–225. Cham: Springer International Publishing, 2016. https://doi.org/10.1007/978-3-319-50137-6_8. bibTeX abstract There exists a proliferation of different approaches to using portfolios and algorithm selection to make solving combinatorial search and optimisation problems more efficient, as surveyed in the previous chapter. In this chapter, we take a look at a detailed case study that leverages transformations between problem representations to make portfolios more effective, followed by extensions to the state of the art that make algorithm selection more robust in practice. 2015 * Kotthoff, Lars, Pascal Kerschke, Holger Hoos, and Heike Trautmann. “Improving the State of the Art in Inexact TSP Solving Using Per-Instance Algorithm Selection.” In LION 9, 202–17, 2015. preprint PDF bibTeX abstract We investigate per-instance algorithm selection techniques for solving the Travelling Salesman Problem (TSP), based on the two state-of-the-art inexact TSP solvers, LKH and EAX. Our comprehensive experiments demonstrate that the solvers exhibit complementary performance across a diverse set of instances, and the potential for improving the state of the art by selecting between them is significant. Using TSP features from the literature as well as a set of novel features, we show that we can capitalise on this potential by building an efficient selector that achieves significant performance improvements in practice. Our selectors represent a significant improvement in the state-of-the-art in inexact TSP solving, and hence in the ability to find optimal solutions (without proof of optimality) for challenging TSP instances in practice. * Kotthoff, Lars, Mirco Nanni, Riccardo Guidotti, and Barry O’Sullivan. “Find Your Way Back: Mobility Profile Mining with Constraints.” In CP, 638–53. Cork, Ireland, 2015. preprint PDF bibTeX abstract Mobility profile mining is a data mining task that can be formulated as clustering over movement trajectory data. The main challenge is to separate the signal from the noise, i.e.{\textbackslash} one-off trips. We show that standard data mining approaches suffer the important drawback that they cannot take the symmetry of non-noise trajectories into account. That is, if a trajectory has a symmetric equivalent that covers the same trip in the reverse direction, it should become more likely that neither of them is labelled as noise. We present a constraint model that takes this knowledge into account to produce better clusters. We show the efficacy of our approach on real-world data that was previously processed using standard data mining techniques. * Kotthoff, Lars, Barry O’Sullivan, S. S. Ravi, and Ian Davidson. “Complex Clustering Using Constraint Programming: Modelling Electoral Map Creation.” In 14th International Workshop on Constraint Modelling and Reformulation, 2015. preprint PDF bibTeX abstract Traditional clustering is limited to a single collection of objects, described by a set of features under simple objectives and constraints. Though this setting can scale to huge data sets, many real world problems do not fit it. Consider the problem motivating this work: creating electoral district maps. Not only are two sets of objects (electoral districts and elected officials) separately clustered simultaneously under complex constraints, the clusters must be matched and it is required to find a global optimum. Existing formulations of clustering such as those using procedural languages or convex programming cannot handle such complex settings. In this paper we explore clustering this complex setting using constraint programming. We implement our methods in the Numberjack language and make use of large-scale solvers such as Gurobi which exploit multi-core architectures. * Chue Hong, Neil P., Tom Crick, Ian P. Gent, Lars Kotthoff, and Kenji Takeda. “Top Tips to Make Your Research Irreproducible.” CoRR abs/1504.00062 (2015). preprint PDF bibTeX abstract It is an unfortunate convention of science that research should pretend to be reproducible; our top tips will help you mitigate this fussy conventionality, enabling you to enthusiastically showcase your irreproducible work. * Kotthoff, Lars. “ICON Challenge on Algorithm Selection.” CoRR abs/1511.04326 (2015). preprint PDF bibTeX 2014 * ———. “Reliability of Computational Experiments on Virtualised Hardware.” Journal of Experimental and Theoretical Artificial Intelligence 26, no. 1 (2014): 33–49. preprint PDF bibTeX abstract We present a large-scale investigation of the variability of run times on physical and virtualised hardware. The aim of our investigation is to establish whether cloud infrastructures are suitable for running computational experiments where the results rely on reported run times. Our application is the use of the Minion constraint solver as an example of an Artificial Intelligence experiment. We include two major providers of public cloud infrastructure, Amazon and Rackspace, as well as a private Eucalyptus cloud. While there are many studies in the literature that investigate the performance of cloud environments, the problem of whether this performance is consistent and run time measurements are reliable has been largely ignored. Our comprehensive experiments and detailed analysis of the results show that there is no intrinsic disadvantage of virtualised hardware over physical hardware and that in general cloud environments are suitable for running computational experiments. Our meticulous investigation reveals several interesting patterns in the variability of run times that researchers using a cloud for this purpose should be aware of. We close by giving recommendations as to which type of virtual machine with which cloud provider should be used to achieve reproducible results. * ———. “Algorithm Selection for Combinatorial Search Problems: A Survey.” AI Magazine 35, no. 3 (2014): 48–60. preprint PDF bibTeX abstract The Algorithm Selection Problem is concerned with selecting the best algorithm to solve a given problem instance on a case-by-case basis. It has become especially relevant in the last decade, with researchers increasingly investigating how to identify the most suitable existing algorithm for solving a problem instance instead of developing new algorithms. This survey presents an overview of this work focusing on the contributions made in the area of combinatorial search problems, where algorithm selection techniques have achieved significant performance improvements. We unify and organise the vast literature according to criteria that determine algorithm selection systems in practice. The comprehensive classification of approaches identifies and analyses the different directions from which algorithm selection has been approached. This paper contrasts and compares different methods for solving the problem as well as ways of using these solutions. * ———. “Ranking Algorithms by Performance.” In LION 8, 16–19, 2014. preprint PDF bibTeX * Geschwender, Daniel, Frank Hutter, Lars Kotthoff, Yuri Malitsky, Holger H. Hoos, and Kevin Leyton-Brown. “Algorithm Configuration in the Cloud: A Feasibility Study.” In LION 8, 41–44, 2014. preprint PDF bibTeX * Hurley, Barry, Lars Kotthoff, Yuri Malitsky, and Barry O’Sullivan. “Proteus: A Hierarchical Portfolio of Solvers and Transformations.” In CPAIOR, 301–17, 2014. preprint PDF bibTeX abstract In recent years, portfolio approaches to solving SAT problems and CSPs have become increasingly common. There are also a number of different encodings for representing CSPs as SAT instances. In this paper, we leverage advances in both SAT and CSP solving to present a novel hierarchical portfolio-based approach to CSP solving, which we call Proteus, that does not rely purely on CSP solvers. Instead, it may decide that it is best to encode a CSP problem instance into SAT, selecting an appropriate encoding and a corresponding SAT solver. Our experimental evaluation uses an instance of Proteus that involved four CSP solvers, three SAT encodings, and six SAT solvers, evaluated on the most challenging problem instances from the CSP solver competitions, involving global and intensional constraints. We demonstrate that significant performance improvements can be achieved by Proteus obtained by exploiting alternative view-point and solvers for combinatorial problem-solving. * Kelsey, Thomas W., Lars Kotthoff, Christopher A. Jefferson, Stephen A. Linton, Ian Miguel, Peter Nightingale, and Ian P. Gent. “Qualitative Modelling via Constraint Programming.” Constraints 19, no. 2 (2014): 163–73. preprint PDF bibTeX abstract Qualitative modelling is a technique integrating the fields of theoretical computer science, artificial intelligence and the physical and biological sciences. The aim is to be able to model the behaviour of systems without estimating parameter values and fixing the exact quantitative dynamics. Traditional applications are the study of the dynamics of physical and biological systems at a higher level of abstraction than that obtained by estimation of numerical parameter values for a fixed quantitative model. Qualitative modelling has been studied and implemented to varying degrees of sophistication in Petri nets, process calculi and constraint programming. In this paper we reflect on the strengths and weaknesses of existing frameworks, we demonstrate how recent advances in constraint programming can be leveraged to produce high quality qualitative models, and we describe the advances in theory and technology that would be needed to make constraint programming the best option for scientific investigation in the broadest sense. * Johnson, Peter George, Tina Balke, and Lars Kotthoff. “Integrating Optimisation and Agent-Based Modelling.” In 28th European Conference on Modelling & Simulation. Brescia, Italy, 2014. preprint PDF bibTeX abstract A key strength of agent-based modelling is the ability to explore the upward-causation of micro-dynamics on the macro-level behaviour of a system. However, in policy contexts, it is also important to be able to represent downward-causation from the macro and meso-levels to the micro, and to represent decision-making at the macro level (i.e., by governments) in a sensible way. Though we cannot model political processes easily, we can try to optimise decisions given certain stated goals (e.g., minimum cost, or maximum impact). Optimisation offers one potential method to model macro-level decisions in this way. This paper presents the implementation of an integration of optimisation with agent-based modelling for the example of an auction scenario of government support for the installation of photovoltaic solar panels by households. Auction type scenarios of this kind, in which large groups of individuals or organisations make bids for subsidies or contracts from government, are common in many policy domains. * Hussain, Bilal, Ian P. Gent, Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, Glenna F. Nightingale, and Peter Nightingale. “Discriminating Instance Generation for Automated Constraint Model Selection.” In 20th International Conference on Principles and Practice of Constraint Programming, 356–65. Lyon, France, 2014. preprint PDF bibTeX abstract One approach to automated constraint modelling is to generate, and then select from, a set of candidate models. This method is used by the automated modelling system CONJURE. To select a preferred model or set of models for a problem class from the candidates, CONJURE uses a set of training instances drawn from the target class. It is important that the training instances are discriminating. If all models solve a given instance in a trivial amount of time, or if no models solve it in the time available, then the instance is not useful for model selection. This paper addresses the task of generating small sets of discriminating training instances automatically. The instance space is determined by the parameters of the associated problem class. We develop a number of methods of finding parameter configurations that give discriminating training instances, some of them leveraging existing parameter-tuning techniques. Our experimental results confirm the success of our approach in reducing a large set of input models to a small set that we can expect to perform well for the given problem class. * Kelsey, Thomas W., Martin McCaffery, and Lars Kotthoff. “Web-Scale Distributed EScience AI Search across Disconnected and Heterogeneous Infrastructures.” In 10th IEEE International Conference on EScience, 39–46. Guarujá, Brazil, 2014. preprint PDF bibTeX abstract We present a robust and generic framework for web-scale distributed e-Science Artificial Intelligence search. Our validation approach is to distribute constraint satisfaction problems that require perfect accuracy to 10, 12 and 15 digits. By checking solutions obtained using the framework against known results, we can ensure that no errors, duplications nor omissions are introduced. Unlike other approaches, we do not require dedicated machines, homogeneous infrastructure or the ability to communicate between nodes. We give special consideration to the robustness of the framework, minimising the loss of effort even after a total loss of infrastructure, and allowing easy verification of every step of the distribution process. The unique challenges our framework tackles are related to the combinatorial explosion of the space that contains the possible solutions, and the robustness of long-running computations. Not only is the time required to finish the computations unknown, but also the resource requirements may change during the course of the computation. We demonstrate the applicability of our framework by using it to solve challenging problems using two separate large-scale distribution paradigms. The results show that our approach scales to e-Science computations of a size that would have been impossible to tackle just a decade ago. * Gent, Ian P., and Lars Kotthoff. “Recomputation.org: Experience of Its First Year and Lessons Learned.” In Recomputability 2014. London, UK, 2014. preprint PDF bibTeX abstract We founded recomputation.org about 18 months ago as we write. The site is intended to serve as a repository for computational experiments, embodied in virtual machines so that they can be recomputed at will by other researchers. We reflect in this paper on those aspects of recomputation.org that have worked well, those that have worked less well, and to what extent our views have changed on reproducibility in computational science. * Kotthoff, Lars. “Algorithm Selection in Practice.” AISB Quarterly, no. 138 (2014): 4–8. preprint PDF bibTeX * Wilson, Nic, and Lars Kotthoff. “Taking into Account Expected Future Bids in EPolicy Optimisation Problem.” Insight Centre for Data Analytics, July 2014. preprint PDF bibTeX * Arabas, Sylwester, Michael R. Bareford, Lakshitha R. de Silva, Ian P. Gent, Benjamin M. Gorman, Masih Hajiarabderkani, Tristan Henderson, et al. “Case Studies and Challenges in Reproducibility in the Computational Sciences.” arXiv, 2014. https://doi.org/10.48550/ARXIV.1408.2123. preprint PDF bibTeX abstract This paper investigates the reproducibility of computational science research and identifies key challenges facing the community today. It is the result of the First Summer School on Experimental Methodology in Computational Science Research (this https URL). First, we consider how to reproduce experiments that involve human subjects, and in particular how to deal with different ethics requirements at different institutions. Second, we look at whether parallel and distributed computational experiments are more or less reproducible than serial ones. Third, we consider reproducible computational experiments from fields outside computer science. Our final case study looks at whether reproducibility for one researcher is the same as for another, by having an author attempt to have others reproduce their own, reproducible, paper. This paper is open, executable and reproducible: the whole process of writing this paper is captured in the source control repository hosting both the source of the paper, supplementary codes and data; we are providing setup for several experiments on which we were working; finally, we try to describe what we have achieved during the week of the school in a way that others may reproduce (and hopefully improve) our experiments. 2013 * Kotthoff, Lars, and Barry O’Sullivan. “Constraint-Based Clustering.” In 10th International Conference on Integration of Artificial Intelligence (AI) and Operations Research (OR) Techniques in Constraint Programming, 2013. bibTeX abstract Machine learning and constraint programming are almost completely independent research fields. However, there are significant opportunities for synergy between them. In this presentation, we introduce a constraint programming approach to the classification problem in machine learning. Specifically, we treat classification as a clustering problem. Previous approaches have used constraints as a means of representing background knowledge to improve the quality of the resulting clustering. We show how to use such constraints to not only guide the machine learning algorithms, but replace them entirely. Our approach uses an off-the-shelf constraint solver to find the clustering that reflects as much background knowledge as possible. A second formulation allows us to optimise for the objectives commonly used in machine learning algorithms, such as maximising the inter-cluster distances. We present an evaluation results of our approaches on a variety of well-known benchmarks covering a range of different application domains. Our approaches can significantly outperform standard clustering methods used in machine learning in terms of the quality of the resulting clustering and classification. In addition, the constraint programming formulation provides much more flexibility and customisation opportunities than standard machine learning approaches. * Akgun, Ozgur, Alan M. Frisch, Bilal Hussain, Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “An Automated Constraint Modelling and Solving Toolchain.” In 20th Automated Reasoning Workshop, 2013. preprint PDF bibTeX * Prokopas, Arunas, Alan M. Frisch, Ian P. Gent, Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “Constructing Constraint Solvers Using Monte Carlo Tree Search.” In 20th Automated Reasoning Workshop, 2013. preprint PDF bibTeX abstract Constraint solvers are complex pieces of software that are capable of solving a wide variety of problems. Customisation and specialisation opportunities are usually very limited and require specialist knowledge. The Dominion constraint solver synthesizer automatically creates problem-specific solvers. The configuration of a constraint solver is highly complex, especially if the aim is to achieve high performance. We demonstrate how Monte Carlo Tree Search can be employed to tackle this problem. * Kotthoff, Lars. “LLAMA: Leveraging Learning to Automatically Manage Algorithms.” arXiv, June 2013. preprint PDF bibTeX abstract Algorithm portfolio approaches have achieved remarkable improvements over single solvers. However, the implementation of such systems is often highly specialized and specific to the problem domain. This makes it difficult for researchers to explore different techniques for their specific problems. We present LLAMA, a modular and extensible toolkit that facilitates the exploration of a range of different portfolio techniques on any problem domain. We describe the current capabilities and limitations of the toolkit and illustrate its usage on a set of example SAT problems. * Hurley, Barry, Lars Kotthoff, Yuri Malitsky, and Barry O’Sullivan. “Proteus: A Hierarchical Portfolio of Solvers and Transformations.” arXiv, June 2013. preprint PDF bibTeX abstract In recent years, portfolio approaches to solving SAT problems and CSPs have become increasingly common. There are also a number of different techniques for converting SAT problems into CSPs. In this paper, we leverage advances in both areas and present a novel hierarchical portfolio-based approach to CSP solving that does not rely purely on CSP solvers, but may convert a problem to SAT choosing a conversion technique and the accommodating SAT solver. Our experimental evaluation relies on competition CSP instances and uses eight CSP solvers, three SAT encodings and eighteen SAT solvers. We demonstrate that significant performance improvements can be obtained by considering alternative view-points of a combinatorial problem. * Akgun, Ozgur, Alan M. Frisch, Ian P. Gent, Bilal Hussain, Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “Automated Symmetry Breaking and Model Selection in Conjure.” In 19th International Conference on Principles and Practice of Constraint Programming, 107–16. Uppsala, Sweden, 2013. preprint PDF bibTeX abstract Constraint modelling is widely recognised as a key bottleneck in applying constraint solving to a problem of interest. The Conjure automated constraint modelling system addresses this problem by automatically refining constraint models from problem specifications written in the Essence language. Essence provides familiar mathematical concepts like sets, functions and relations nested to any depth. To date, Conjure has been able to produce a set of alternative model kernels (i.e. without advanced features such as symmetry breaking or implied constraints) for a given specification. The first contribution of this paper is a method by which Conjure can break symmetry in a model as it is introduced by the modelling process. This works at the problem class level, rather than just individual instances, and does not require an expensive detection step after the model has been formulated. This allows Conjure to produce a higher quality set of models. A further limitation of Conjure has been the lack of a mechanism to select among the models it produces. The second contribution of this paper is to present two such mechanisms, allowing effective models to be chosen automatically. * Mehta, Deepak, Barry O’Sullivan, Lars Kotthoff, and Yuri Malitsky. “Lazy Branching for Constraint Satisfaction.” In ICTAI, 1012–19, 2013. preprint PDF bibTeX abstract When solving a constraint satisfaction problem using a systematic backtracking method, the branching scheme normally selects a variable to which a value is assigned. In this paper we refer to such strategies as eager branching schemes. These contrast with the alternative class of novel branching schemes considered in this paper whereby having selected a variable we proceed by removing values from its domain. In this paper we study such lazy branching schemes in depth. We define three lazy branching schemes based on k-way, binary and split branching. We show how each can be incorporated into MAC, and define a novel value ordering heuristic that is suitable in this setting. Our results show that lazy branching can significantly out-perform traditional branching schemes across a variety of problem classes. While, in general, neither lazy nor eager branching dominates the other, our results clearly show that choosing the correct branching scheme for a given problem instance can significantly reduce search effort. Therefore, we implemented a variety of branching portfolios for choosing amongst all of the branching strategies studied in this paper. The results demonstrate that a good branching scheme can be automatically selected for a given problem instances and that including lazy branching schemes in the portfolio significantly reduces runtime. * Kotthoff, Lars. “Ranking Algorithms by Performance,” November 2013. preprint PDF bibTeX abstract A common way of doing algorithm selection is to train a machine learning model and predict the best algorithm from a portfolio to solve a particular problem. While this method has been highly successful, choosing only a single algorithm has inherent limitations -- if the choice was bad, no remedial action can be taken and parallelism cannot be exploited, to name but a few problems. In this paper, we investigate how to predict the ranking of the portfolio algorithms on a particular problem. This information can be used to choose the single best algorithm, but also to allocate resources to the algorithms according to their rank. We evaluate a range of approaches to predict the ranking of a set of algorithms on a problem. We furthermore introduce a framework for categorizing ranking predictions that allows to judge the expressiveness of the predictive output. Our experimental evaluation demonstrates on a range of data sets from the literature that it is beneficial to consider the relationship between algorithms when predicting rankings. We furthermore show that relatively naive approaches deliver rankings of good quality already. 2012 * Kelsey, Thomas W., Lars Kotthoff, Christopher A. Jefferson, Stephen A. Linton, Ian Miguel, Peter Nightingale, and Ian P. Gent. “Qualitative Modelling via Constraint Programming: Past, Present and Future.” In 18th International Conference on Principles and Practice of Constraint Programming (Position Paper), 2012. preprint PDF bibTeX abstract Qualitative modelling is a technique integrating the fields of theoretical computer science, artificial intelligence and the physical and biological sciences. The aim is to be able to model the behaviour of systems without estimating parameter values and fixing the exact quantitative dynamics. Traditional applications are the study of the dynamics of physical and biological systems at a higher level of abstraction than that obtained by estimation of numerical parameter values for a fixed quantitative model. Qualitative modelling has been studied and implemented to varying degrees of sophistication in Petri nets, process calculi and constraint programming. In this paper we reflect on the strengths and weaknesses of existing frameworks, we demonstrate how recent advances in constraint programming can be leveraged to produce high quality qualitative models, and we describe the advances in theory and technology that would be needed to make constraint programming the best option for scientific investigation in the broadest sense. * Distler, Andreas, Christopher A. Jefferson, Tom Kelsey, and Lars Kotthoff. “The Semigroups of Order 10.” In 18th International Conference on Principles and Practice of Constraint Programming, 883–99, 2012. preprint PDF bibTeX abstract The number of finite semigroups increases rapidly with the number of elements. Since existing enumeration formulae do not give the complete number of semigroups of given order up to symmetric equivalence, the remainder can only be found by careful search. We describe the use of mathematical results combined with distributed Constraint Satisfaction to show that the number of non-equivalent semigroups of order 10 is 12,418,001,077,381,302,684. This solves a previously open problem in Mathematics, and has directly led to improvements in Constraint Satisfaction technology. * Kotthoff, Lars, Ian P. Gent, and Ian Miguel. “An Evaluation of Machine Learning in Algorithm Selection for Search Problems.” AI Communications 25, no. 3 (2012): 257–70. preprint PDF bibTeX abstract Machine learning is an established method of selecting algorithms to solve hard search problems. Despite this, to date no systematic comparison and evaluation of the different techniques has been performed and the performance of existing systems has not been critically compared with other approaches. We compare the performance of a large number of different machine learning techniques from different machine learning methodologies on five data sets of hard algorithm selection problems from the literature. In addition to well-established approaches, for the first time we also apply statistical relational learning to this problem. We demonstrate that there is significant scope for improvement both compared with existing systems and in general. To guide practitioners, we close by giving clear recommendations as to which machine learning techniques are likely to achieve good performance in the context of algorithm selection problems. In particular, we show that linear regression and alternating decision trees have a very high probability of achieving better performance than always selecting the single best algorithm. * Hammond, Gail, Samantha Krause, Lars Kotthoff, and Thomas H. Guderjan. “Continuing Research Using Landscape Archaeology and GIS at Nojol Nah, Belize.” In 77th Annual Meeting of the Society for American Archaeology. Memphis, TN, 2012. preprint PDF bibTeX abstract We present preliminary results of interdisciplinary efforts in determining the use and management of ancient Maya rural landscapes in northwestern Belize. Through our comprehensive data set combining archaeology, GIS, land survey and soil science, we provide a singular insight into the strategies used by the people that inhabited this site on the periphery of the Maya world. Georeferenced maps - the results of intensive pedestrian survey - have been combined with a NASA digital elevation model as well as hydrological and soil data into a regional geodatabase, which includes the results of ongoing test units, and larger excavations. * Kotthoff, Lars, and Thomas H. Guderjan. “An Interactive Atlas of Maya Sites.” In 77th Annual Meeting of the Society for American Archaeology. Memphis, TN, 2012. preprint PDF bibTeX abstract Every year, a large amount of geographical data on Maya settlements is generated in the course of excavations, surveys and similar operations. Only some of those data are published in reports and most of it is not used more than once. We present a new effort that aims to make publicly available such data as the location of individual sites as well as detailed maps, excavation pictures and contextual information. All this is available through an easy-to-use web interface at www.mayamap.org that brings the power of traditional GIS systems to the layman user. * Kotthoff, Lars. “On Algorithm Selection, with an Application to Combinatorial Search Problems.” Ph.D., University of St Andrews, 2012. preprint PDF bibTeX abstract The Algorithm Selection Problem is to select the most appropriate way for solving a problem given a choice of different ways. Some of the most prominent and successful applications come from Artificial Intelligence and in particular combinatorial search problems. Machine Learning has established itself as the de facto way of tackling the Algorithm Selection Problem. Yet even after a decade of intensive research, there are no established guidelines as to what kind of Machine Learning to use and how. This dissertation presents an overview of the field of Algorithm Selection and associated research and highlights the fundamental questions left open and problems facing practitioners. In a series of case studies, it underlines the difficulty of doing Algorithm Selection in practice and tackles issues related to this. The case studies apply Algorithm Selection techniques to new problem domains and show how to achieve significant performance improvements. Lazy learning in constraint solving and the implementation of the alldifferent constraint are the areas in which we improve on the performance of current state of the art systems. The case studies furthermore provide empirical evidence for the effectiveness of using the misclassification penalty as an input to Machine Learning. After having established the difficulty, we present an effective technique for reducing it. Machine Learning ensembles are a way of reducing the background knowledge and experimentation required from the researcher while increasing the robustness of the system. Ensembles do not only decrease the difficulty, but can also increase the performance of Algorithm Selection systems. They are used to much the same ends in Machine Learning itself. We finally tackle one of the great remaining challenges of Algorithm Selection — which Machine Learning technique to use in practice. Through a large-scale empirical evaluation on diverse data taken from Algorithm Selection applications in the literature, we establish recommendations for Machine Learning algorithms that are likely to perform well in Algorithm Selection for combinatorial search problems. The recommendations are based on strong empirical evidence and additional statistical simulations. The research presented in this dissertation significantly reduces the knowledge threshold for researchers who want to perform Algorithm Selection in practice. It makes major contributions to the field of Algorithm Selection by investigating fundamental issues that have been largely ignored by the research community so far. * Balasubramaniam, Dharini, Christopher Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “An Automated Approach to Generating Efficient Constraint Solvers.” In 34th International Conference on Software Engineering, 661–71, 2012. preprint PDF bibTeX abstract Combinatorial problems appear in numerous settings, from timetabling to industrial design. Constraint solving aims to find solutions to such problems efficiently and automatically. Current constraint solvers are monolithic in design, accepting a broad range of problems. The cost of this convenience is a complex architecture, inhibiting efficiency, extensibility and scalability. Solver components are also tightly coupled with complex restrictions on their configuration, making automated generation of solvers difficult. We describe a novel, automated, model-driven approach to generating efficient solvers tailored to individual problems and present some results from applying the approach. The main contribution of this work is a solver generation framework called Dominion, which analyses a problem and, based on its characteristics, generates a solver using components chosen from a library. The key benefit of this approach is the ability to solve larger and more difficult problems as a result of applying finer-grained optimisations and using specialised techniques as required. * Kotthoff, Lars. “Hybrid Regression-Classification Models for Algorithm Selection.” In 20th European Conference on Artificial Intelligence, 480–85, 2012. preprint PDF bibTeX abstract Many state of the art Algorithm Selection systems use Machine Learning to either predict the run time or a similar performance measure of each of a set of algorithms and choose the algorithm with the best predicted performance or predict the best algorithm directly. We present a technique based on the well-established Machine Learning technique of stacking that combines the two approaches into a new hybrid approach and predicts the best algorithm based on predicted run times. We demonstrate significant performance improvements of up to a factor of six compared to the previous state of the art. Our approach is widely applicable and does not place any restrictions on the performance measure used, the way to predict it or the Machine Learning used to predict the best algorithm. We investigate different ways of deriving new Machine Learning features from the predicted performance measures and evaluate their effectiveness in increasing performance further. We use five different regression algorithms for performance prediction on five data sets from the literature and present strong empirical evidence that shows the effectiveness of our approach. * ———. “Algorithm Selection for Combinatorial Search Problems: A Survey.” University College Cork, 2012. preprint PDF bibTeX abstract The Algorithm Selection Problem is concerned with selecting the best algorithm to solve a given problem on a case-by-case basis. It has become especially relevant in the last decade, as researchers are increasingly investigating how to identify the most suitable existing algorithm for solving a problem instead of developing new algorithms. This survey presents an overview of this work focusing on the contributions made in the area of combinatorial search problems, where Algorithm Selection techniques have achieved significant performance improvements. We unify and organise the vast literature according to criteria that determine Algorithm Selection systems in practice. The comprehensive classification of approaches identifies and analyses the different directions from which Algorithm Selection has been approached. This paper contrasts and compares different methods for solving the problem as well as ways of using these solutions. It closes by identifying directions of current and future research. 2011 * Kelsey, Tom, and Lars Kotthoff. “Exact Closest String as a Constraint Satisfaction Problem.” In Proceedings of the International Conference on Computational Science, 1062–71, 2011. preprint PDF bibTeX abstract We report the first evaluation of Constraint Satisfaction as a computational framework for solving closest string problems. We show that careful consideration of symbol occurrences can provide search heuristics that provide several orders of magnitude speedup at and above the optimal distance. We also report the first analysis and evaluation – using any technique – of the computational difficulties involved in the identification of all closest strings for a given input set. We describe algorithms for web-scale distributed solution of closest string problems, both purely based on AI backtrack search and also hybrid numeric-AI methods. * Kotthoff, Lars, Ian P. Gent, and Ian Miguel. “A Preliminary Evaluation of Machine Learning in Algorithm Selection for Search Problems.” In Fourth Annual Symposium on Combinatorial Search, 84–91, 2011. preprint PDF bibTeX abstract Machine learning is an established method of selecting algorithms to solve hard search problems. Despite this, to date no systematic comparison and evaluation of the different techniques has been performed and the performance of existing systems has not been critically compared to other approaches. We compare machine learning techniques for algorithm selection on real-world data sets of hard search problems. In addition to well-established approaches, for the first time we also apply statistical relational learning to this problem. We demonstrate that most machine learning techniques and existing systems perform less well than one might expect. To guide practitioners, we close by giving clear recommendations as to which machine learning techniques are likely to perform well based on our experiments. * Gent, Ian P., Christopher A. Jefferson, Lars Kotthoff, and Ian Miguel. “Modelling Constraint Solver Architecture Design as a Constraint Problem.” In Annual ERCIM Workshop on Constraint Solving and Constraint Logic Programming, 87–96, 2011. preprint PDF bibTeX abstract Designing component-based constraint solvers is a complex problem. Some components are required, some are optional and there are interdependencies between the components. Because of this, previous approaches to solver design and modification have been ad-hoc and limited. We present a system that transforms a description of the components and the characteristics of the target constraint solver into a constraint problem. Solving this problem yields the description of a valid solver. Our approach represents a significant step towards the automated design and synthesis of constraint solvers that are specialised for individual constraint problem classes or instances. * Gent, Ian P., and Lars Kotthoff. “Reliability of Computational Experiments on Virtualised Hardware.” In AAAI Workshop AI for Data Center Management and Cloud Computing, 2011. preprint PDF bibTeX abstract We present preliminary results of an investigation into the suitability of virtualised hardware -- in particular clouds -- for running computational experiments. Our main concern was that the reported CPU time would not be reliable and reproducible. The results demonstrate that while this is true in cases where many virtual machines are running on the same physical hardware, there is no inherent variation introduced by using virtualised hardware compared to non-virtualised hardware. * Balasubramaniam, Dharini, Lakshitha de Silva, Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “Dominion: An Architecture-Driven Approach to Generating Efficient Constraint Solvers.” In 9th Working IEEE/IFIP Conference on Software Architecture, 228–31, 2011. preprint PDF bibTeX abstract Constraints are used to solve combinatorial problems in a variety of industrial and academic disciplines. However most constraint solvers are designed to be general and monolithic, leading to problems with efficiency, scalability and extensibility. We propose a novel, architecture-driven constraint solver generation framework called Dominion to tackle these issues. For any given problem, Dominion generates a lean and efficient solver tailored to that problem. In this paper, we outline the Dominion approach and its implications for software architecture specification of the solver. 2010 * Kotthoff, Lars, and Neil C.A. Moore. “Distributed Solving through Model Splitting.” In 3rd Workshop on Techniques for Implementing Constraint Programming Systems (TRICS), 26–34, 2010. preprint PDF bibTeX abstract Constraint problems can be trivially solved in parallel by exploring different branches of the search tree concurrently. Previous approaches have focused on implementing this functionality in the solver, more or less transparently to the user. We propose a new approach, which modifies the constraint model of the problem. An existing model is split into new models with added constraints that partition the search space. Optionally, additional constraints are imposed that rule out the search already done. The advantages of our approach are that it can be implemented easily, computations can be stopped and restarted, moved to different machines and indeed solved on machines which are not able to communicate with each other at all. * Gent, Ian P., Lars Kotthoff, Ian Miguel, and Peter Nightingale. “Machine Learning for Constraint Solver Design – a Case Study for the Alldifferent Constraint.” In 3rd Workshop on Techniques for Implementing Constraint Programming Systems (TRICS), 13–25, 2010. preprint PDF bibTeX abstract Constraint solvers are complex pieces of software which require many design decisions to be made by the implementer based on limited information. These decisions affect the performance of the finished solver significantly. Once a design decision has been made, it cannot easily be reversed, although a different decision may be more appropriate for a particular problem. We investigate using machine learning to make these decisions automatically depending on the problem to solve. We use the alldifferent constraint as a case study. Our system is capable of making non-trivial, multi-level decisions that improve over always making a default choice and can be implemented as part of a general-purpose constraint solver. * Kotthoff, Lars, Ian Miguel, and Peter Nightingale. “Ensemble Classification for Constraint Solver Configuration.” In 16th International Conference on Principles and Practices of Constraint Programming, 321–29, 2010. preprint PDF bibTeX abstract The automatic tuning of the parameters of algorithms and automatic selection of algorithms has received a lot of attention recently. One possible approach is the use of machine learning techniques to learn classifiers which, given the characteristics of a particular problem, make a decision as to which algorithm or what parameters to use. Little research has been done into which machine learning algorithms are suitable and the impact of picking the "right" over the "wrong" technique. This paper investigates the differences in performance of several techniques on different data sets. It furthermore provides evidence that by using a meta-technique which combines several machine learning algorithms, we can avoid the problem of having to pick the "best" one and still achieve good performance. * Gent, Ian P., Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, Neil Moore, Peter Nightingale, and Karen E. Petrie. “Learning When to Use Lazy Learning in Constraint Solving.” In 19th European Conference on Artificial Intelligence, 873–78, 2010. preprint PDF bibTeX abstract Learning in the context of constraint solving is a technique by which previously unknown constraints are uncovered during search and used to speed up subsequent search. Recently, lazy learning, similar to a successful idea from satisfiability modulo theories solvers, has been shown to be an effective means of incorporating constraint learning into a solver. Although a powerful technique to reduce search in some circumstances, lazy learning introduces a substantial overhead, which can outweigh its benefits. Hence, it is desirable to know beforehand whether or not it is expected to be useful. We approach this problem using machine learning (ML). We show that, in the context of a large benchmark set, standard ML approaches can be used to learn a simple, cheap classifier which performs well in identifying instances on which lazy learning should or should not be used. Furthermore, we demonstrate significant performance improvements of a system using our classifier and the lazy learning and standard constraint solvers over a standard solver. Through rigorous cross-validation across the different problem classes in our benchmark set, we show the general applicability of our learned classifier. * Kotthoff, Lars, Ian P. Gent, and Ian Miguel. “Using Machine Learning to Make Constraint Solver Implementation Decisions.” In SICSA PhD Conference, 2010. preprint PDF bibTeX abstract Programs to solve so-called constraint problems are complex pieces of software which require many design decisions to be made more or less arbitrarily by the implementer. These decisions affect the performance of the finished solver significantly. Once a design decision has been made, it cannot easily be reversed, although a different decision may be more appropriate for a particular problem. We investigate using machine learning to make these decisions automatically depending on the problem to solve with the alldifferent constraint as an example. Our system is capable of making non-trivial, multi-level decisions that improve over always making a default choice. 2009 * Kotthoff, Lars. “Constraint Solvers: An Empirical Evaluation of Design Decisions.” CIRCA preprint. University of St Andrews, Centre for Interdisciplinary Research in Computational Algebra, 2009. preprint PDF bibTeX abstract This paper presents an evaluation of the design decisions made in four state-of-the-art constraint solvers; Choco, ECLiPSe, Gecode, and Minion. To assess the impact of design decisions, instances of the five problem classes n-Queens, Golomb Ruler, Magic Square, Social Golfers, and Balanced Incomplete Block Design are modelled and solved with each solver. The results of the experiments are not meant to give an indication of the performance of a solver, but rather investigate what influence the choice of algorithms and data structures has. The analysis of the impact of the design decisions focuses on the different ways of memory management, behaviour with increasing problem size, and specialised algorithms for specific types of variables. It also briefly considers other, less significant decisions. * Gent, Ian P., Christopher A. Jefferson, Lars Kotthoff, Ian Miguel, and Peter Nightingale. “Specification of the Dominion Input Language Version 0.1.” University of St Andrews, 2009. preprint PDF bibTeX * Kotthoff, Lars. “Dominion – A Constraint Solver Generator.” In Doctoral Program of CP, 2009. preprint PDF bibTeX abstract This paper proposes a design for a system to generate constraint solvers that are specialised for specific problem models. It describes the design in detail and gives preliminary experimental results showing the feasibility and effectiveness of the approach. 2007 * ———. “Using Constraints to Render Websites — Applications of Artificial Intelligence in E-Commerce Environments.” Diplom, University of Leipzig, 2007. preprint PDF bibTeX abstract Constraint programming is an area of Artificial Intelligence which has many applica- tions. This thesis applies its techniques to a new kind of problem – the rendering of online retailer websites. First, in-depth introductions to constraint programming and the problem of rendering a shop website will be given. A prototypical implementation of a constraint problem solver and a system to solve and illustrate the problem will be described. The architecture of the prototypical implementation and specific features, algo- rithms, and design decisions will be detailed, analysed, and illustrated. An overview of related work both in the fields of constraint programming and website generation will be presented and existing technologies evaluated. Features and concepts unique to this thesis, like real-time constraint satisfaction, will be introduced and discussed. Finally, a comprehensive example will illustrate the problem, means of modelling it, and possible solutions. An outlook to future work and a summary conclude the thesis. SOFTWARE * Maintainer of the FSelector R package. * Author and maintainer of LLAMA, an R package to simplify common algorithm selection tasks such as training a classifier as portfolio selector. * Core contributor to the mlr R package (Github) for all things machine learning in R. * Leading the Auto-WEKA project, which brings automated machine learning to WEKA. TEACHING * I am teaching COSC 3020 (Algorithms and Data Structures) and COSC 4552/5552. Lecture materials, assignments, announcements, etc. are available on WyoCourses. * I am teaching a practical machine learning course using mlr. The slides are available here. AWARDS * Co-PI on NSF award 2055621, RET Site: WySTACK - Supporting Teachers And Computing Knowledge ($600,000). * Co-PI on NASA EPSCoR award for advanced manufacturing of flexible electronics ($749,997). * Open-Source Machine Learning Award at ODSC West 2019 for the mlr package. * Co-PI on NSF award 1923542, CS For All:RPP - Booting Up Computer Science in Wyoming ($999,929). * PI on University of Wyoming College of Engineering and Applied Sciences Engineering Initiative center seed grant ($300,000). See all NSF award 1813537, Robust Performance Models ($462,148). Outstanding PC award at AAAI 2018. €3,000 from Artificial Intelligence Journal for the ACP summer school 2018. University of Wyoming Global Engagement Office travel grant worth $2000. Best Paper Award at the Computational Intelligence and Data Mining workshop at the ITAT conference 2016. I was awarded an EPSRC Doctoral Prize. Best Student Paper Prize at the Symposium on Combinatorial Search 2011. OTHER Apart from my main affiliation, I am a research associate with the Maya Research Program. If I'm not in the office, it's possible that you can find me in the jungle of Belize excavating and/or mapping Maya ruins. Check out the interactive map. I am also involved with the OpenML project project and a core contributor to ASlib, the benchmark library for algorithm selection. While you're here, have a look at my overview of the Algorithm Selection literature. For something more visual, have a look at my pictures on Flickr.