1 code implementation • 2 Jul 2022 • Alexander Thebelt, Calvin Tsay, Robert M. Lee, Nathan Sudermann-Merx, David Walz, Behrang Shafei, Ruth Misener
Tree ensembles can be well-suited for black-box optimization tasks such as algorithm tuning and neural architecture search, as they achieve good predictive performance with little or no manual tuning, naturally handle discrete feature spaces, and are relatively insensitive to outliers in the training data.
1 code implementation • 4 Feb 2022 • Francesco Ceccon, Jordan Jalving, Joshua Haddad, Alexander Thebelt, Calvin Tsay, Carl D. Laird, Ruth Misener
The optimization and machine learning toolkit (OMLT) is an open-source software package incorporating neural network and gradient-boosted tree surrogate models, which have been trained using machine learning, into larger optimization problems.
no code implementations • 25 Jan 2022 • Alexander Thebelt, Johannes Wiebe, Jan Kronqvist, Calvin Tsay, Ruth Misener
For each of these four data characteristics, we discuss applications where these data characteristics arise and show how current chemical engineering research is extending the fields of data science and machine learning to incorporate these challenges.
1 code implementation • 4 Nov 2021 • Alexander Thebelt, Calvin Tsay, Robert M. Lee, Nathan Sudermann-Merx, David Walz, Tom Tranter, Ruth Misener
Energy systems optimization problems are complex due to strongly non-linear system behavior and multiple competing objectives, e. g. economic gain vs. environmental impact.
1 code implementation • NeurIPS 2021 • Calvin Tsay, Jan Kronqvist, Alexander Thebelt, Ruth Misener
This paper introduces a class of mixed-integer formulations for trained ReLU neural networks.
1 code implementation • 10 Mar 2020 • Alexander Thebelt, Jan Kronqvist, Miten Mistry, Robert M. Lee, Nathan Sudermann-Merx, Ruth Misener
Gradient boosted trees and other regression tree models perform well in a wide range of real-world, industrial applications.