Automated Machine Learning (AutoML) systems are an important part of the machine learning engineering toolbox. However, existing solutions often require substantial amounts of compute, and therefore extended running times, to achieve competitive performance.
In this post, we evaluate Plexe, an AutoML framework designed to deliver competitive or superior predictive performance while demanding fewer computational resources compared to other popular frameworks. We use as our baseline performance results published in the 2024 paper by Gijsbers et al. In this post, we frequently refer to their methodology and findings, so we encourage you to check out their work.
Across a set of 20 benchmark datasets, relative to the AutoML tools we are evaluating against, Plexe consistently achieves superior or competitive prediction accuracy, while converging to a solution in a shorter wall time on comparable hardware.
While we are not yet ready to release Plexe itself to the world, the benchmark experiments themselves are fully reproducible, with code and data publicly accessible at GitHub – plexe-ai/plexe-results. We plan to release Plexe in the coming months.
We compare against leading AutoML frameworks:
We evaluate Plexe following the same protocol described in the paper by Gijsbers et al:
All experiments were conducted on standardized cloud infrastructure:
This is comparable to the hardware setup used by Gijsbers et al in their paper.
Results across the OpenML benchmark suite show consistent performance improvements:
Dataset | Size | Features | Domain | Plexe | AutoGluon(B) | H2O AutoML | LightAutoML | FLAML | AUTO-SKLEARN | MLJAR(B) | TPOT |
---|---|---|---|---|---|---|---|---|---|---|---|
kr-vs-kp | 3,196 | 36 | Games (Chess) | 1.000 | 1.000 | 1.000 | 1.000 | 0.961 | 1.000 | 0.999 | |
adult | 48,842 | 14 | Census | 0.933 | 0.932 | 0.931 | 0.932 | 0.932 | 0.930 | 0.927 | |
bank-marketing | 45,211 | 16 | Finance | 0.938 | 0.941 | 0.938 | 0.940 | 0.937 | 0.939 | 0.935 |
Note: *Steel plates uses logloss metric (lower is better)’-‘ indicates framework failed to complete or produce valid results
In the AutoML benchmark experiment, while a one-hour budget was set for each framework’s training, not all frameworks consistently used the full time. In fact, there were instances where frameworks finished well before the allotted hour, indicating variability in processing time requirements across frameworks and datasets. Additionally, certain frameworks occasionally exceeded the time budget by a few minutes, though significant overruns were rare and generally controlled within an acceptable range of leniency. In Plexe’s case, all best solutions were found well within the 1 hour limit, with 75% finishing in less than 30 minutes.
Performance on Kaggle competitions provides independent validation and an example for a situation where Plexe can be immediately useful:
Domain | Competition | Plexe’s Percentile Rank | Plexe’s absolute Rank |
---|---|---|---|
Finance | Loan Approval Prediction | 73.14% | 765/2845* |
Real Estate | House Price Prediction | 83.12% | 800/4735* |
Analytics | Titanic Survival Prediction | 85.73% | 2200/15422* |
Note: * represents competition ongoing at the time of submission
Side note: Test set accuracy for the Titanic Survival dataset reported by MLJar (running 4 hours job on m5.24xlarge machine (96CPU, 384GB RAM)): 77.99% whereas Plexe reported an accuracy of 84.51%.(with a 1 hour limit running on the same hardware as defined on the top of this blog post)
While these early results are only indicative, we do find them encouraging. In particular, we observed:
We acknowledge that the experiment has a number of limitations, which constrains the conclusions we can draw:
Nonetheless, we feel these initial results warrant further work on the tool. There are several directions that we want to explore:
Our empirical evaluation of Plexe’s early prototype suggests that the tool is capable of achieving competitive or superior performance compared to current AutoML frameworks, while reducing computational requirements. This provides an encouraging foundation for us to continue building upon, and we plan to release a fully fledged beta to the public in the coming months. Stay tuned for updates!
Gijsbers, P., Bueno, J., Coors, S., LeDell, E., Poirier, S., Thomas, J., Bischl, B., & Vanschoren, J. (2023). AMLB: An AutoML Benchmark. 22-0493.pdf, benchmarking results and framework evaluation.
Source: Plexe: Production-ready custom AI from natural language