Abstract
The training error of Machine Learning (ML) methods has been extensively used for performance assessment, and its low values have been used as a main justification for complex methods such as estimator fusion and ensembles, and hyper parameter tuning. We present two practical cases where independent tests indicate that the low training error is more of a reflection of over-fitting rather than the generalization ability. We derive a generic form of the generalization equations that separates the training error terms of ML methods from their epistemic terms that correspond to approximation and learnability properties. It provides a framework to separately account for both terms to ensure an overall high generalization performance. For regression estimation tasks, we derive conditions for performance enhancements achieved by hyper parameter tuning, and fusion and ensemble methods over their constituent methods. We present experimental measurements and ML estimates that illustrate the analytical results for the throughput profile estimation of a data transport infrastructure.
Original language | English |
---|---|
Title of host publication | 2023 26th International Conference on Information Fusion, FUSION 2023 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9798890344854 |
DOIs | |
State | Published - 2023 |
Event | 26th International Conference on Information Fusion, FUSION 2023 - Charleston, United States Duration: Jun 27 2023 → Jun 30 2023 |
Publication series
Name | 2023 26th International Conference on Information Fusion, FUSION 2023 |
---|
Conference
Conference | 26th International Conference on Information Fusion, FUSION 2023 |
---|---|
Country/Territory | United States |
City | Charleston |
Period | 06/27/23 → 06/30/23 |
Funding
This research is sponsored in part by RAMSES project of Advanced Scientific Computing Research program, U.S. Department of Energy, and in part by the Office of Basic Energy Sciences, Division of Materials Sciences and Engineering, U.S. Department of Energy, and is performed at Oak Ridge National Laboratory managed by UT-Battelle, LLC for U.S. Department of Energy under Contract No. DE-AC05-00OR22725. The United States Government retains and the publisher, by accepting the article for publication, acknowledges that the United States Government retains a nonexclusive, paid-up, irrevocable, world-wide license to publish or reproduce the published form of this manuscript, or allow others to do so, for United States Government purposes. The Department of Energy will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).
Keywords
- fusion and ensemble
- generalization bounds
- hyper-parameter tuning
- machine learning
- over-fitting
- regression
- throughput profile