Understanding the Effect of Hyperparameter Optimization on Machine Learning Models for Structure Design Problems. (June 2021)
- Record Type:
- Journal Article
- Title:
- Understanding the Effect of Hyperparameter Optimization on Machine Learning Models for Structure Design Problems. (June 2021)
- Main Title:
- Understanding the Effect of Hyperparameter Optimization on Machine Learning Models for Structure Design Problems
- Authors:
- Du, Xianping
Xu, Hongyi
Zhu, Feng - Abstract:
- Abstract: To relieve the computational cost of design evaluations using expensive finite element (FE) simulations, surrogate models have been widely applied in computer-aided engineering design. Machine learning algorithms (MLAs) have been implemented as surrogate models due to their capability of learning the complex interrelations between the design variables and the response from big datasets. Typically, an MLA regression model contains model parameters and hyperparameters. The model parameters are obtained by fitting the training data. Hyperparameters, which govern the model structures and the training processes, are assigned by users before training. There is a lack of systematic studies on the effect of hyperparameters on the accuracy and robustness of the surrogate model. In this work, we proposed to establish a hyperparameter optimization framework to deepen our understanding of the effect. Based on the sequential model-based optimization method, the Pareto front is generated by running the optimal acquisition and updating the surrogate model iteratively. The optimum acquisition works by repeating a design space shrinking process. Using the acquired optimum, the surrogate model is updated, which describes the relationship between the hyperparameter combinations (inputs) generated by Latin hypercube sampling from the design space and structural response (outputs) to evaluate the modeling accuracy. The updated model will then be used for the next iteration of optimalAbstract: To relieve the computational cost of design evaluations using expensive finite element (FE) simulations, surrogate models have been widely applied in computer-aided engineering design. Machine learning algorithms (MLAs) have been implemented as surrogate models due to their capability of learning the complex interrelations between the design variables and the response from big datasets. Typically, an MLA regression model contains model parameters and hyperparameters. The model parameters are obtained by fitting the training data. Hyperparameters, which govern the model structures and the training processes, are assigned by users before training. There is a lack of systematic studies on the effect of hyperparameters on the accuracy and robustness of the surrogate model. In this work, we proposed to establish a hyperparameter optimization framework to deepen our understanding of the effect. Based on the sequential model-based optimization method, the Pareto front is generated by running the optimal acquisition and updating the surrogate model iteratively. The optimum acquisition works by repeating a design space shrinking process. Using the acquired optimum, the surrogate model is updated, which describes the relationship between the hyperparameter combinations (inputs) generated by Latin hypercube sampling from the design space and structural response (outputs) to evaluate the modeling accuracy. The updated model will then be used for the next iteration of optimal acquisition until the termination criterion is met. Four frequently used MLAs, namely Gaussian Process Regression (GPR), Support Vector Machine (SVM), Random Forest Regression (RFR), and Artificial Neural Network (ANN), are tested on four benchmark examples of structure design optimization. For each MLA model, the model accuracy and robustness before and after the hyperparameters optimization (HOpt) are compared. The results show that HOpt can generally improve the performance of the MLA models in general with dependency on model complexity. HOpt leads to unstable improvements in the MLAs accuracy and robustness for complex problems, which are featured by high-dimensional mixed-variable design space. We also investigated the additional computational costs incurred by HOpt. The training cost is closely related to the MLA architecture. After HOpt, the training cost of ANN and RFR is increased more than that of the GPR and SVM. In summary, this study benefits the selection of HOpt method for different types of design problems based on their complexity (i.e. design domain continuity and the number of design variables, etc.). Highlights: A framework was established for optimizing the machine learning-based surrogate models in structural design. The impact of HOpt on the performance of the surrogate models was investigated using benchmark problems. A parametric study was conducted on the hyperparameters' influence on modeling performance. … (more)
- Is Part Of:
- Computer aided design. Volume 135(2021)
- Journal:
- Computer aided design
- Issue:
- Volume 135(2021)
- Issue Display:
- Volume 135, Issue 2021 (2021)
- Year:
- 2021
- Volume:
- 135
- Issue:
- 2021
- Issue Sort Value:
- 2021-0135-2021-0000
- Page Start:
- Page End:
- Publication Date:
- 2021-06
- Subjects:
- Structure design -- Surrogate models -- Machine learning -- Hyperparameters optimization
Computer-aided design -- Periodicals
Engineering design -- Data processing -- Periodicals
Computer graphics -- Periodicals
Conception technique -- Informatique -- Périodiques
Infographie -- Périodiques
Computer graphics
Engineering design -- Data processing
Periodicals
Electronic journals
620.00420285 - Journal URLs:
- http://www.journals.elsevier.com/computer-aided-design/ ↗
http://www.elsevier.com/journals ↗ - DOI:
- 10.1016/j.cad.2021.103013 ↗
- Languages:
- English
- ISSNs:
- 0010-4485
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 3393.520000
British Library DSC - BLDSS-3PM
British Library STI - ELD Digital store - Ingest File:
- 16329.xml