dc.date.accessioned | 2021-10-04T23:00:58Z | |
dc.date.available | 2021-10-04T23:00:58Z | |
dc.date.created | 2021-10-04T23:00:58Z | |
dc.date.issued | 2021 | |
dc.identifier | https://hdl.handle.net/20.500.12866/9838 | |
dc.identifier | https://doi.org/10.1007/s10462-021-10011-5 | |
dc.description.abstract | SVM with an RBF kernel is usually one of the best classification algorithms for most data sets, but it is important to tune the two hyperparameters C and γ to the data itself. In general, the selection of the hyperparameters is a non-convex optimization problem and thus many algorithms have been proposed to solve it, among them: grid search, random search, Bayesian optimization, simulated annealing, particle swarm optimization, Nelder Mead, and others. There have also been proposals to decouple the selection of γ and C. We empirically compare 18 of these proposed search algorithms (with different parameterizations for a total of 47 combinations) on 115 real-life binary data sets. We find (among other things) that trees of Parzen estimators and particle swarm optimization select better hyperparameters with only a slight increase in computation time with respect to a grid search with the same number of evaluations. We also find that spending too much computational effort searching the hyperparameters will not likely result in better performance for future data and that there are no significant differences among the different procedures to select the best set of hyperparameters when more than one is found by the search algorithms | |
dc.language | eng | |
dc.publisher | Springer | |
dc.relation | Artificial Intelligence Review | |
dc.relation | 1573-7462 | |
dc.rights | https://creativecommons.org/licenses/by-nc-nd/4.0/deed.es | |
dc.rights | info:eu-repo/semantics/restrictedAccess | |
dc.subject | Bayesian optimization | |
dc.subject | Classification (of information) | |
dc.subject | Classification algorithm | |
dc.subject | Computation time | |
dc.subject | Computational effort | |
dc.subject | Convex optimization | |
dc.subject | Empirical evaluations | |
dc.subject | Grid search | |
dc.subject | Hyperparameters | |
dc.subject | Learning algorithms | |
dc.subject | Non-convex optimization algorithms | |
dc.subject | Nonconvex optimization | |
dc.subject | Particle swarm optimization (PSO) | |
dc.subject | Parzen estimators | |
dc.subject | Random search | |
dc.subject | Search Algorithms | |
dc.subject | Simulated annealing | |
dc.subject | Support vector machines | |
dc.subject | SVM | |
dc.title | How to tune the RBF SVM hyperparameters? An empirical evaluation of 18 search algorithms | |
dc.type | info:eu-repo/semantics/article | |