Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions

Gradient variance errors in gradient-based search methods are largely mitigated using momentum, however the bias gradient errors may fail the numerical search methods in reaching the true optimum. We investigate the reduction in both bias and variance errors attributed to the simplex gradient estima...

Full description

Saved in:
Bibliographic Details
Main Authors: Mircea-Bogdan Radac, Titus Nicolae
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10843234/
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832586880881786880
author Mircea-Bogdan Radac
Titus Nicolae
author_facet Mircea-Bogdan Radac
Titus Nicolae
author_sort Mircea-Bogdan Radac
collection DOAJ
description Gradient variance errors in gradient-based search methods are largely mitigated using momentum, however the bias gradient errors may fail the numerical search methods in reaching the true optimum. We investigate the reduction in both bias and variance errors attributed to the simplex gradient estimated from noisy function measurements, in favor of the finite-differences gradient, when both are used for black-box optimization methods. Regardless of the simplex orientation, while reducing the gradient bias error owned to several factors such as truncation, numerical or measurement noise, we claim and verify that, under relaxed assumptions about the underlying function’s differentiability, the estimated gradient by the simplex method has at most half the variance of the finite-difference gradient. The findings are validated with two comprehensive and representative case studies, one related to the minimization of a nonlinear feedback control system cost function and the second related to a deep machine learning classification problem whose hyperparameters are tuned. We conclude that in up to medium-size practical black-box optimization problems with unknown variable domains and where the noisy function measurements are expensive, a simplex gradient-based search is an attractive option.
format Article
id doaj-art-aba0787d35f14259acb177dd3c30cadb
institution Kabale University
issn 2169-3536
language English
publishDate 2025-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj-art-aba0787d35f14259acb177dd3c30cadb2025-01-25T00:02:48ZengIEEEIEEE Access2169-35362025-01-0113143041431610.1109/ACCESS.2025.352991510843234Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly FunctionsMircea-Bogdan Radac0https://orcid.org/0000-0001-8410-6547Titus Nicolae1https://orcid.org/0009-0006-5757-5464Department of Automation and Applied Informatics, Politehnica University of Timisoara, Timisoara, RomaniaSmilecloud, United Business Center 0, Piaţa Consiliul Europei, Timisoara, RomaniaGradient variance errors in gradient-based search methods are largely mitigated using momentum, however the bias gradient errors may fail the numerical search methods in reaching the true optimum. We investigate the reduction in both bias and variance errors attributed to the simplex gradient estimated from noisy function measurements, in favor of the finite-differences gradient, when both are used for black-box optimization methods. Regardless of the simplex orientation, while reducing the gradient bias error owned to several factors such as truncation, numerical or measurement noise, we claim and verify that, under relaxed assumptions about the underlying function’s differentiability, the estimated gradient by the simplex method has at most half the variance of the finite-difference gradient. The findings are validated with two comprehensive and representative case studies, one related to the minimization of a nonlinear feedback control system cost function and the second related to a deep machine learning classification problem whose hyperparameters are tuned. We conclude that in up to medium-size practical black-box optimization problems with unknown variable domains and where the noisy function measurements are expensive, a simplex gradient-based search is an attractive option.https://ieeexplore.ieee.org/document/10843234/Black-boxcontrol systemsdeep learningfinite differencesmachine learningoptimization
spellingShingle Mircea-Bogdan Radac
Titus Nicolae
Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
IEEE Access
Black-box
control systems
deep learning
finite differences
machine learning
optimization
title Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
title_full Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
title_fullStr Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
title_full_unstemmed Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
title_short Leveraging Simplex Gradient Variance and Bias Reduction for Black-Box Optimization of Noisy and Costly Functions
title_sort leveraging simplex gradient variance and bias reduction for black box optimization of noisy and costly functions
topic Black-box
control systems
deep learning
finite differences
machine learning
optimization
url https://ieeexplore.ieee.org/document/10843234/
work_keys_str_mv AT mirceabogdanradac leveragingsimplexgradientvarianceandbiasreductionforblackboxoptimizationofnoisyandcostlyfunctions
AT titusnicolae leveragingsimplexgradientvarianceandbiasreductionforblackboxoptimizationofnoisyandcostlyfunctions