Sequential dimension reduction for learning features of expensive black-box functions

Abstract : Learning a feature of an expensive black-box function (optimum, contour line,...) is a difficult task when the dimension increases. A classical approach is two-stage. First, sensitivity analysis is performed to reduce the dimension of the input variables. Second, the feature is estimated by considering only the selected influential variables. This approach can be computationally expensive and may lack flexibility since dimension reduction is done once and for all. In this paper, we propose a so called Split-and-Doubt algorithm that performs sequentially both dimension reduction and feature oriented sampling. The 'split' step identifies influential variables. This selection relies on new theoretical results on Gaussian process regression. We prove that large correlation lengths of covariance functions correspond to inactive variables. Then, in the 'doubt' step, a doubt function is used to update the subset of influential variables. Numerical tests show the efficiency of the Split-and-Doubt algorithm.
Document type :
Preprints, Working Papers, ...
Complete list of metadatas

Cited literature [34 references]  Display  Hide  Download


https://hal.archives-ouvertes.fr/hal-01688329
Contributor : Malek Ben Salem <>
Submitted on : Wednesday, February 27, 2019 - 11:09:48 AM
Last modification on : Monday, January 20, 2020 - 12:14:05 PM
Long-term archiving on: Tuesday, May 28, 2019 - 12:18:16 PM