Emergent Mind

Abstract

We present a flexible trust region descend algorithm for unconstrained and convexly constrained multiobjective optimization problems. It is targeted at heterogeneous and expensive problems, i.e., problems that have at least one objective function that is computationally expensive. The method is derivative-free in the sense that neither need derivative information be available for the expensive objectives nor are gradients approximated using repeated function evaluations as is the case in finite-difference methods. Instead, a multiobjective trust region approach is used that works similarly to its well-known scalar pendants. Local surrogate models constructed from evaluation data of the true objective functions are employed to compute possible descent directions. In contrast to existing multiobjective trust region algorithms, these surrogates are not polynomial but carefully constructed radial basis function networks. This has the important advantage that the number of data points scales linearly with the parameter space dimension. The local models qualify as fully linear and the corresponding general scalar framework is adapted for problems with multiple objectives. Convergence to Pareto critical points is proven and numerical examples illustrate our findings.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.