Goal: minimize unknown function ϕ in as few evaluations as possible
Modeling:
Decision-making:
Joint work with Qian Xie, Raul Astudillo, Peter Frazier, and Ziv Scully
Goal: minimize unknown function ϕ in as few evaluations as possible
Algorithm:
Optimal choice xt+1 and when to stop: intractable dynamic program
Cost-aware baseline: expected improvement per unit cost αtEIPC(x)=c(x)EIf∣y1,..,yt(x;max1≤τ≤tyτ)EIψ(x;y)=Emax(0,ψ(x)−y)
Invariant to multiplying costs by a scalar: do we want this behavior?
Cost-aware analog of expected improvement:
Assumptions:
These are restrictive! But they lead to an interesting, general solution
Theorem (Weitzman, 1979). Let:
Then, for the cost-per-sample problem, the policy defined by maximizing the Gittins index acquisition function α⋆ with its associated stopping rule is Bayesian-optimal.
Gittins index α⋆: optimal for cost-per-sample problem
Theorem. Assume the expected budget constraint is feasible and active. Then there exists a λ>0 and a tie-breaking rule such that the policy defined by maximizing the Gittins index acquisition function α⋆(⋅), defined using costs λc(x), is Bayesian-optimal.
Proof idea: Lagrangian duality
Our work: extends special case of a result of Aminian et al. (2024) to non-discrete reward distributions
Bayesian optimization: posterior distribution is correlated
Define Pandora's Box Gittins Index acquisition function:
αPBGI(x)=g where g solves EIf∣y(x;g)=c(x)
Correlations: incorporated into acquisition function via the posterior
Novel acquisition function: Pandora's Box Gittins Index
Performance of Pandora's Box Gittins Index
Gittins Index Theory
Bayesian Optimization: high-dimension and complex feedback models
Q. Xie, R. Astudillo, P. Frazier, Z. Scully, and A. Terenin. Cost-aware Bayesian optimization via the Pandora's Box Gittins index. NeurIPS, 2024.