Abstract
Introduction
Thermoacoustic oscillations pose a major challenge to designers of gas turbines, jet engines and rockets. These instabilities arise from positive feedback loops between fluctuations in the heat release rate and acoustic waves in the combustor. They may result in unacceptable levels of noise, enhanced heat transfer through the chamber walls and structural damage.
Thermoacoustic stability is exceedingly sensitive to changes in the acoustic characteristics of a combustion chamber and the acoustic/ hydrodynamic response of the flame. As a result, small changes in the system geometry, boundary conditions or flame behaviour can often stabilise an unstable combustor. However, coming up with a suitable design modification through pure trial-and-error experimentation is usually infeasible. A famous example is the F1 engine of the Saturn V rocket whose stabilization required 2000 full-scale engine tests; baffles were introduced in the injector plate to suppress thermoacoustic oscillations. 1 Adjoint-based sensitivity analysis in thermoacoustic models offers a systematic procedure for discovering appropriate design changes.
The advantage of adjoint methods lies in the efficient computation of derivatives of a given objective function at a cost that scales independently of the number of design variables. The sensitivity of eigenvalues (natural frequencies and growth rates) of a thermoacoustic model w.r.t. to design variables can therefore be computed cheaply using adjoint methods; this information is then utilized by an optimization routine from the gradient descent family to arrive at a stabilized design. Adjoint-based sensitivity analysis has been applied to low-order network models by Magri and Juniper 2 , Mensah and Moeck 3 , Silva et al. 4 , to 2D Helmholtz equation models by Falco and Juniper 5 and to 3D Helmholtz models by Mensah et al. 6 . A comprehensive review of the use of adjoints in thermoacoustics can be found in Magri 7 .
Other techniques have also been used to stabilize designs. Aguilar Perez 8 used an exhaustive grid search while Jones et al. 9 used a genetic algorithm to optimize the shape of a thermoacoustically unstable combustor. Compared to gradient descent using adjoints, these tools have the advantage of not getting trapped in local minima, but are much slower in comparison and require many expensive simulations to be run for each configuration they evaluate.
In this paper, we propose using gradient-augmented Bayesian optimization (BO) to optimize adjoint models. BO is able to find global optima in expensive, multi-modal functions with strikingly few function evaluations. It does so by building a Gaussian Process (GP) metamodel of the objective function at every iteration. To decide where to evaluate the objective function next, an acquisition function such as expected improvement,
10
upper confidence bound
11
or knowledge gradient
12
is used. This balances exploration (i.e., moving to places where the GP has high predictive uncertainty) and exploitation (i.e., moving to where the GP has a high predictive mean). Vanilla BO, however, does not utilize derivative information. Nevertheless, since the derivative of a GP with a twice-differentiable kernel is also a GP,
13
it is possible to model derivatives in a GP metamodel. The incorporation of derivative observations in the GP model leads to lower predictive variances (Figure 1 shows a simple example of a 1D GP with and without derivatives) and makes our exploration of the search space much more efficient. The main challenge is the impractical

A 1D Gaussian Process metamodel with and without derivatives. Including derivative information results in a GP which has a higher confidence in its predictions. The gray area represents 3 s.d. uncertainty bounds in the Gaussian process metamodel, the red crosses are function evalulations, the red lines denote function and derivative evaluations and the dotted blue lines are randomly sampled functions from the Gaussian process metamodel.
To demonstrate the effectiveness of gradient-augmented Bayesian optimization, we consider two optimization problems from the literature on adjoint models in thermoacoustics. The first one is a simple toy problem 15 where the iris diameter and heater position in a 1D Helmholtz equation model of a Rijke tube must be optimized. The second involves geometry optimization in a low-order network model of a longitudinal combustor. 16 We compare gradient-augmented BO with BFGS, a standard quasi-Newton optimizer which uses an estimate of the inverse Hessian matrix to improve convergence. Compared to BFGS, we find that the gradient-augmented BO does not get stuck in local optima and requires fewer evaluations of the solver to arrive at more thermacoustically stable configurations. It is also able to efficiently explore the whole design space and locate multiple suitable combinations of design parameters, which is useful further downstream in the design cycle.
Methods
Thermoacoustic models
Our goal is to design systems in which all eigenmodes decay in time, so we use the summed exponential of the growth rates as our objective function. This strongly penalizes positive growth rates, but the rewards for stabilizing the system diminish progressively as growth rates become less than zero. If
The first test case considered in this paper comes from Juniper 15 where optimization using adjoints is demonstrated on a 1D Helmholtz model of a Rijke tube. A finite element discretization of the weak form is used and the discrete adjoint framework is used to compute eigenvalue sensitivities. The design parameters are the heater position and the diameter of a variable-diameter iris placed at the downstream boundary. The growth rate of the fundamental mode, which we seek to minimize, has two local minima (Figure 2) in the search space.

Contour plots of non-dimensionalized growth rate and frequency of the Rijke tube fundamental mode plotted against heater position and iris diameter, reproduced from. 15 The numbered black contour lines correspond to frequencies and the colorued contour plot represents the growth rate. White arrows indicate the direction and magnitude of the growth rate gradients, as computed by the adjoint model.
The second test case is taken from Aguilar and Juniper
16
and relates to geometry optimization in a network model of Rama Balachandran’s 10 kW longitudinal combustor built in Cambridge, originally intended for the experimental investigation of the response of turbulent premixed flames to acoustic oscillations.
17
The geometry consists of an inlet duct connected to a plenum with a linearly varying cross section on either end. This leads to the neck which contains the fuel injection plane and a centred bluff body used to stabilize the flame. The outlet is a cylindrical pipe which contains the flame. Figure 3 shows a schematic of the combustor model and the geometric parameters being optimized. The rig is modeled using a one-dimensional network model with 124 straight ducts. This model has 4 unstable eigenmodes between 0 and 1000 Hz that need to be stabilized with geometry modifications. We restrict the search space to a box bounded by

Geometric parameters of the network model from 8 used in the optimization routine. The parameters in black are allowed to be updated. The parameters in red are kept constant.
For more details on the test cases, we refer the interested reader to the original papers.
Gradient-augmented Bayesian optimization
Bayesian optimization is a powerful tool but the inclusion of
Once we have the GP surrogate model of our function, we use its mean and uncertainty to compute the expected improvement acquisition function
19
which helps us decide where best to sample the function and its derivative in each iteration. Expected improvement is defined as
The parameter
In most high-dimensional optimization problems of practical significance, there exist only a few relevant directions that capture most of the variation in the function. This is exploited by the active subspace dimension reduction method.
21
The optimal subspace is given by the dominant eigenvectors of the covariance matrix
Results
For the 1D Rijke tube problem, the gradient-augmented BO consistently finds the known global optimum near

Surrogate Gaussian process model of the Rijke tube cost function at different iterations. Black dots indicate where the Bayesian optimization has evaluated the function and its gradients. The bottom right plot is a contour plot of the true cost function, with the black dot indicating the location of the true global minimum.
To compare between gradient-augmented BO and BFGS, we conduct 10 trials of both algorithms on the longitudinal combustor test problem. In each trial, the initial points are sampled uniformly from the design space with a different seed for the pseudo-random number generator. When the number of iterations is very small, the BFGS appears to outperform Bayesian optimization, but this is because Bayesian optimization explores more early on due to the high initial uncertainty in the Gaussian Process surrogate. BO quickly overtakes the BFGS average, however, and by the 50th iteration, 9 out of 10 BO runs have found a configuration as stable as the best BFGS run. The final average cost function after 50 runs is 1.62 for BFGS and 1.01 for BO. As in the Rijke tube case, multiple distinct stable geometries are found in each run. In general, stable configurations have a reduced flame holder area and a reduced neck area, because the eigenvalues are quite sensitive to these two parameters. Figure 5 shows a stable geometry found by one of the BO runs.

A stable combustor configuration, shown in red, found by a gradient-augmented Bayesian optimization run after 50 evaluations. Original unstable configuration in black.
Figure 6 also shows the comparison between the gradient-augmented BO to BO without gradient values. As expected, it converges to the global optimum more slowly. This shows the benefit of having derivative information from the adjoint models.

Best objective function values achieved by the BFGS, BO without gradient and gradient-augmented BO routines in the longitudinal combustor test case, averaged across 10 trials, plotted against the number of cost function evaluations. Errorbars indicate 1 s.d. variation across trials.
To be fair to BFGS, it should be pointed out the BO involves a modest computational overhead whereas BFGS has nearly none. For the longitudinal combustor model, it takes around 30 seconds per iteration for the gradient-augment BO on a Dell G7 7590 laptop with a 4 core 8th Generation Intel i5-8300HQ processor. However, when optimizing expensive models with this algorithm, the evaluation of the cost function and its gradient becomes the computational bottleneck while the cost per BO iteration becomes a non-issue.
Conclusions
In this study, we demonstrate that scalable, gradient-augmented BO pairs perfectly with adjoint methods in thermoacoustics. We use two optimization problems from the literature as our test cases: a 1D Helmholtz equation model of a Rijke tube and a low-order network model of a longitudinal combustor. We showed that the BO algorithm consistently finds more stable parameter combinations using fewer iterations than BFGS, a popular quasi-Newton optimizer. It also builds a metamodel of the objective function in the entire design space and explores multiple feasible designs. The metamodel can naturally be used for uncertainty quantification as well.
There are several potential avenues for future work. While this paper illustrates the untapped potential of gradient-augmented BO in thermoacoustics and other engineering disciplines where adjoint models are commonly used, the models we used in our test cases were relatively cheap to evaluate. BO really shines when the objective is expensive to evaluate, so a truly convincing use case would involve more expensive models, e.g. 3D adjoint Helmholtz solvers and a complex combustor geometry with many hundreds of parameters. Another interesting line of research would be looking at different acquisition functions in the BO routine. One acquisition function of particular interest is Noisy-Input Entropy Search 23 which searches for robust optima that are less sensitive to noise in their input. Since thermoacoustic stability is highly sensitive to small perturbations in model parameters and geometry, finding more robust optima may lead to better designs.
