Space mapping-based Receding Horizon Control for Stochastic Interacting Particle Systems: dogs herding sheep

Control of stochastic interacting particle systems is a non-trivial task due to the high dimensionality of the problem and the lack of fast algorithms. Here, we propose a space mapping-based approximation of the stochastic control problem by solutions of the deterministic one. In combination with the receding horizon control technique this yields a reliable and fast numerical scheme for the closed loop control of stochastic interacting particle systems. As a numerical example we consider the herding of sheep with dogs. The numerical results underline the feasibility of our approach and further show stabilizing behaviour of the closed loop control.


Introduction
Collective behaviour of crowds or swarms has been investigated by various researchers in the past decades [14,15,16,17,33]. First, the focus was on the simulation of large groups, like flocks of birds and schools of fish, and their attractive and repulsive self-interaction [11,18]. The resulting models are able to reflect major properties of the interaction such as flocking and the formation of mills [10]. Further, the stability of these patterns was analysed [1,12,18]. Later, the models were refined to take into account view cones or topographical aspects like walls [9,13,24]. To include a random disturbance of the individuals' behaviour one introduces an additive Brownian motion in the velocity component of the dynamics [6,28]. Mathematically, this changes the model from ordinary differential equations (ODEs) to a system of stochastic differential equations (SDEs).
Based on this knowledge, the investigation of the interaction of crowds and external agents became of interest [2,21]. In particular, the idea of controlling crowds with the help of the external agents [7,8]. The corresponding optimal control problem (OCP) is then constrained by the dynamics of the respective ODE or SDE system. For the deterministic problem one can employ standard techniques from variational calculus to derive the gradient of the cost functional and to implement a tailored iterative scheme to compute the controls.
Unfortunately, these classical methods cannot be directly adapted to the stochastic problems [34]. In fact, the stochastic influence forces the decoupled forward and backward equations of an deterministic optimal control problem, to be a fully coupled Forward-Backward SDE system involving a ghost process to capture the uncertain terminal condition, see, e.g., [20] for the derivation of such a system based on a Hamiltonian formulation. First, steps towards a numerical realization in special cases can be found in [22].
Here, we are interested in controlling the crowd over a large time horizon, such that open loop control is not appropriate. Instead we use the closed loop receding horizon control to allow for feedback during the time evolution (see also [3]). To deal with the stochastic nature of the model we employ the space mapping approach [4,36], which allows for the control of a high fidelity model (here the stochastic one) by the optimization of a surrogate model (here the deterministic one).
The space mapping approach first came up in the engineering community [5] as a tool to solve large scale optimization problems with the help of an easier surrogate model. Through the years the technique became well-established in engineering and has been also recognized by the mathematics community for various applications like radiative heat transfer, control the dispersion of particles in a fluid, dynamic compressor optimization of gas networks and optimal inflow control of transmission lines, see, e.g., [19,23,25,26,29,31,35].
Here, we consider as new application the herding of a crowd of sheep using dogs with repulsive influence on the crowd. The combination of the space mapping technique with the receding horizon control will finally allow for the construction of a tailored closed loop algorithm to control interacting stochastic particle systems.
The manuscript is organized as follows: in the next section the details of a general class optimal control problems with SDE constraints are given. Then, the space mapping approach is discussed in Section 3 and the Aggressive Monte Carlo Space Mapping Algorithm is presented. We derive the first-order optimality system of the deterministic ODE model and the gradient of the respective reduced cost functional which is needed for the numerical implementation in Section 4. The algorithms for the numerical investigation are described in Section 5. We present a projected gradient method for the deterministic optimization and discuss the receding horizon procedure for the closed loop control of the stochastic particle system. The feasibility of our approach is underlined by the numerical results presented in Section 6. We discuss a space-mapping approach based on a mean-field approximation in Section 7, before we give conclusions and an outlook in Section 8.

The Control Problem
In this section we define the general class of control problems constrained by a stochastic interacting particle system.
and combined in the vectors for each t ∈ [0, T ], respectively. In analogy, we consider M external agents having positions Later, they act as control functions. We assume u ∈ L 2 ([0, T ], R M D ) .
The self-organisation of the crowd and the interaction of the particles with the agents is modelled with the help of radially symmetric interaction potentials For the sake of well-posedness, we assume that their first and second derivative ∇ r Φ j (r) =: G j (r) and ∇ 2 r Φ j (r) =: H j (r) are locally Lipschitz and globally bounded, i.e., Φ j ∈ C 2 b (R + 0 ) for j = 1, 2. Further, we include a friction term with parameter α > 0 and additive stochastic noise with strength σ ≥ 0 influencing the velocities of the individuals. The friction models the lethargy of the individuals, while the stochasticity allows for disturbances of the surroundings, that are not considered explicitly. Let B i t , i = 1, . . . , N , denote independent D-dimensional Brownian motions. Then, the stochastic state system is given by supplemented with initial data Y 0 := (X 0 , V 0 , a 0 ). The full state is a random variable Y = (X, V, a).
Remark 2.1. Clearly, for σ = 0 the above system reduces to an ODE system, which we are going to use as the surrogate model for the space mapping procedure.
It would be interesting to generalize the approach for common noise situations, i.e., B i t = B t in order to model effects on the system as a whole, instead of single particles. This will have impacts on the cost-functional and also on the mean-field equation discussed later on. For simplicity, we restrict ourselves here to the case of individual noise.

2.2.
Well-posedness of the State Systems. Assuming a maximal velocity u max for the agents, we define the set of admissible controls Note, that the ODE for the agents can be solved explicitly for given u ∈ U ad which yields Indeed, we get an absolutely continuous function a, which can be plugged into the SDE system governing the dynamic of the crowd. Using the assumption Φ j ∈ C 2 b (R + 0 ), we obtain weak solutions of the stochastic system in the sense of Itô due to standard SDE theory, see, e.g., [32]. Further, the state fulfills Remark 2.2. Note, that the stochastic system can be generalized to space and time dependent σ = σ(x, t) without any effect on the well-posedness as long as the following conditions are satisfied for some positive constants C, D and x, y ∈ R D .
would be enough regularity to obtain the absolutely continuous function a. But to identify the gradient for the numerical algorithm later on, we need a Hilbert-space structure. Thus, we choose the stronger assumption for U ad .
In the case σ = 0, we obtain a deterministic ODE system which attains a unique solution by standard results from ODE theory. This allows us to define the control-to-state map S c which assigns to each u ∈ U ad the unique solution y of the ODE system. In analogy, we define S f (u) = Y for the solution of the the SDE system. For better readability we refer to states of the ODE system with lower-case letters and states corresponding to the SDE system with upper-case letters.
2.3. The Cost Functional. In general, cost functionals involving empirical quantities, like expectation, variance or other kind of moments of the particle crowd are appropriate for the space mapping approach.
In the following, we consider a specific cost functional that is based on the expected trajectory of the centre of mass of the crowd reflecting the aim of our application, i.e., steering the crowd to a predefined destination Z des . To do so we define a time dependent reference stateZ : [0, T ] → R D . Similar to the approach in [8], we measure the spread of the crowd aroundZ. In particular, due to the stochastic behaviour of the state system we use the expected paths E[X].
This leads to the following cost functional where the first term tracks the expected centre of mass of the crowd and penalizes its distance to the desired trajectory. The second term measures the control costs and is weighted with the parameter γ > 0.
Remark 2.4. The predefined desired trajectoryZ(t) and the reference velocities u are input parameters for the cost functional. In the space mapping procedure, Z shall be replaced by the expected centre of mass andū by the optimal control of the surrogate model.
To sum up, the SDE constrained optimal control problem of consideration is given by Remark 2.5. The existence of an optimal control can be shown with standard techniques from variational calculus [27]. In fact, an existence result can be obtained for all sequentially weak lower semicontinuous and coercive cost functionals J(Y, u). Note that, in general, we cannot expect its uniqueness due to the non-convexity which is introduced by the nonlinearity in the state system.

The Space Mapping Approach
The direct solution of this SDE constrained optimal control problem is a nontrivial task. Nevertheless, we can exploit the fact that the deterministic ODE model is for small noise σ a good approximation for the stochastic one in combination with the space mapping procedure.
The general idea of space mapping for optimization problems is to approximate a complex (fine) model by a simple (coarse) surrogate model such that its main features are still resolved and the coarse model allows for a fast optimization. In particular, no gradient information of the fine model needs to be computed. Space mapping goes back to Bandler [5] and an excellent introduction is given in the review [4] and the references therein.
Let G f and G c be two operators mapping the fine control and the coarse control to some observable, respectively. To get an approximation of the fine model optimization for a desired valuew, one uses optimizers of the coarse model, i.e., For a better approximation the space mapping function is introduced, which assigns to an input u f of the fine model a control u c for the coarse model, yielding the best approximation of the fine model output G f (u f ) by the coarse model output G c (u c ).
If the observables to the respective optimizers are similar, i.e., Indeed, the space mapping is fixed by the observable defined by the operators G f and G c . In the following we use and compute E[X] with the help of a Monte Carlo simulation as proposed in [31]. Therefore, it makes sense to setw = 0. That means, in the following, the stochastic interacting particle system (σ > 0) will act as the fine model, while the coarse model is given by the deterministic particle system (σ = 0). Another possible choice could use the solution operators of the state equations, i.e., G f = S f and G c = S c withw being an desired state.
Remark 3.1. Note, that the space mapping function T might be formally set valued if the optimization problem admits multiple solutions. Assumptions on the models ensuring that T is well defined are discussed in detail in [19,26].
In general, the space mapping function T is directly not accessible, such that there are several approximations proposed in the literature [5,19,4]. These update the controls of the fine models iteratively. For example, Aggressive Space Mapping (ASM) and Trust Region Aggressive Space Mapping (TRASM) borrow the idea from quasi-Newton methods to approximate the Jacobian with the help of Broyden-type matrices. On the other hand, Hybrid Aggressive Space Mapping (HASM) combines the classical space mapping method with classical optimization techniques (cf. [4]).
We use the ASM approach for the numerical computations below. Hence, the update h k for the next iterate is given by where B k is the k-th Broyden matrix iterate and ρ > 0 the step-length.
For a smooth presentation of the algorithm, we define the expected centre of mass of the stochastic particle crowd as The resulting Aggressive Monte Carlo Space Mapping (AMCSM) approach [31] is stated in all details in Algorithm 1.
For the present application of dogs herding sheep, we need just one solve of the fine stochastic model, which involves the expensive Monte Carlo simulation in each step of the algorithm. The optimization step is only involving the coarse deterministic model, for which fast numerical algorithms based on gradient information are available. Clearly, for N 1, the determinsitic problem is still challenging, one idea is to use a mean-field appoximation in this case. See the discussion on the mean-field limit in Section 7. Further note, that the space-mapping approach discussed here does in general not yield a perfect space mapping, such that the algorithm might terminate with a suboptimal solution (c.f. [19]). This does not matter in our case, since we are designing a closed loop control with the help of the receding horizon control technique. The numerical results below indicate that the method proposed here, works fine for the problem at hand. Nevertheless, for other problems with short time horizons the space mapping solutions may fail to be robust. A qualitative study of the approximation and the robustness are subject to future work.

Optimal Control of the Coarse Model
The core of the space mapping approach is the fast optimization of the coarse model. Since we intend to use a steepest descent algorithm, we derive the firstorder optimality conditions for the coarse optimization problem. The derived adjoint information can then be used for the evaluation of the gradient of the reduced cost functional.

First-Order Optimality Condition.
For the deterministic optimal control problem with ODE constraints we can derive the adjoint system and the optimality condition with the help of the extended Lagrangian. Note, that the calculations are very similar to [8]. The deterministic system d dt can be compactly denoted by d dt y = F (y, u) , supplemented with the initial conditions y(0) = y 0 .
We define the set of controls U and the state space Y as Obviously it holds U ad ⊂ U.
As usual the first-order optimality condition of the coarse problem is given by dL(y, u, ξ, η;Z,ū) = 0.
Following the standard approach from variational calculus for the derivation of the adjoint equations (cf. [27]), we obtain the following first order optimality system.
Remark 4.2. The variational inequality in (5) can be derived as well with the help of the Pontryagin maximum principle. In view of the numerical implementation, the inequality is not handy. We therefore choose the Lagrangian approach here, leading to explicit expressions for the adjoint which can be used in the algorithm.
Together with a projection onto the feasible set U ad we can design a projected gradient-descent method for the optimization problem, see Algorithm 2.

Gradient of the Reduced Cost Functional.
In this section we introduce the reduced cost functional for the coarse model constraint and formally calculate its gradient which we need for the descent algorithm. Using the control-to-state map S c we define the reduced cost functional aŝ Since U is a Hilbert space, we may use the Riesz representation theorem to identify the gradient of the reduced cost functional as Now, we have all ingredients at hand to state the gradient descent method for the numerical simulations.

Numerical Schemes
The Aggressive Monte Carlo Space Mapping algorithm (AMCSM) proposed in Algorithm 1 uses solutions of the coarse optimal control problem and only evaluations of the fine stochastic particle system.

Optimization Algorithm for the Coarse Model.
We solve the deterministic ODE systems of state and adjoint problem with the explicit Euler scheme. In the optimal control loop for deterministic problem, we update the controls using nonlinear conjugate gradient (NCG) steps.
The step size for the gradient update is obtained by a line search based on the Armijo rule with projection (cf. [27]).
These ingredients define the numerical scheme for the deterministic optimization stated in Algorithm 2, where we denote by u n the control of the n-th optimization iteration. When the optimal solution of the coarse problem u * c is found, where the x i refer to the optimal positions extracted from S c (u * c ).

Algorithm 2: Optimal Control Algorithm for the Coarse Problem
Data: initial data for states and control, stopping tolerance opt , time steps K, desired destination Z * Result: optimal control u, optimal states y initialize; while u n+1 − u n L 2 > opt do solve deterministic state system (4); solve adjoint problem given in (6); compute gradient corresponding to (7); compute step size using the Armijo rule with projection; update controls by nonlinear conjugate gradient; end In our particular case, the projection P U has the explicit representation otherwise, for m = 1, . . . , M and t ∈ [0, T ].

5.2.
Receding Horizon Control. The appropriate time horizon for steering the crowd to the given destination depends on the distance of the crowd to the destination and might be large. Since the space mapping procedure is based on optimal controls, we need to store the full forward information to compute the adjoints. On large time intervals this leads to an extensive memory consumption. Additionally, having the application of dogs herding sheep in mind, an open loop approach is rather unrealistic. In reality, a dog will react on the current state of the crowd, such that it makes more sense to model the problem using a closed loop ansatz. This is why a closed loop control for a large time horizon is preferable. Now, we are going to combine the above numerical approaches with the receding horizon control [3]. In more detail, we split the time interval of interest [0, T ] into K smaller intervals I 1 , . . . , I K . Then, we apply the space mapping algorithm to these smaller intervals. In fact, we compute the stochastic output by an Euler-Mayurama scheme on I 1 but store only the first half of the solution. Then, we initialize the values using the optimal values at time t = I 1 /2 and compute the solution on the interval [ I 1 2 , I 2 2 ] and glue half of this solution to the one stored before. After two steps, we have the optimal control on the full interval I 1 available. We proceed iteratively until we reach the terminal time T . The receding horizon procedure is visualized in Figure 1. Note that here is some freedom in choosing the length of the smaller interval. Numerical studies motivated us to use I k 2 . Figure 1. Visualization of the receding horizon procedure. The first iteration computes the optimal control on the interval I 1 . Only the first half of it, u 1 * , is accepted as optimal solution. Then the optimal control on the interval [I 1 /2, I 2 /2] is computed. The first half, u 2 * , is accepted and clued to u 1 * . These two steps give us the optimal control on I 1 . We proceed iteratively up to the terminal time T .
Remark 5.1. Using this receding horizon procedure we need to adapt the desired trajectoryZ. Indeed, we cannot expect that the controls lead the crowd to the destination in one subinterval. Hence, we adaptZ on I k in the following way: we interpolate the distance of the initial centre of mass of the crowd and the desired destination Z des with the time steps used in one subinterval. Of course, this is not attainable for small k, nevertheless we simulate the deterministic optimal control using this interpolation asZ on t ∈ [I k−1 , I k ]. Then, we compute the trajectory of the center of mass corresponding to this solution. We expect this trajectory to be appropriate and use it in the space mapping procedure on the interval [I k−1 , I k ].

Numerical Results
In the following we present numerical results underlining the feasibility of our approach. In particular, we investigate the number of space mapping iterations needed to obtain appropriate results. Further, we shall see how the number of dogs is influencing the success of the herding procedure. Finally, we analyze numerically if the system is stabilized for large times T 1.
For the simulations we choose Morse potentials [18] to model the interaction: To realize the self-organization of the sheep we assume that they have some long range attraction and short range repulsion, i.e., we set C r,1 = 1, C a,1 = 5e −4 , r,1 = 2, a,1 = 1e −2 .
Further, we assume the dogs to scare the sheep and therefore have stronger repulsive influence. This leads to C a,2 = C a,1 , a,2 = a,1 , C r,2 = 1e −2 , r,2 = 0.5.
Remark 6.1. We emphasize that the space mapping control approach discussed here can be adapted to various other applications by changing the interaction potentials or the cost functional.
The following parameters are fixed for all simulations where dt denotes the time step size. Moreover, we choose σ(x, t) = σ, i.e. the stochastic force is independent of space and time. Nevertheless, σ will be changed for different simulations and is thus specified explicitly later on as well as other parameters.
for two consecutive iterates u n f and u n+1 f . The accuracy of the deterministic controls deteriorates as the stochastic influence increases, see Figure 2 (up) as well as Table 1. For σ = 0.03 the stochastic influence starts to superimpose the crowd behaviour. Figure 2 (down) shows the trajectories of the center of mass of the crowd using space mapping. We see that space mapping works well for small values of σ. As the stochasticity starts to superimpose the crowd behaviour, the space mapping technique is not so efficient. This is expected, since for large volatility the deterministic model is not a good approximation of the stochastic one. The second part of Table 1 shows results obtained with 1000 Monte Carlo samples. The values change only slightly, which justifies to fix the number of MC samples to 100 for the following computations. Remark 6.2. We would like to emphasize that a basic Monte Carlo approach works fine in the present setting. For problems that are more involved, it may be necessary to use Multi-level Monte Carlo techniques in order to get efficient approximations for the stochastic states.

MC = 100
σ = 0.01 σ = 0.02 σ = 0.03 σ = 0.04 space mapping iterations 0 1 3 -L 2 error with deterministic control 7.00 · e −3 3.73 · e −2 9.09 · e −2 1.57 · e −1 L 2 error after space mapping 7.00 · e −3 1.05 · e −2 7.05 · e −2 -MC = 1000 σ = 0.01 σ = 0.02 σ = 0.03 σ = 0.04 space mapping iterations 0 1 3 -L 2 error with deterministic control 9.28 · e −3 4.25 · e −2 9.70 · e −2 1.60 · e −1 L 2 error after space mapping 9.28 · e −3 1.31 · e −2 7, 07 · e −2 - Table 1. Numerical investiation of the space mapping procedure. For σ = 0.01 no space mapping is needed, the optimal deterministic control is accepted. The number of space mapping steps increases with increasing stochastic strength. The L 2 -error of the trajectory of the center of mass compared to the center of mass of the optimal deterministic solution increases as well for larger σ. The space mapping procedure is decreasing the error by a factor three for σ = 0.02. As the stochastic starts to superimpose the crowd behaviour for σ ≥ 0.03, we see that the space mapping approach decreases the error only marginally. The second part of the Varying the number of dogs leads to very different controls which can be visualized implicitly by the trajectories of the dogs. For this study we chose the parameter values SM = 0.5, N = 20, σ = 0.01. Moreover, instead of fixing T we used |X − Z des | < 0.05 as stopping criterion and did 100 Monte Carlo runs. The change of the stopping criterion is necessary because we expect that a different number of dogs will need different times to steer the crowd to the desired destination. Figure 3 compares the trajectories of the dogs (red) and the resulting trajectory of the centre of mass of the crowd (blue). Note, that one dog has a hard time of leading the crowd as the iteration stops at T 1 = 3400. The situation is getting better for two dogs. They are successful at time T 2 = 80. Three dogs finish at time T 3 = 60. In the other cases we have T 4 = 60, T 5 = 70, T 6 = 40. Remark 6.3. We emphasize that the initial positions of the dogs were chosen manually and not included in the optimization. Hence, we cannot deduce the optimal number of dogs from these results. 6.3. Stabilization. Next, we show snapshots of a simulation with T = 250, γ = 1e −3 and 5 dogs in order to investigate if the herding process stabilizes. Indeed, we see in Figure 4 that the dogs begin to circle around the crowd when the task of steering the centre of mass to the destination Z des is achieved. This behaviour can be interpreted as stabilization of the system. For this simulation the maximum We see that already one dog is able to steer the crowd. Nevertheless, more dogs significantly decrease the time needed for the steering process. The stochastic influence in the system is implicitly displayed in the trajectory of the dogs in the figure on the top left. As for deterministic systems one would expect to have a homogeneous helix. number of space mapping iterations was limited to two. The stabilization indicated by this example is underlined by the following computations. For simplicity we set the friction parameter to α = 0 and consider the deterministic model.
Herding one sheep with two dogs. We first investigate the case having two dogs and only one sheep. Suppose the sheep is located at the destination Z des and the dogs are initially positioned at a circle with radius αp 1 around Z des , one at each end of a diameter. Then, the positions of the dogs can be parametrized with the help of p 1 as In this setting the deterministic state equations are given bẏ Choosing the initial conditions x 1 = Z des , v 1 (0) = 0, a 1 = Z des + αp 1 , a 2 = Z des − αp 1 , we obtain due to the radial symmetry of the interaction potentialṡ Thus, the system is stable for u i = 0. Note, that we have reduced the system from two controls to only one control affecting both dogs at the same time. One can even change the positions of the dogs byṗ 1 = f for some function f without any effect on the position of the sheep. This shows the stability of the configuration in this toy example.
Herding 2N sheep with 2M dogs. The observation of the previous section can be generalized to the following framework having an even number of dogs 2M and an even number of sheep 2N involved.
We assume the initial configuration to fulfil the following assumptions: (A1) v i = 0, i = 1, . . . , 2N. (A2) The centre of mass is located at the destination Z des , i.e.
(A3) For each x i exists x j with i = j and x i − a k = −(x j − a ) where a k and a are positioned at a circle around Z des each on one end of a diameter.
Further, we note that we have assumed the interaction potentials to be radially symmetric. Using these assumptions we find Thus, setting u k = 0 for all k = 1, . . . , 2M we obtain stability for the location of the centre of mass. Moreover, changing the velocities of the dogs as discussed in the framework with one sheep and two dogs, the position of the centre of mass is conserved as well. These findings are illustrated by the black circle added to the terminal configuration of the simulation shown in Figure 4. The dog at the bottom of the picture is far away from the crowd such that its contribution to the forces can be neglected.
Of course, the stochastic influence and the error in the numerical integration will lead to dogs circling around the crowd. Thus, we do not expect to obtain numerically a stable setting with all sheep and dogs standing still. . We see that the five dogs are able to steer the centre of mass of the crowd to the destination and that the crowd stays together. The latter is a new information which is not accessible by investigating only the centre of mass. Moreover, we see a stabilization as the dogs begin to circle around the crowd (t = 75, 125, 250) after the centre of mass reached the desired destination. The black circle is underlining the discussion of the stabilization of the crowd. 7. Space-mapping using the mean-field limit Crowds consisting of many individuals, i.e. N 1, are often investigated from a mesoscopic point of view with the help of a mean-field equation, see e.g. [2,6,7,9]. Following the steps in [8], this equation can be obtained via the empirical density Formally passing to the mean-field limit N → ∞ leads the deterministic optimization problem (9) J(f, u;Z,ū) : subject to the state system given by Now, drawing the random initial conditions i.i.d. from f 0 (x, v), it is well-known that f (t, x, v) assigns the probability of finding a particle at time t at position x with velocity v. Hence, one could use the deterministic mean-field problem for f as coarse model for a space-mapping in order to control the stochastic limit for many particles. Note that a similar deterministic optimization problem was solved in [8].
Remark 7.1. We want to emphasize that in the common noise case the limiting equation is not deterministic but a stochastic PDE [30]. Thus, it is not clear whether it is an appropriate choice as coarse model. The cost for computing optimal controls with the SPDE are probably very high.

Conclusion and Outlook
We discussed a space mapping approach in combination with receding horizon control for the closed loop control of a stochastic interacting particle system. The numerical results underline that the method is feasible for interacting particle systems with small stochastic perturbation. Further, they indicate that a suboptimal control for the stochastic system is found efficiently already after few space mapping iterations.
In near future, we plan to use the space mapping approach to control a stochastic system involving a large number of interacting particles. In this case, the mean-field approximation can be used as coarse model for the space mapping approach. Moreover, a rigorous analysis of the space mapping procedure applied to stochastic problems is of interest.
Further, an investigation of the stabilizing effect of the feedback control is planned. A rigorous generalization to the setting with common noise and its influence on the space-mapping performance are interesting future projects as well.