Escolar Documentos
Profissional Documentos
Cultura Documentos
1, JANUARY 2016
Abstract— Many controlled systems suffer from unmodeled If these errors can be anticipated, the control performance
nonlinear effects that recur periodically over time. Model-free can be significantly improved. While errors arising stochas-
controllers generally cannot compensate these effects, and good tically cannot be preempted systematically, periodic effects
physical models for such periodic dynamics are challenging to
construct. We investigate nonparametric system identification are amenable for prediction: since their future resembles their
for periodically recurring nonlinear effects. Within a Gaussian past, extrapolation (prediction) is easier and more structured.
process (GP) regression framework, we use a locally periodic Based on this idea, we present a framework for identification
covariance function to shape the hypothesis space, which allows and control of periodic effects. Our framework continually
for a structured extrapolation that is not possible with more performs identification at runtime, and is thus applicable to
widely used covariance functions. We show that hyperparameter
estimation can be performed online using the maximum stochastic time-varying systems.
a posteriori point estimate, which provides an accuracy The correction of periodic errors has repeatedly been
comparable with sampling methods as soon as enough data studied. The Very Large Telescope uses an internal parametric
to cover the periodic structure has been collected. It is also model for the known error sources, and a Kalman filter [1]
shown how the periodic structure can be exploited in the as an estimator for the model parameters [2]. High precision
hyperparameter optimization. The predictions obtained from
the GP model are then used in a model predictive control tracking of spacecrafts on periodic trajectories was addressed
framework to correct the external effect. The availability of in [3], based on predictive filtering using an extended
good continuous predictions allows control at a higher rate than Kalman filter [1]. To predict the beating motion of a human
that of the measurements. We show that the proposed approach heart, extended Kalman filtering for state estimation was used
is particularly beneficial for sampling times that are smaller in [4], allowing the nonlinear model to change over time.
than, but of the same order of magnitude as, the period length
of the external effect. In experiments on a physical system, an Concerning the use of learning-based models for control, there
electrically actuated telescope mount, this approach achieves a is a wide range of the literature available in the context
reduction of about 20% in root mean square tracking error. of adaptive control. For methods based on model predictive
Index Terms— Adaptive control, nonlinear control systems, control (MPC) [5]–[7].
optimal control. In contrast to previous methods for periodic error correction,
I. I NTRODUCTION the approach presented here does not rely on a prespecified
finite-dimensional model class. Instead, we propose a
S CREWS and gears are not the only source of periodically
recurring errors in dynamical systems. Every system that
is tied to the ubiquitous day-and-night cycle (like building con-
nonparametric framework based on Gaussian process (GP)
regression that is frequently used for system identification
trol or energy systems) or to recurring movements (like a beat- (see [8] for a survey). It is closely related to least-squares
ing heart or a satellite) suffers from periodic errors. Since these regression, which is the most commonly employed technique
effects are often small relative to the required control precision, in system identification, but is based on a probabilistic
they are in practice usually neglected in the controller design. interpretation, which can be used to guide exploration during
For high precision control systems, however, such errors can identification [9]. There is recent work on using GPs for
be the dominant source of problems. Correcting errors only state filtering [10] and on modeling and control of nonlinear
after they are measured leads to a delay in the error correction. systems [8], [11]. The idea of using the learned model in
predictive control is conceptually similar to [5], [6], and [12],
Manuscript received September 22, 2014; revised January 26, 2015; with the key difference that we use a GP to predict
accepted March 13, 2015. Date of publication May 11, 2015; date of current time-varying effects. Reference [13] provides a general
version December 21, 2015. Manuscript received in final form March 26,
2015. The work of M. N. Zeilinger was supported by the European Union’s introduction to GPs.
through the Seventh Framework Programme under Grant PIOF-GA-2011- A GP model is parametrized by two objects: 1) mean and
301436-COGENT. Recommended by Associate Editor A. Chiuso. 2) covariance function. When used in system identification, in
E. D. Klenske, B. Schölkopf, and P. Hennig are with the Max Planck
Institute for Intelligent Systems, Tübingen 72076, Germany (e-mail: particular, the choice of covariance function has a strong effect
edgar.klenske@tuebingen.mpg.de; bernhard.schoelkopf@tuebingen.mpg.de; on performance, and requires consideration of the particular
philipp.hennig@tuebingen.mpg.de). dynamics to be identified. Although the literature knows
M. N. Zeilinger is with the Max Planck Institute for Intelligent Systems,
Tübingen 72076, Germany, and also with the Department of Electrical universal covariance functions that can technically approxi-
Engineering and Computer Sciences, University of California at Berkeley, mate any continuous function [14], [15], this notion applies
Berkeley, CA 94720 USA (e-mail: melanie.zeilinger@eecs.berkeley.edu). only in the infinite limit, and can be subject to extremely slow
Color versions of one or more of the figures in this paper are available
online at http://ieeexplore.ieee.org. (logarithmic) convergence [16], [17]. Hence, the choice of the
Digital Object Identifier 10.1109/TCST.2015.2420629 covariance function is often critical in practice.
1063-6536 © 2015 IEEE. Translations and content mining are permitted for academic research only. Personal use is also permitted,
but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
KLENSKE et al.: GP-BASED PREDICTIVE CONTROL FOR PERIODIC ERROR CORRECTION 111
In this paper, the focus lies on the identification of often not perfectly periodic in this sense, they show various
quasi-periodic systems (Section II-A) with GPs (Section II-B) forms of phase shift, deformation, and desynchronization.
using a specific model class involving periodicity To address this issue, we generalize to consider locally
(Section III-A). A key element of GP regression is the periodic functions. These are functions for which
estimation of the hyperparameters (Section III-B), which is g(t) ≈ g(t +nλ) for nλ and g(t) ≈ g(t +nλ) for nλ ,
performed by exploiting the structure of the problem at hand where is some measure of temporal locality. Intuitively
(Section III-C). We focus on the case where the system up to speaking, local periodicity means that the periodicity is
the periodic effect is linear and use MPC to achieve optimal not strict, i.e., variations are allowed. In particular, this
closed-loop performance (Section IV-A). A reformulation covers functions that vary on a slower time scale, e.g.,
in the form of a tracking problem is proposed, which a decaying oscillation or an oscillation with long-term change
offers simple implementation and facilitates analysis of the in shape.
control performance (Section IV-B). To show the qualitative We consider the case where a linear model
properties of this framework, we apply it to a toy problem first (matrices A and B) is available and the goal is to infer
(Section V-B). As the development of this method was driven the disturbance function g(t) online from measurements.
by a real problem in astronomy, the method is evaluated on This is motivated by the fact that often a nominal model
this problem both in simulation (Sections V-C1 and V-C2) and is derived either from physical considerations or an offline
hardware experiments (Section V-C3). While GPs and MPC system identification step.
are well-studied techniques, the main contribution of this paper At every measurement time tk , the system goes through the
is a combination that is tailored to quasi-periodic functions, following process.
allowing for extrapolation and efficient computation, which 1) Measure x(tk ), ẋ(tk ).
is crucial for online identification and control. A practical 2) Construct observation for g(tk ), update model for g(t).
technique for hyperparameter optimization is proposed 3) Compute control input u(tk ) = u(x(tk ), g(tk )) and apply
and the effectiveness of the approach for compensation of to system (1).
quasi-periodic errors is demonstrated on the experiment of a It should be noted that in practice, measurements of ẋ are
telescope mount. generally not available and will be approximated numerically
This paper extends the preliminary conference version on (see also the experimental details in Section V).
the same topic [18]. We here provide a more extensive At this point, it should be intuitively clear that the
treatment of hyperparameter estimation, discuss Markov chain performance gain one can expect from the use of a periodic
Monte Carlo (MCMC) versus maximum a posteriori (MAP) model for feed-forward compensation depends on the
estimation, and propose a custom optimization method for the sampling rate of the control system: if the external error is
latter. Additional experiments include an analysis of system slow compared with the measurement rate, a locally linear
behavior under sensor failure. model is sufficient. But if the external error is on the same
time scale as the measurement, it helps to use feed-forward
II. P RELIMINARIES control based on GP predictions. With the presented approach
A. Problem Statement it is even possible to choose the control interval smaller than
the actual measurement interval (see Section V-C1 for a more
Consider the dynamical system
detailed discussion).
ẋ(t) = Ax(t) + Bu(t) + g(t) (1)
B. Gaussian Process Regression
composed of a linear system with dynamic matrices A, B,
and a nonlinear time-varying function g : R+ R E , where For notational simplicity, this section only covers the scalar
x ∈ R E denotes the state and u ∈ R F the input. For case of GPs. In the case of a vector-valued function g, one
simplicity, full state measurement is assumed. The structure GP is trained for every dimension. For our purposes, a GP
of (1) could be chosen more generally—GP models can GP(g; μ, k) is an infinite-dimensional probability distribution
also learn nonlinear functions of the state and input. We opt over the space of real-valued functions g : R R, such
for this linear formulation with nonlinear external reference that every finite, D-dimensional linear restriction to function
here to keep the resulting control problem conceptually clear values g(T ) ∈ R D (measurements) at times T ∈ R D
and computationally simple. If needed, the definition can (measurement locations) is a D-variate Gaussian distribution
also be adapted to a nonlinear system using a nonlinear N (g(T ); μ(T ), k(T, T )). It is parametrized by a mean func-
MPC technique [19]. tion μ(T ) : R D R D , and a covariance function k(T1 , T2 ) :
The function g captures nonlinear time-dependent effects, R D1 × R D2 R D1 ×D2 . The mean has a relatively straight-
in particular, we will focus on systems exhibiting some form forward role; it simply shifts predictions. The covariance
of periodic behavior. Systems with such time-dependent errors function’s responsibility is more intricate. It can be interpreted
of periodic characteristic appear in different application areas, as a similarity measure over function values in R D and
such as building temperature control, beating-heart surgery, controls the shape of the GP belief in the space of functions.
or electrical power grids. For strictly periodic functions, It has to be chosen such that, for any T1 , T2 ∈ R D , the
there exists a constant period λ, such that g(t + nλ) = g(t) matrix k(T1 , T2 ) ∈ R D×D , also known as the kernel, is positive
for n ∈ N. However, error sources in real systems are semidefinite.
112 IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, VOL. 24, NO. 1, JANUARY 2016
Fig. 3. Comparison of MCMC inference (left column) with MAP point estimation (right column) on hyperparameters of the GP model. Top: initial phase
of learning, after only a few observations. Bottom: convergence after observation of several periods.
Algorithm 1 Customized Optimization A common cost function in stochastic MPC for regulating
the system state to the origin is the expected value of the sum
of stage costs
N−1
V (x( p(g)), u) := E l(x n ( p(g)), u n ) (14)
n=0
where U ⊆ R F is a polytopic set defining the input constraints, experiment on a real telescope system, showing substantial
x := [x 0 , . . . , x N ] is the state trajectory and similarly for u. improvements in control performance.
The resulting problem is a quadratic program that can be
solved efficiently using available optimization software [30] A. Implementation Details
or fast MPC techniques proposed in recent years, such
as code generation [31]. Because this is an instance of a Since in practice the state and derivative cannot be measured
basic MPC technique, the standard properties of MPC apply. directly, they have to be approximated from potentially noisy
It also allows for a more principled analysis of the closed-loop measurements. An observer (like the Kalman filter [1]) can be
properties: extensions in the field of tracking MPC can be used to estimate the state, which increases the performance,
applied to ensure stability, such as reference governors, or especially when the measurement noise is high.
the periodic MPC approach in [32]. Applying the modified Since the function to be inferred acts as additional input to
tracking formulation in [32], convergence can be guaranteed the dynamics, observations of the disturbance are constructed
if the model g̃(t) converges to a periodic function. from observations of the change in state, x = x k+1 − x k .
Remark 1: The discrete-time model (13) with the predicted We therefore assume g(t) to be piece-wise constant,
nonlinear term ãk can in principle be used in any linear here a zero-order-hold linearization is used (chosen to be
or linearized state estimation or control method based on at t + (1/2)t)
t
state prediction. One example is the Kalman filter. The 1
nonlinear prediction from the GP can be incorporated into ak = G d g t + t , Gd = e Aτ dτ (19)
2 0
the state prediction without complicating the Kalman filter
equations. The measurement update of the Kalman filter where t is the sampling time. Note that this is only used
remains unchanged. The GP predictions increase the to generate the data point, the resulting GP model is still a
performance by providing a better state estimate. This leads continuous function. The value g(t + (1/2)t) is obtained as
to smaller correction terms and smaller posterior variance. solution of the linear system
Remark 2: Because the posterior of the GP is a Gaussian 1
distribution, state constraints can be included in the form of G d g t + t ≈ (x k+1 − Ad x k − Bd u k ) (20)
2
soft constraints, penalizing the amount of constraint violation,
or chance constraints, ensuring constraint satisfaction with a which is the data point used to train the GP.
certain probability [33], [34]. In the experiments presented in the following, the discretiza-
Remark 3: The approach can also be applied to stage tion of the mean prediction g̃(t) is obtained from the exact
cost functions and constraints that do not allow for a discretization:
t
deterministic representation using the GP model, e.g.,
ãk = e Aτ g̃((k + 1)t − τ )dτ (21)
a value-at-risk formulation involving the variance of the cost 0
using sample-based methods to approximate the stochastic
evaluated with a standard ordinary differential equation (ODE)
MPC problem [35], [36]. GPs fit well in this framework by
solver [37].
being generative models from which sample trajectories can
be easily drawn.
Remark 4: Depending on the discretization used, a tradeoff B. Toy Problem
between the mean and variance of a quadratic cost function can As a simple problem for providing intuition, consider the
be formulated as a deterministic optimization problem using following linear dynamic (double integrator) system with an
the posterior GP prediction additive time-periodic component g:
1 0 1 0
E x n Qx n = μn Qμn + tr(Qn ) (18a) ẋ = x(t) + u(t) + g(t) (22a)
2 0 0 1
1 1 sin(t)
V x n Qx n = μn Qn Qμn + tr(Qn Qn ). (18b) g(t) = . (22b)
2 2 cos(1.3t)
This is the case whenever the distribution N (μn , n ) of ak The goal is to control the first state of the system to the origin.
can be directly obtained from the distribution of g(t), e.g., in We use the quadratic cost
the case of Euler discretization. 1 1
l(x n , u n ) =
x n Qx n + u n Ru n (23)
2 2
V. E XPERIMENTS with diagonal state weight Q = diag(100, 0). The weight on
The presented method was implemented and used on the control input is set to R = 1, allowing for aggressive
different test problems. After providing a toy example that control behavior. The horizon length of the MPC is set to
demonstrates how the state evolution is anticipated by the N = 15. State and input constraints are omitted for simplicity.
use of the GP prediction, the method is evaluated for a The system was simulated numerically with a sampling
simulated telescope system where the performance under rate of 1 Hz. Fig. 5 shows control inputs and resulting state
different measurement frequencies and under sensor failure trajectories for a MPC controller without information about
is analyzed. Finally, the proposed method is tested in an the disturbance g, and a MPC controller using the posterior
KLENSKE et al.: GP-BASED PREDICTIVE CONTROL FOR PERIODIC ERROR CORRECTION 117
Fig. 5. Closed-loop input and output trajectories under the MPC controller.
MPC using a linear model is shown as thin lines and the GP-based
MPC controller as thick lines.
Fig. 8. Telescope mount used for the tracking experiments. On the right side
is the camera lens used as guiding telescope. A main telescope is not used
for the tests.
Fig. 10. Measurements of the state for the physical experiments with the
telescope (a) without controller, where the shaded area shows the vertical
range of the two other plots; (b) for plain MPC using only a linear model;
and (c) for the periodic GP-based MPC. Results (b) and (c) are from run 2
Fig. 9. Gearbox of the telescope mount. One of the motors is visible on of our experiments, which resulted in the highest (worst) rms error for both
the left. The two cogs on the right transmit the motor’s rotation to a worm models (Table I).
gear (not visible), which sits on the same axis and thus has the same period.
These cogs and worm gear are the likely source of the periodic error. TABLE I
E XPERIMENTAL R ESULTS (rms E RROR , IN asec)
VI. C ONCLUSION
High precision control of dynamical systems requires
precise models of even minor external error sources. Where
analytic models are not available, they can only be constructed
numerically from measurements of the system. Periodic error
sources are an especially promising domain in this regard,
as they can be extrapolated well into the future. We have
studied a nonparametric modeling framework based on a
carefully crafted GP prior exhibiting a weak, localized form of
periodicity. Because Gaussian regression returns models in the
form of stochastic differential equations, they can be combined
directly with existing control frameworks. Integration into
an MPC scheme was investigated, which can leverage the
prediction at a desired resolution, even below the sampling
time. Numerical and physical experiments confirm the
intuitive result that the benefit of periodic models depends on
the relative size of state sampling and disturbance frequencies.
Fig. 11. Power spectra of the measurements of Fig. 10. (a) Plain MPC using
only a linear model. (b) Periodic GP model-based MPC.
We showed that, even in cases where the gain of a periodic
prediction is only marginal during normal operation, such
models are beneficial when sensors fail temporarily. The
presented method also shows considerable increases in
control performance, confirming the practical utility of this
has to be built and inverted at every evaluation of the objective framework.
function, the number of evaluations has to be kept small at
every sampling time. We use a numerical optimizer based R EFERENCES
on the BFGS update [38]–[41], a quasi-Newton algorithm
[1] S. Särkkä, Bayesian Filtering and Smoothing. Cambridge, U.K.:
that updates an estimate of the inverse Hessian in each Cambridge Univ. Press, 2013.
iteration. In the standard implementation, the estimate of the [2] T. Erm and S. Sandrock, “Adaptive periodic error correction for the
Hessian obtained at one time step is discarded after each VLT,” Proc. SPIE, vol. 4837, p. 900, Feb. 2003.
[3] J. L. Crassidis and F. L. Markley, “Predictive filtering for nonlinear
individual call to the optimizer. For the use in the control systems,” J. Guid., Control, Dyn., vol. 20, no. 3, pp. 566–572,
setting, we altered the algorithm so that the estimate of the 1997.
inverse Hessian is stored and used to initialize the optimizer’s [4] S. G. Yuen, P. M. Novotny, and R. D. Howe, “Quasiperiodic predictive
filtering for robot-assisted beating heart surgery,” in Proc. IEEE Int.
estimate at the next sampling time. This makes it possible to Conf. Robot. Autom., May 2008, pp. 3875–3880.
do one iteration per sampling interval (in a sense, threading [5] J. Kocijan, R. Murray-Smith, C. E. Rasmussen, and A. Girard, “Gaussian
the optimization algorithm into the learning algorithm). This process model based predictive control,” in Proc. Amer. Control Conf.,
2004, pp. 2214–2219.
significantly reduces computation time. [6] A. Aswani, H. González, S. S. Sastry, and C. Tomlin, “Provably safe
The presented method was tested on two different setups, and robust learning-based model predictive control,” Automatica, vol. 49,
one with an MPC based on the linear model without g and no. 5, pp. 1216–1226, 2013.
[7] M. Tanaskovic, L. Fagiano, R. Smith, P. Goulart, and M. Morari,
one with the GP prediction for the periodic error. The test “Adaptive model predictive control for constrained linear systems,” in
was run three times for 25 min each. Both the sampling and Proc. Eur. Control Conf., 2013, pp. 382–387.
the discretization time were set to 3 s. The horizon length [8] G. Pillonetto, F. Dinuzzo, T. Chen, G. De Nicolao, and L. Ljung,
“Kernel methods in system identification, machine learning and
was N = 10; the state and control weights were set to function estimation: A survey,” Automatica, vol. 50, no. 3, pp. 657–682,
Q = 102 and R = 103 . The results of these runs are 2014.
shown in Table I. The rms error drops by 22.64% through [9] P. Hennig, “Optimal reinforcement learning for Gaussian systems,” in
Advances in Neural Information Processing Systems. Red Hook, NY,
the use of GP predictions in this hardware setup. Baseline USA: Curran & Associates Inc., 2011, pp. 325–333.
measurements without movement showed about 0.25 asec– [10] J. Ko and D. Fox, “GP-BayesFilters: Bayesian filtering using Gaussian
0.35 asec of noise. The noise introduced by the stepper motor process prediction and observation models,” Auto. Robots, vol. 27, no. 1,
pp. 75–90, 2009.
could not be quantified with the current measurement system. [11] J. Hall, C. Rasmussen, and J. Maciejowski, “Modelling and control of
Overall, the presented method eliminated at least a third of the nonlinear systems using Gaussian processes with partial model infor-
controllable error, after subtracting baseline noise but without mation,” in Proc. IEEE 51st Annu. Conf. Decision Control, Dec. 2012,
pp. 5266–5271.
taking the stepper motor into account. That is a good result [12] J. M. Maciejowski and X. Yang, “Fault tolerant control using Gaussian
in this domain, but could probably be improved, which is also processes and model predictive control,” in Proc. Conf. Control
visible through the light residual structure visible in Fig. 10(c). Fault-Tolerant Syst. (SysTol), 2013, pp. 1–12.
[13] C. E. Rasmussen and C. K. I. Williams, Gaussian Processes for Machine
To further investigate this issue and assess the performance Learning. Cambridge, MA, USA: MIT Press, 2006.
of the method, Fig. 11 shows the power spectra of the measure- [14] I. Steinwart, “On the influence of the kernel on the consistency of
ments, obtained from the Fourier transform. It is noticeable support vector machines,” J. Mach. Learn. Res., vol. 2, pp. 67–93,
Mar. 2002.
that the strong periodic components near 100 and 500 s are [15] C. A. Micchelli, Y. Xu, and H. Zhang, “Universal kernels,” J. Mach.
highly damped with the presented method. Learn. Res., vol. 7, pp. 2651–2667, Dec. 2006.
KLENSKE et al.: GP-BASED PREDICTIVE CONTROL FOR PERIODIC ERROR CORRECTION 121
[16] A. W. van der Vaart and J. H. van Zanten, “Rates of contraction of Edgar D. Klenske received the Diploma degree
posterior distributions based on Gaussian process priors,” Ann. Statist., in engineering cybernetics from the University of
vol. 36, no. 3, pp. 1435–1463, 2008. Stuttgart, Stuttgart, Germany, in 2012. He is cur-
[17] A. W. van der Vaart and J. H. van Zanten, “Information rates of rently pursuing the Ph.D. degree with the Depart-
nonparametric Gaussian process methods,” J. Mach. Learn. Res., vol. 12, ment of Empirical Inference, Max Planck Institute
pp. 2095–2119, Feb. 2011. for Intelligent Systems, Tübingen, Germany.
[18] E. D. Klenske, M. N. Zeilinger, B. Schölkopf, and P. Hennig, “Non- He conducted his Diploma thesis research with the
parametric dynamics estimation for time periodic systems,” in Proc. Max Planck Institute for Intelligent Systems from
51st Annu. Allerton Conf. Commun., Control, Comput., Oct. 2013, 2011 to 2012. His current research interests include
pp. 486–493. probabilistic view on controllers and the connections
[19] M. Diehl, H. J. Ferreau, and N. Haverbeke, “Efficient numerical methods between control theory and machine learning.
for nonlinear MPC and moving horizon estimation,” in Nonlinear Model
Predictive Control, vol. 384. Berlin, Germany: Springer-Verlag, 2009,
pp. 391–417.
[20] D. J. C. MacKay, “Introduction to Gaussian processes,” NATO ASI Series
F Comput. Sys. Sci., vol. 168, pp. 133–166, 1998. Melanie N. Zeilinger received the Diploma degree
[21] D. Duvenaud, J. R. Lloyd, R. Grosse, J. B. Tenenbaum, and in engineering cybernetics from the University of
Z. Ghahramani, “Structure discovery in nonparametric regression Stuttgart, Stuttgart, Germany, in 2006, and the
through compositional kernel search,” in Proc. Int. Conf. Mach. Ph.D. (Hons.) degree in electrical engineering from
Learn. (ICML), 2013, pp. 1166–1174. ETH Zurich, Zurich, Switzerland, in 2011.
[22] D. Barber, Bayesian Reasoning and Machine Learning. Cambridge, She conducted her Diploma thesis research with
U.K.: Cambridge Univ. Press, 2011. the University of California at Santa Barbara, Santa
[23] M. B. Thompson and R. M. Neal. (2010). “Slice sampling with adaptive Barbara, CA, USA, from 2005 to 2006. She was
multivariate steps: The shrinking-rank method.” [Online]. Available: a Post-Doctoral Researcher with the Department
http://arxiv.org/abs/1011.4722 of Electrical Engineering and Computer Sciences,
[24] L. Wasserman, All of Statistics: A Concise Course in Statistical Infer- University of California at Berkeley, Berkeley, CA,
ence. New York, NY, USA: Springer-Verlag, 2010. USA, from 2012 to 2014, and a Post-Doctoral Fellow with the École Poly-
[25] D. J. C. MacKay, “Comparison of approximate methods for handling technique Fédérale de Lausanne, Lausanne, Switzerland, from 2011 to 2012.
hyperparameters,” Neural Comput., vol. 11, no. 5, pp. 1035–1068, She is currently a Research Scientist and Marie Curie Fellow with the
1999. Department of Empirical Inference, Max Planck Institute for Intelligent
[26] R. R. Bitmead, M. Gevers, and V. Wertz, Adaptive Optimal Control: Systems, Tübingen, Germany. Her current research interests include
The Thinking Man’s GPC. New York, NY, USA: Prentice-Hall, 1990. distributed control and optimization, and safe learning-based control, with
[27] J. B. Rawlings and D. Q. Mayne, Model Predictive Control: Theory and applications to energy distribution and management systems.
Design. Madison, WI, USA: Nob Hill Publishing, 2009.
[28] B. Kouvaritakis and M. Cannon, “Stochastic model predictive control,”
in Encyclopedia of Systems and Control, J. Baillieul and T. Samad, Eds.
London, U.K.: Springer-Verlag, 2014, pp. 1–9. Bernhard Schölkopf (SM’08) received the
[29] D. Q. Mayne, “Model predictive control: Recent developments and Ph.D. degree in computer science from Technical
future promise,” Automatica, vol. 50, no. 12, pp. 2967–2986, 2014. University Berlin, Berlin, Germany. His thesis was
[30] M. Grant and S. Boyd. (2014). CVX: MATLAB Software for involved in Support Vector Learning. He is director
Disciplined Convex Programming, Version 2.1. [Online]. Available: of the Department of Empirical Inference, Max
http://cvxr.com/cvx Planck Institute for Intelligent Systems, Tübingen,
[31] A. Domahidi. (2012). FORCES: Fast optimization for Real-Time Control Germany.
on Embedded Systems. [Online]. Available: http://forces.ethz.ch He has been involved in research with AT&T Bell
[32] D. Limon, T. Alamo, D. Muñoz de la Peña, M. N. Zeilinger, C. N. Jones, Labs, Murray Hill, NJ, USA, GMD FIRST, Berlin,
and M. Pereira, “MPC for tracking periodic reference signals,” in Proc. Australian National University, Canberra, ACT,
4th IFAC Nonlinear Model Predict. Control Conf., 2012, pp. 490–495. Australia, and Microsoft Research Cambridge,
[33] A. T. Schwarm and M. Nikolaou, “Chance-constrained model predictive Cambridge, U.K. The ISI lists him as a Highly Cited Researcher.
control,” AIChE J., vol. 45, no. 8, pp. 1743–1752, 1999. Prof. Schölkopf is a member of the IEEE Computer Society. He was
[34] M. Ono and B. C. Williams, “Iterative risk allocation: A new appointed as a Scientific Member of the Max Planck Society in 2001. He
approach to robust model predictive control with a joint chance received the J. K. Aggarwal Prize of the International Association for Pattern
constraint,” in Proc. 47th IEEE Conf. Decision Control, Dec. 2008, Recognition in 2006. His thesis received the Annual Dissertation Prize of
pp. 3427–3432. the German Association for Computer Science. He is a Board Member of
[35] M. C. Campi, S. Garatti, and M. Prandini, “The scenario approach the NIPS Foundation and the International Machine Learning Society.
for systems and control design,” Annu. Rev. Control, vol. 33, no. 2,
pp. 149–157, 2009.
[36] G. Schildbach, L. Fagiano, C. Frei, and M. Morari, “The scenario
approach for stochastic model predictive control with bounds on Philipp Hennig received the Diploma degree
closed-loop constraint violations,” Automatica, vol. 50, no. 12, in physics from the University of Heidelberg,
pp. 3009–3018, 2014. Heidelberg, Germany, in 2007, and the Ph.D. degree
[37] J. R. Dormand and P. J. Prince, “A family of embedded Runge–Kutta from the University of Cambridge, Cambridge, U.K.,
formulae,” J. Comput. Appl. Math., vol. 6, no. 1, pp. 19–26, 1980. in 2011.
[38] C. G. Broyden, “A new double-rank minimization algorithm,” Notices He is currently an Emmy Noether Group Leader
Amer. Math. Soc., vol. 16, p. 670, 1969. with the Department of Empirical Inference, Max
[39] R. Fletcher, “A new approach to variable metric algorithms,” Comput. J., Planck Institute for Intelligent Systems, Tübingen,
vol. 13, no. 3, pp. 317–322, 1970. Germany, where he has resided since 2011. He is
[40] D. Goldfarb, “A family of variable-metric methods derived by variational particularly interested in descriptions of numerical
means,” Math. Comput., vol. 24, no. 109, pp. 23–26, 1970. algorithms as probabilistic inference, in the use of
[41] D. F. Shanno, “Conditioning of quasi-Newton methods for func- probability distributions to encode, share, and transfer information between
tion minimization,” Math. Comput., vol. 24, no. 111, pp. 647–656, related computations, and to control the use of computational budget.
1970. His current research interests include the use of uncertainty in computation.