Skip to content
BY 4.0 license Open Access Published by De Gruyter (O) September 9, 2021

Constrained stochastic predictive control of linear systems with uncertain communication

Stochastische prädiktive Regelung beschränkter linearer Systeme mit unsicherer Kommunikation
  • Jannik Hahn

    Jannik Hahn is research assistant with the Control and System Theory Group, Department of Electrical Engineering and Computer Science, Universität Kassel. His research activities are focussed on robust and stochastic distributed model predictive control.

    EMAIL logo
    and Olaf Stursberg

    Olaf Stursberg is Full Professor and Head of the Control and System Theory Group in the Department of Electrical Engineering and Computer Science at University of Kassel. His main research areas include methods for optimal and predictive control of networked and hierarchical systems, techniques for analysis and design of hybrid dynamic systems, and the control of stochastic, uncertain and learning systems in different domains of application.

Abstract

This paper proposes a scheme of model predictive control for single-loop networked control system (NCS) with probabilistically modeled communication channel and disturbances. Uncertainties of the communication network are projected onto a tailored probability for the satisfaction of state and input constraints. The proposed receding horizon control scheme uses a probabilistic terminal state and set to establish a balance between control performance and state probability distribution, while satisfying the given constraints. In addition to describing the control approach, its properties are discussed, and it is illustrated by an example.

Zusammenfassung

Dieser Artikel schlägt ein Schema zur modellprädiktiven Regelung einschleifiger Regelkreise mit probabilistisch modelliertem Kommunikationskanal und Störung vor. Unsicherheiten des Kommunikationsnetzes werden auf eine angepasste Wahrscheinlichkeit zur Erfüllung von Zustands- und Eingangsbeschränkungen projiziert. Die Nutzung von probabilistischen Terminalmengen sowie -zuständen erlaubt die Einstellung eines Kompromisses zwischen der Regelgüte und der Zustandsverteilung bei Einhaltung der Beschränkungen. Der Beitrag stellt den Ansatz vor, diskutiert seine Eigenschaften und beschreibt die Anwendung anhand eines Beispiels.

1 Introduction

This article combines principles of model predictive control (MPC) and networked control systems (NCS) to address timing effects that arise from wireless components within control loops [1]. As wireless sensors have become widely available and are simple to integrate [2], they also contribute to enabling concepts of Industry 4.0, e. g., the control of mobile systems with stationary motion capture systems [3]. On the downside, the drawback of electromagnetic susceptibility (like packet loss and/or latency) arises if used in control systems. As imperfect wireless communication can be interpreted as a stochastic process, the behavior, control performance, and stability of the NCS depend on stochastic information flow. In case the communication delays are in the order of the dominant time constants of the plant dynamics, the consideration of the stochastic process within the controller synthesis is mandatory, while being omitted in the vast majority of methods proposed for NCS. Some approaches model the imperfections by considering a worst case delay [4], [5], [6] – however, the use of constant worst case delays within controller synthesis typically leads to overly conservative results and degraded performance. Own work in this direction aimed at the reduction of conservatism by use of variable communication delays within the controller synthesis [7], [8]. The key point there is to predict and robustly bound the delay of the communication network, and to consider the time-varying bounds in distributed MPC.

However, as all these approaches aim at robust satisfaction of constraints (while the underlying communication process is of probabilistic nature), the reduction of conservatism is limited. Therefore, the objective of this contribution is to model the NCS completely stochastic, and to use stochastic MPC (SMPC). In SMPC the probabilistic system behavior is explicitly described by its underlying probability distribution to directly address the probability to satisfy or violate associated constraints. For a brief overview of SMPC and its application to control systems without network structures, see [9]. For more explicit literature on SMPC with state and input constraints, the interested reader is referred to [10], [11]. In the context of NCS, the satisfaction of state constraints – especially in the probabilistic setting – is still a largely open issue. Thus, [12] has further investigated the stochastic problem setup and provides a stochastic optimal control scheme, which is able to handle both, uncertain communication and additive disturbances affecting the plant as stochastic processes. The uncertainty of when future sensor information will be received by the controller are projected onto a tailored likelihood to satisfy state and input constraints. The promising result is that a poorly performing communication link will affect the uncertainty, but not necessarily the performance of controlled behavior. In this article, the findings of [12] are used to develop an MPC scheme, in order to cope with larger horizons and to further reduce the conservatism. As usual for MPC, this implies the formulation and use of a terminal control law and set for the stochastic system to ensure recursive feasibility and stability in a probabilistic sense.

The article is structured as follows: Section 2 introduces notation and clarifies different mathematical aspects used throughout the article. Section 3 specifies the networked control system and the control problem. Section 4, as main part of the paper, details aspects of the stochastic communication channel, the computation of the terminal control law, and the stochastic MPC scheme including properties. The scheme is then applied in Sec. 5 to the example of an aerial vehicle, and conclusions are contained in Sec. 6.

2 Preliminaries

The symbol s ( k ) denotes a discrete-time value of a signal s ( t ) R n s at time t = k · Δ t, with k N + , and constant time-step Δ t R + . The value of the signal s for a future time-step k + t , t N + as predicted in k is denoted by s t | k .

A convex polytope for s, parameterized by a pair ( C s , b s ) of matrix and vector of appropriate dimension, is referenced by:

S = s C s · s b s .

An ellipsoidal set is defined by its center point and shape matrix ( s ¯ , S ):

ε s ¯ , S = s ( s s ¯ ) S 1 ( s s ¯ ) 1 .

A multivariate normal distribution of an n s -dimensional random vector s with covariance S R n s × n s and a mean-value s ¯ R n s is denoted by s N s ¯ , S . Correspondingly, the expectation of a value s is denoted by s ¯ = E s . An affine transformation of a normal distribution with matrix M and vector v is again a normal distribution according to:

M · N s ¯ , S + v = N M s ¯ + v , M S M ,

and the sum of two normally distributed random vectors s 1 N s ¯ 1 , S 1 and s 2 N s ¯ 2 , S 2 is also again a normal distribution:

s 1 + s 2 N s ¯ 1 + s ¯ 2 , S 1 + S 2 .

The level-curves of a Gaussian probability density function are ellipsoidal, such that the δ-confidence ellipsoid of a normal distribution with rank n and with probability δ 0 , 1 is defined by:

S [ δ ] = ε s ¯ , c [ δ ] S ,

using a scaling factor c [ δ ] = F χ 2 1 δ , n , in which F χ 2 denotes the cumulative distribution function of a χ 2 -distribution. Throughout the paper, the mean value of a distribution coincides with the center point of the confidence ellipsoid, and the shape matrix is equal to the covariance matrix of the distribution. Thus, the same notation is used for these quantities.

The probability for an event A is referred to by P ( A ), and if this probability is specific for the time-step k, it is denoted by P k ( A ). Likewise, the conditioned expectation for a signal s t | k is denoted by E k s t | k = : s ¯ t | k , since prediction and expectation rely on the same time-step k throughout the paper.

Finally, the symbol s Q 2 = s · Q · s denotes a weighted 2-norm of the vector s with weight Q.

3 Considered class of networked control systems

The class of systems under consideration in this paper represent single-loop feedback structures with wireless communication network in the signal link between plant and controller, see Fig. 1a. This setting can be motivated, e. g., in the context of controlling aerial vehicles, as explained in Sec. 5. The plant is modeled by discrete-time linear systems with probabilistic additive disturbances:

(1) x ( k + 1 ) = A x ( k ) + B u ( k ) + E w ( k ) , w ( k ) N 0 , W ( k ) ,

with state vector x R n x , input vector u R n u , and a vector of disturbances w R n w . The additive disturbances cannot be measured and are assumed to be normally distributed with zero-mean and covariance matrix W, and to be independent, identically distributed (i. i. d.).

Figure 1 
Structure of the networked control system and the corresponding Markov chain modeling the communication network.
Figure 1

Structure of the networked control system and the corresponding Markov chain modeling the communication network.

Assumption 1.

The stochastic disturbance process may vary over time, but is assumed to be known with an upper bound W ˆ of the covariance: W ( k ) W ˆ.

The states x ( k ) and inputs u ( k ) have to satisfy polytopic chance-constraints with likelihoods δ x , and δ u respectively (both typically chosen close to 1):

(2a) P ( x ( k ) X k ) δ x , X k = x C x k · x b x k ,
(2b) P ( u ( k ) U ) δ u , U = u C u · u b u .
The admissible sets X k and U are assumed to be convex, compact, and to contain the origin in their interior for all k.

The controller and plant are statically connected in wired form, but sensor and controller are loosely connected by wireless communication. The communication network is established as a simple one-link channel and consists of a sender S, a receiver R, and the link-probability p ( k ), see Fig. 1a. The link probability can be understood as an abstract representation of the network characteristics (or protocol), as, e. g., resending information in case of failed transmission, or activation/deactivation of the channel by a network controller, but it could also be interpreted as the probability that measured sensor data is available. However, for assuming a time-varying link-probability, the communication network is modeled by a Markov chain [13] in each time-step k, as shown in Fig. 1b.

Assumption 2.

The link-probability p ( k ) may vary over time, but is assumed to be known (by measurement, or possibly according to a predictive network control scheme, see [14]). Furthermore, possible dropouts of information at times k and k + t are i. i. d. for all { k , t } N + .

In time-step k, the current state x ( k ) is available to S. The information is broadcast to R, and thus available to the controller with probability p ( k ). If information is not received, the information remains with S. Then, the last information received by the controller is outdated, such that the age of the currently available state information increments. Similar to the previous work in [8], [12], the following definition is proposed:

Definition 1.

The quantity a ( k ) N 0 denotes the age of the newest state information x ( t ) , t N k available to the controller (short: AoI), and it quantifies the difference between the time-instances of sending and using the information x ( t ).

If x ( k + 1 ) is received, the AoI is reset to zero, otherwise it is incremented again.

The main control objective for this networked control system is to steer the states to the origin with acceptable costs for the input, i. e., to find a control sequence { u ( k ) } k = 0 which minimizes the infinite horizon cost function:

J = k = 0 l x ( k ) , u ( k )

subject to the state and input constraints, and the uncertain AoI respectively. The stage costs are formulated in the following common form:

(3) l ( x , u ) = x Q x 2 + u Q u 2

with symmetric and positive semi-definite weighting matrix Q x = Q x 0 for the states, and definite weighting matrix Q u = Q u > 0 for the inputs.

4 SMPC based on stochastic optimization

To cope with the challenges of control under uncertainty of the reception of data and the satisfaction of chance constraints, the following control scheme is proposed: In each time-step k in which information is received (i. e., a ( k ) = 0), the controller solves a finite-time stochastic optimal control problem with horizon H. The solution of the optimal control problem is a control sequence { u t | k } t = 0 H 1 , which has to guarantee the satisfaction of the constraints (2) with respect to the upcoming communication link probabilities { p ( k + t ) } t = 1 H 1 . Then, as usual for MPC, the first control input is applied to the system, i. e., u ( k ) = u 0 | k . In case no information is received, i. e., a ( k ) > 0, the controller solves no optimization problem, but applies the subsequent control input of the optimal control law computed in the previous optimization. This was accomplished at time k a ( k ), thus the input is chosen according to u ( k ) = u a ( k ) | k a ( k ) . Note that the use of a previously computed control input is relatively unusual in predictive control, but here it is used to avoid accounting for uncertain initial states within the optimization problem, discussed later on.

This scheme is repeated recursively, thus the case of an infinite horizon control, as mentioned in Sec. 3, would be approximated. However, due to the properties of the probabilistic communication link, the possibility exists that the communication link fails for more than H subsequent time-steps, i. e., P a ( k ) H > 0. In this case, no control law computed before is available. This problem is immanent to the considered setting, and cannot be healed by any closed-loop control strategy (while its probability of occurrence can, of course, be reduced by increasing H within finite-horizon MPC). Consequently, closed loop control may not be possible, if the lack of measured information persists permanently, while open-loop control remains to be an option (obvious from Fig. 1a). Thus, the following assumption needs to be stated:

Assumption 3.

A fallback strategy, like stalling or shutting-down operation, exists, which can be applied in the case that the communication link fails for H or more subsequent time-steps, to ensure safety of operation.

Obviously, the age of information (AoI) is a crucial value in operating the NCS and for the success of the predictive control policy. Hence, this quantity is detailed in the following subsection.

4.1 Modelling the age of information

According to the one-link communication network in Fig. 1, the AoI depends on the behavior of the communication link, and its transmission probability. In each time-step, the transmission succeeds if x ( k ) is made available to the controller, implying according to Def. 1 that a ( k ) = 0. If the transmission fails, the newest information previously sent to the controller remains the same as before. But since this information is older by one time-step, a ( k ) : = a ( k 1 ) + 1 results. According to this scheme, the behavior of the AoI is modeled (and becomes therefore predictable) according to a discrete-time Markov chain with states σ j ( k + t ) implying a ( k + t ) = j, and with probabilities q ( k + t ) = 1 p ( k + t ), as shown in Fig. 2.

Figure 2 
Example of a Markov chain to model the AoI for 3 subsequent time-steps with 
k
=
0k=0 and 
a
(
0
)
=
0a(0)=0.
Figure 2

Example of a Markov chain to model the AoI for 3 subsequent time-steps with k = 0 and a ( 0 ) = 0.

With an initial distribution μ ( k ) (determined by the current AoI), the distribution for future values of the AoI results to:

(4) μ t | k = r = 1 t P ( r ) · μ ( k )

with time-varying transition probability matrix P ( k ). For the example of Fig. 2, P ( k ) is given by:

P ( k ) = p ( k ) p ( k ) p ( k ) 1 q ( k ) 0 0 0 0 q ( k ) 0 0 0 0 q ( k ) 0 .

The probability for a specific value of the AoI in time-step k + t predicted in k is given by:

(5) P k ( a ( k + t ) = j ) = μ t | k [ j ] ,

where μ [ j ] denotes the ( j + 1 )-th entry of the vector μ.

4.2 AoI-dependent constraints

The predicted evolution of the NCS is subject to two types of uncertainties, the additive disturbances and the AoI. While the disturbances may lead to violation of the state constraints, the AoI may prevent necessary control actions. To minimize the uncertainty of prediction (and thus to enhance the control performance), a control law is synthesized which compensates the additive disturbances. To obtain a synthesis problem with admissible solution, a scheme of disturbance feedback is used. Since the applicability of the control law depends on the AoI, the constraints are formulated depending on this quantity. In any time-step, in which the controller receives new information, i. e., a ( k ) = 0, the predictive scheme requires to solve the stochastic optimal control problem. The solution { u t | k } t = 0 H 1 has to satisfy input and state constraints (2), despite the uncertainty arising from the AoI. This raises the question of how the feedback policies should be chosen, given the situation that it is unclear whether the future states are available to the controller. Therefore, uncertainties of the communication network are projected onto a tailored probability for the satisfaction of state and input constraints.

However, the availability of the states is predictable according to the AoI with (5): x ( k + r ) is available to the controller in time-step k + t , t r, if a ( k + t ) t r holds. Now assume for a moment, that if x ( k + r ) is available, also all prior states { x ( k + t ) } t = 0 r are available to the controller. Then, also all affecting disturbances { w ( k + t ) } t = 0 r 1 (except of the last) could be reconstructed by the controller according to (1).

Hence, the probability that a disturbance w ( k + r ) is available to the controller in k + t results with (5) to:

P k ( r < t a ( k + t ) ) = j = 0 t r 1 μ t | k [ j ] = : p t , r | k ,

which follows from (4) with μ ( k ) = 1 , 0 , , 0 as a ( k ) = 0. With the probabilities p t , r | k , the likelihoods { δ u , δ x } for satisfying the input and state constraints are separated into: the likelihood that the control policy from optimization in k is applicable in k + t: α t | k max { δ u , δ x } , 1 , and the tailored likelihoods to satisfy the input and state constraints: { γ t | k ( u ) , γ t | k ( x ) }, such that:

(6) δ u = γ t | k ( u ) · α t | k , δ x = γ t | k ( x ) · r = 1 t 1 α r | k .

Note, that the choice of the α t | k allows for some degrees of freedom, but is proposed to be as in [12]:

(7) α t | k = min r N 0 { p t , r | k , 1 } , s. t.: p t , r | k > max { δ x , δ u } , r = 1 t 1 α r | k δ x t H .

Finally, the disturbance feedback policy used for the optimization results to:

(8) u t | k = V t | k · x ( k ) + r = 0 r ˆ t | k M t , r | k · w ( k + r ) ,

with feedback matrices V R n u × n x and M R n u × n w , and r ˆ t | k as the minimizer of (7). Note that (8) is equivalent to a linear state-feedback [15], and could easily extended to an affine control law [16]. However, this is not necessary here, since the NCS is controlled to the origin.

In summary, the uncertain AoI requires the computation of the probabilities { μ t | k } t = 1 H 1 to choose the values { α t | k , r ˆ t | k } t = 1 H 1 , and to tighten the state and input constraints with (6) according to:

(9a) P k ( x t | k X k + t ) γ t | k ( x ) , t = { 1 , , H } ,
(9b) P k ( u t | k U ) γ t | k ( u ) , t = { 0 , , H 1 } .
Finally, the uncertainty of the AoI is cast into a variation of the control law, and the state and input constraints, such that a common stochastic optimal control problem can be formulated. In order to adapt the optimal control problem into one of predictive control, the next section clarifies how to specify the terminal conditions for the NCS with uncertain AoI.

Note the following important point: in each time-step k, in which new information arrives, an optimization over the control law (8) is initiated. Then, only the state x ( k ) is needed to compute the input u 0 | k (which has to be applied to the NCS), since w ( k + r ) r 0 is never available in k. In the first succeeding time-step k + t for which the feedback of disturbance w ( k ) (the first disturbance affecting the system after reception of x ( k )) was expected in k, new information is received with probability α t | k . In case this information is received (and w ( k ) could be used within u t | k ), a new optimization is started and the reconstruction of w ( k ) is not necessary anymore. In case the expected information is not available, the input u t | k has to be applied without the expected feedback of disturbance w ( k ). This may violate the state and input constraints, but is scheduled according to the tightened probabilities γ t | k ( u ) , and γ t | k ( x ) . Hence, the assumption to know all prior states and disturbances (which was made above), is not necessary within the predictive control scheme, but allows the prediction of an admissible control sequence according to (8).

4.3 The terminal system and covariance

The control strategy aims at recursively solving the stochastic optimal control problem to obtain a predictive control scheme. In deterministic constrained predictive control, stability is usually guaranteed by i.) the use of an invariant set as terminal constraint for the last state of the prediction horizon, and ii.) by the use of a Lyapunov-based terminal cost term in the objective function. Similarly to the control of delayed systems, the terminal set and cost term have to be modified for the NCS to account for the uncertain AoI. In addition and according to the additive and probabilistic disturbance w ( k ), the use of an invariant covariance is introduced to ensure recursive feasibility. In order to formulate terminal covariance and terminal set, a terminal system is introduced, motivated by the findings in [4].

Terminal system

With delay-time τ N > 0 , the state vector of the terminal system ξ R n ξ with n ξ = ( τ + 1 ) · n x is introduced to consist of multiple state vectors of the NCS for different time instances:

ξ H | k = x H τ | k x H 2 | k x H 1 | k x H | k .

The dynamics of the terminal system is chosen to:

(10) ξ t + 1 | k = A ξ · ξ t | k + B ξ · ϕ t | k + E ξ · w ( k + t )

for t H and with matrices:

A ξ = 0 τ · n x × n x I τ · n x 0 n x × τ · n x A , B ξ = 0 τ · n x × n u B , E ξ = 0 τ · n x × n w E ,

and the input vector ϕ R n u . Obviously, (10) comprises and extends the dynamics (1), if ϕ t | k : = u t | k is chosen.

Now consider a delayed control law (which is later called terminal control law):

(11) ϕ t | k = K τ · x t τ | k = : K ξ · ξ t | k

with K τ R n u × n x such that the terminal system has a stable closed loop matrix A ξ , c l = ( A ξ + B ξ K ξ ). Note that this control law is applicable as long as the AoI does not exceed the value τ, what is assumed temporarily.

Given the probabilistic disturbance w ( k ) with zero mean, the closed loop behavior of the terminal state is predictable with ξ t | k N ξ ¯ t | k Ξ t | k with mean value ξ ¯ t | k and covariance Ξ t | k evolving according to:

(12a) ξ ¯ t + 1 | k = A ξ , c l ξ ¯ t | k ,
(12b) Ξ t + 1 | k = A ξ , c l Ξ t | k A ξ , c l + E ξ W ( k + t ) E ξ .
Due to the structure of the terminal system, the rank of the underlying distribution is limited to n x , regardless of the dimension of n ξ , or τ respectively. Thus, an ellipsoid with confidence δ for containing the predicted state ξ t | k is introduced and denoted by Ξ t | k [ δ ] = ε ξ ¯ t | k , c [ δ ] · Ξ t | k , with a scaling factor c [ δ ] = F χ 2 1 δ , n x .

Invariant covariance and probabilistic invariant set

By applying (12) recursively, the evolution of the terminal state is obtained. With stable closed-loop matrix A ξ , c l and for t , the expected value ξ t | k converges to zero. With respect to the upper bound W ˆ W ( k ) from Asm. 1, an invariant covariance can be obtained, similar to the work in [17].

Definition 2.

A matrix Σ ξ = Σ ξ 0 is called invariant covariance for system (10) with control law (11) and upper bound W ˆ W ( k ), if it satisfies the Lyapunov equation:

Σ ξ = A ξ , c l Σ ξ A ξ , c l + E ξ W ˆ E ξ .

Corollary 1.

If Σ ξ is an invariant covariance, the following holds:

Ξ t + 1 | k Σ ξ , Ξ t | k Σ ξ , W ( k ) W ˆ .

Obviously from Def. 2 and Corollary 1, the predicted terminal state converges to a neighborhood around the origin, i. e., ξ | k N 0 , Σ ξ .

Next, the satisfaction of the constraints (2) is considered for the terminal system: From Corollary 1 follows directly that the terminal system is held within an ellipsoid of δ x -confidence:

Corollary 2.

Given an invariant covariance Σ ξ , the probability δ x 0 , 1 , and a corresponding confidence ellipsoid Σ ξ [ δ x ] = ε 0 , c [ δ x ] · Σ ξ with scaling factor c [ δ x ] = F χ 2 1 δ x , n x , the relation:

(13) Ξ t | k [ δ x ] Σ ξ [ δ x ]

implies for all r t that:

P k ξ ( k + r ) Σ ξ [ δ x ] δ x .

In consequence, satisfaction of (2a) can be guaranteed, if the confidence ellipsoid lies within the admissible state-space, i. e., Σ ξ [ δ x ] X k + t τ × × X k + t , and thus if the ellipsoid-in-ellipsoid constraint (13) is satisfied. In general, an ellipsoid-in-ellipsoid problem can be solved according to the S-procedure proposed in [18] – the solution may be difficult, however, e. g., if the center points of the ellipsoids differ. In [19], a general ellipsoid-in-ellipsoid constraint was reformulated with respect to the linearization of a shape-matrix with heuristically chosen linearization point. The use of heuristics is not appropriate for the purpose of recursive feasibility in this paper. Another possibility is to use polytopic under-approximations Σ ˇ ξ [ δ x ] Σ ξ [ δ x ] , leading to an ellipsoid-in-polytope constraint. This is simple to formulate, but the computation of Σ ˇ ξ [ δ x ] is in general only feasible for small dimensions n ξ . Since the uncertainties arising from the AoI typically lead to terminal systems of higher dimension, polytopic computations are also not applicable here. Therefore, ellipsoid-in-ellipsoid constraints with equivalent center-points are used here:

Proposition 1.

Given two ellipsoids E 1 ε q , E 1 and E 2 ε q , E 2 with q R n q , then E 1 E 2 iff:

E 2 E 1 0 .

Hence, constraint (13) is reformulated to:
(14a) ξ ¯ t | k = 0 ,
(14b) Σ ξ Ξ t | k 0 .
Since the terminal system is used within the optimization problem for the end of the prediction horizon, the conservatism introduced by (14a) reduces if the prediction horizon increases. To likewise satisfy the input constraint (2b) with use of (11) and the terminal state ξ t | k , the δ u -confidence ellipsoid for the input ϕ t | k N ϕ ¯ t | k , Φ t | k has to lie within the admissible input set, i. e.:

(15) Φ t | k [ δ u ] = ε K ξ ξ ¯ t | k , c [ δ u ] · K ξ Ξ t | k K ξ U .

If the constraints (14) are satisfied, (15) is satisfied too, if:

(16) ε 0 , c [ δ u ] · K ξ Σ ξ K ξ U

is guaranteed. Note that the satisfaction of (16) can be proven offline with determination of K ξ and Σ ξ .

4.4 Constrained stochastic optimization

Before the optimization problem is stated, the existence of a terminal control law, which complies with the restrictions above, is assumed.

Assumption 4.

For the terminal system (10) with delay-time τ and the communication link probabilities { p ( k + t ) } t = 1 , there exist a linear control law (11), a terminal weight P ξ and a probabilistic terminal set Σ ξ [ δ x ] such that the following holds:

  1. The delay-time τ is chosen to satisfy the threshold:

    (17) P ( a ( k ) τ ) δ τ , k N 0

  2. The terminal set is the δ x -confidence ellipsoid of the invariant covariance Σ ξ for the closed loop system:

    ξ ( k + 1 ) = A ξ , c l ξ ( k ) + E ξ w ( k ) , w ( k ) N 0 , W ˆ .

  3. The terminal set is a subset of the state admissible set, i. e., Σ ξ [ δ x ] X k τ × × X k , k N 0 .

  4. The terminal control law u ( k ) = K ξ · ξ ( k ) satisfies the probabilistic input constraint according to (16).

  5. The weighting matrix P ξ of the terminal cost term solves a Lyapunov function in the sense that:

    ξ ( k + 1 ) P ξ 2 ξ ( k ) P ξ 2 + ξ ( k ) Q ξ 2 0

    holds for the weights of the stage cost function (3):

    Q ξ = diag K ξ Q u K ξ , 0 , , 0 , Q x .

With these assumptions, a common stochastic finite-horizon cost-function is used with the terminal weight P ξ and the stage cost (3):

(18) J H ( k ) = E k ξ H | k P ξ 2 + t = 0 H 1 l x t | k , u t | k

The optimization problem to be solved in each k with a ( k ) = 0 is eventually:

(19a) J H ( k ) = min { V t | k , M t , r | k } { r , t } = 0 { r ˆ t | k , H 1 } J H ( k )
(19b) s. t.: x t + 1 | k = A x t | k + B u t | k + E w ( k + t ) , u t | k satisfies (8) , γ t | k ( x ) , γ t | k ( u ) given by (9a) , (9b) ,
(19c) ξ H | k = x H τ | k x H | k ξ ¯ H | k = 0 , Σ ξ Ξ H | k 0 .

Note, that reformulation of (18) into a quadratic cost-functional is well-known in literature, e. g., cf. [10], [20], and that [15] already provides a method to reformulate (19b) into an LMI, such that (19) results in an SDP. The optimal solution { V t | k , M t , r | k } { t , r } = 0 H 1 together with (8) provides admissible control inputs for the next H time-steps.

As the NCS is affected by an unbounded additive disturbance, and thus the state may become unbounded, too, the optimization problem (19) may also become infeasible. Therefore, and similar to the findings in [20], the use of positive slack variables is proposed to soften both types of constraints (19b) according to [21]. As discussed in [20], this simply augments the set of feasible initial states to R n x , and recursive feasibility of (19) can be ensured.

Theorem 1.

If (19) is feasible in k = 0 without slack variables, the proposed control scheme guarantees the satisfaction of constraints (2) for all k > 0 at least with probability δ τ in each k.

Proof.

For k + 1, consider the following solution candidate:

(20) { u t | k + 1 } t = 0 H 1 = u 1 | k , , u H 1 | k , K ξ ξ H | k .

Each predicted input { u r | k } r = 1 H 1 is applicable to the NCS with probability α r | k , and the γ r | k ( u ) -confidence ellipsoids of the input lie within U. Hence, the probability that each input lies within the input admissible set U is at least δ u . For k + H, an admissible input can be predicted with respect to the terminal control law u H 1 | k + 1 = K ξ · ξ H | k . The use of the terminal control law implies the satisfaction of the input at least with probability δ u , cf. Asm. 4. 4. Eventually, the input constraint (2b) is satisfied by (20), under the condition of an applicable terminal control law, which is given at least with probability δ τ , see Asm. 4. 1. The satisfaction of the state constraints follows analogously. Note, that possible communication dropouts are stochastically independent, such that α t 1 | k + 1 : = α t | k holds recursively (cf. Asm. 2), and satisfaction of constraints (2) for all times k + t , t N > 0 follows directly by induction over k.  □

Obviously, constraint satisfaction of the predictive control scheme depends on the threshold δ τ for the AoI and the underlying probability of the communication link. In general, the link probability does not depend on the control system and is therefore a quantity solely depending on the communication network. However, Asm. 4. 1 applies to the communication network and is in general guaranteed with values δ τ close to one, see, e. g., [22] for variable transmission power of wireless networks.

Since the disturbance w ( k ) is stochastically distributed with unbounded support, the controlled NCS results in a perturbed Markov process. Stability of a Markov process (in the sense of mean-square stability) implies a bounded average for the expected step costs [23, Prop. 1. iv)], which is often used to show stability in stochastic model predictive control schemes, e. g., [24], [25], [26], [10]. With Asm. 4. 5, problem (19) is a common stochastic control problem with tailored control law and likelihoods to satisfy the constraints. In addition, the existence of an admissible solution-candidate is given by (20). Therefore, the decrease of expected costs can be shown analogously to the references above, and mean-square stability is obtained, but conditioned by δ τ .

5 Numerical simulation

This section describes the application of the proposed method to the example of an unmanned aerial vehicle (UAV) moving in a constrained area. Usually, the control of UAVs is twofold: i) an inner control loop stabilizes the UAV in hovering mode, and ii) an outer control loop determines the path to follow. While the inner loop controls states which are locally measurable, as rotor speeds (in the case of drones) or the yaw, roll, and pitch angles, the outer loop controls the position and height in the 3-dimensional air-space. Therefore, the outer control loop is responsible for safe operation with respect to the UAV’s environment, i. e., physical and virtual obstacles like buildings or no flight zones in urban areas. In contrast to the inner loop, the states of the outer loop, i. e., the exact position in 3D, is often not locally measurable because of an unmeasurable drift of the gyros used for stabilization, uneven terrain (e. g., if cameras or infrared sensors are used), or varying air pressure. Hence, the position is often measured by GPS for outdoor motion, and via motion tracking systems in case of indoor motion. Both methods have the drawback that the control loop relies on these external signals. In motion tracking systems, the position of the UAV is pre-processed by a stationary unit, and afterwards transmitted to the UVA via a communication link (commonly via wireless LAN). In contrast, GPS is based on the run-time measurement of electromagnetic satellite-signals and relies therefore on possibly faulty connectivity to satellites. However, in both cases the availability of the exact position measurements for the outer UAV control loop can be modeled by the communication link in Fig. 1b, and its transmission probability. Here, the UAV motion is modeled in a two-dimensional space with integrator dynamics, and with an underlying inner control loop, as proposed in [11]. The dynamics (1) is specified by:

A = A ˜ A ˜ , B = B ˜ B ˜ , E = E ˜ E ˜ , A ˜ = 1 0.79 0 0.61 , B ˜ = 0.21 0.39 , E ˜ = 0.5 1 ,

with state vector x = p x , v x , p y , v y , input vector u = a x , a y , and disturbances w = w x , w y . Here, { p x , p y } are the positions, { v x , v y } the velocities, { a x , a y } the acceleration, and { w x , w y } the disturbances in x- and y-direction. In contrast to [11], only one disturbance per direction is considered, modeling a wind force, and affecting both the position and the velocity of the UAV. The disturbance is modeled as a Gaussian random variable, and the communication link is considered to transmit with constant probability:

p ( k ) = 0.7 , W ( k ) W ˆ = 0.6 0.6 k .

Position and velocity are constrained according to (2a) with likelihood δ x = 0.95 and the admissible set:

X k = x 30 p x 50 8 v x 7 25 p y 20 8 v y 8 , k .

Similarly, the input is constrained according to (2b) with likelihood δ u = 0.9 and the admissible set U = u u 7.5 . Caused by the uncertain AoI, the threshold δ τ is selected to 99 %. With respect to the constant link probability and according to (17), the threshold results in δ τ 99.19 % with delay-length τ = 4 for the terminal system and the control law. Note, that τ = 4 implies H 4, and that the expected value of the last 4 predicted states have to lie in the origin (cf. (19c)).

The feedback matrix of the terminal control law K τ is chosen to:

K τ = 0.058 1.172 0 0 0 0 0.058 1.172

satisfying the assumptions 4. 3 and 4. 4.

The prediction horizon is chosen to H = 12, and the stage cost function from [12] is used with:

Q x = Q ˜ Q ˜ , Q ˜ = 10 0.1 , Q u = 0.1 0.1 , S = 1 .

Starting from an initial state x 0 = 35 , 2 , 20 , 4 (blue square), Fig. 3 shows 200 Monte Carlo simulations with 50 time-steps each in both directions of the state space, where each color denotes a different simulation. Note, that in none of these simulations the use slack variables was needed. For simulation, the covariance of the disturbance is chosen to W ( k ) = W ˆ. All simulations show that system trajectories converge into a neighborhood of the origin. The neighborhood is significantly smaller than the 95 %-confidence ellipsoid of Σ ξ , which is identical to the region for which the control law (11) is defined.

Figure 3 
Simulation results for 200 Monte-Carlo simulations with the initial state marked as blue square, the admissible state space as light gray box, and the projection of the terminal set onto the corresponding direction as red ellipsoid.
Figure 3

Simulation results for 200 Monte-Carlo simulations with the initial state marked as blue square, the admissible state space as light gray box, and the projection of the terminal set onto the corresponding direction as red ellipsoid.

Figure 4 shows the distribution of the AoI values for each time-step k of the simulations. Even if the probability is small, the maximum value for the AoI is 8, thus larger than τ = 4. Note, that the control strategy only guarantees probabilistic constraint satisfaction and mean-square stability as long as a ( k ) τ. By a solution of (20) without the use of slack variables, it provides a suitable control input as long as a ( k ) < H. Here, suitability means that a possible violation of the state and/or input constraints with an aged control input u ( k ) = u k | k t , τ t < H is considered in the optimal control problem, together with the likelihoods { γ k ( u ) , γ k ( x ) }. Since aged control inputs are rarely used (according to Fig. 4 only with probability around 40 % in each time-step), the satisfaction of constraints is much higher than { δ u , δ x }. Here, the admissible state space is never left over all simulations.

Figure 4 
Distribution of the Age of Information for the Monte-Carlo simulations and each time-step of the simulation time.
Figure 4

Distribution of the Age of Information for the Monte-Carlo simulations and each time-step of the simulation time.

6 Conclusion

This article has introduced a model predictive control scheme for constrained linear systems with uncertain availability of state information. The proposed method is based on a recently suggested optimal control scheme for linear systems with uncertain communication, which projects the uncertainty of available information onto a tailored likelihood to satisfy state and input constraints. Here, solving such a stochastic optimal control problem in receding horizon fashion implies the use of terminal constraints to guarantee recursive feasibility and stability of the control scheme. Due to the uncertain communication, the use of a state-augmented terminal system with delayed control law is proposed. In contrast to existing methods, which use reachable sets as terminal sets, the terminal set here is derived with respect to the confidence ellipsoid of an invariant covariance. This approach scales better with the dimension of the system and the communication link probability, compared to the use of polytopic sets.

The simulation results show that the proposed predictive control scheme is able to steer a networked control system with high (and specified probability) into the target, while the AoI changes probabilistically. Although the control scheme is based on probabilistic constraint satisfaction, the system state remains in the admissible state space. At the same time, the state trajectory does not require considerable safety margins to the state constraints, thus introducing only little conservatism. When approaching the target state, the proposed control scheme keeps the system state in a much tighter neighborhood than approaches which are based solely on a worst case delay of the communication link.

It is worth noting, that the computational complexity of the predictive control does not significantly grow with the complexity of the communication network between sensor and controller, or if the threshold δ τ (and therefore the dimension of the terminal system) is increased.

This control scheme opens the field to distributed predictive control systems using collectively a single communication network – this is matter of future work of the authors.

Award Identifier / Grant number: SPP1914

Funding statement: This work is financially supported in part by the German Research Foundation (DFG) within the research priority program SPP1914 – Cyberphysical Networking.

About the authors

M. Sc. Jannik Hahn

Jannik Hahn is research assistant with the Control and System Theory Group, Department of Electrical Engineering and Computer Science, Universität Kassel. His research activities are focussed on robust and stochastic distributed model predictive control.

Prof. Dr.-Ing. Olaf Stursberg

Olaf Stursberg is Full Professor and Head of the Control and System Theory Group in the Department of Electrical Engineering and Computer Science at University of Kassel. His main research areas include methods for optimal and predictive control of networked and hierarchical systems, techniques for analysis and design of hybrid dynamic systems, and the control of stochastic, uncertain and learning systems in different domains of application.

References

1. J. P. Hespanha, P. Naghshtabrizi and Y. Xu, “A survey of recent results in networked control systems,” Proc. of the IEEE, vol. 95, no. 1, pp. 138–162, 2007.10.1109/JPROC.2006.887288Search in Google Scholar

2. I. Akyildiz, W. Su, Y. Sankarasubramaniam, et al., “Wireless sensor networks: A survey,” Computer Networks, vol. 38, no. 4, pp. 393–422, 2002.10.1016/S1389-1286(01)00302-4Search in Google Scholar

3. E. A. Lee, “Cyber physical systems: Design challenges,” in 11th IEEE Int. Symp. on Object and Component-oriented Real-time Distributed Computing, IEEE, 2008, pp. 363–369.10.1109/ISORC.2008.25Search in Google Scholar

4. D. Groß and O. Stursberg, “Distributed predictive control of communicating and constrained systems,” ZAMM-Journal of Applied Mathematics and Mechanics, vol. 94, no. 4, pp. 303–316, 2014.10.1002/zamm.201100166Search in Google Scholar

5. D. Jia, K. Lu and J. Wang, “On the network connectivity of platoon-based vehicular cyber-physical systems,” Transportation Research Part C: Emerging Technologies, vol. 40, pp. 215–230, 2014.10.1016/j.trc.2013.10.006Search in Google Scholar

6. M. H. Yamchi and R. M. Esfanjani, “Distributed predictive formation control of networked mobile robots subject to communication delay,” Robotics and Autonomous Systems, vol. 91, pp. 194–207, 2017.10.1016/j.robot.2017.01.005Search in Google Scholar

7. J. Hahn, R. Schoeffauer, G. Wunder, et al., “Distributed MPC with prediction of time-varying communication delay,” IFAC-PapersOnLine, vol. 51, no. 23, pp. 224–229, 2018.10.1016/j.ifacol.2018.12.039Search in Google Scholar

8. J. Hahn and O. Stursberg, “Robust distributed MPC for disturbed affine systems using predictions of time-varying communication,” in 18th European Control Conf., 2019, pp. 56–62.10.23919/ECC.2019.8795773Search in Google Scholar

9. A. Mesbah, “Stochastic model predictive control: An overview and perspectives for future research,” IEEE Control Systems Magazine, vol. 36, no. 6, pp. 30–44, 2016.10.1109/MCS.2016.2602087Search in Google Scholar

10. K. Okamoto and P. Tsiotras, “Stochastic model predictive control for constrained linear systems using optimal covariance steering,” arXiv:1905.13296, 2019.10.1109/CDC40024.2019.9029353Search in Google Scholar

11. L. Blackmore, M. Ono and B. C. Williams, “Chance-constrained optimal path planning with obstacles,” IEEE Trans. on Robotics, vol. 27, no. 6, pp. 1080–1094, 2011.10.1109/TRO.2011.2161160Search in Google Scholar

12. J. Hahn and O. Stursberg, “Uncertain Aol in stochastic optimal control of networked and constrained LTI systems,” arXiv:2104.12435, 2021.10.1515/auto-2021-0125Search in Google Scholar

13. S. C. Smith and P. Seiler, “Estimation with lossy measurements: Jump estimators for jump systems,” IEEE Transactions on Automatic Control, vol. 48, no. 12, pp. 2163–2171, 2003.10.1109/TAC.2003.820140Search in Google Scholar

14. R. Schoeffauer and G. Wunder, “Predictive network control and throughput sub-optimality of max weight,” in European Conf. on Networks and Comm., 2018, pp. 1–6.10.1109/EuCNC.2018.8442615Search in Google Scholar

15. J. Hahn and O. Stursberg, “Balanced stochastic optimal control of uncertain linear systems with constraints,” in Proc. of the IFAC World Congress, 2020, pp. 7254–7260.10.1016/j.ifacol.2020.12.535Search in Google Scholar

16. P. J. Goulart, E. C. Kerrigan and J. M. Maciejowski, “Optimization over state feedback policies for robust control with constraints,” Automatica, vol. 42, no. 4, pp. 523–533, 2006.10.1016/j.automatica.2005.08.023Search in Google Scholar

17. M. Farina, L. Giulioni, L. Magni, et al., “A probabilistic approach to model predictive control,” in 52nd IEEE Conf. on Decision and Control, 2013, pp. 7734–7739.10.1109/CDC.2013.6761117Search in Google Scholar

18. S. Boyd, L. El Ghaoui, E. Feron, et al., Linear matrix inequalities in system and control theory. SIAM, 1994.10.1137/1.9781611970777Search in Google Scholar

19. L. Asselborn and O. Stursberg, “Probabilistic control of uncertain linear systems using stochastic reachability,” IFAC-PapersOnLine, vol. 48, no. 14, pp. 167–173, 2015.10.1016/j.ifacol.2015.09.452Search in Google Scholar

20. J. A. Paulson, E. A. Buehler, R. D. Braatz, et al., “Stochastic model predictive control with joint chance constraints,” International Journal of Control, vol. 93, no. 1, pp. 126–139, 2020.10.1080/00207179.2017.1323351Search in Google Scholar

21. E. C. Kerrigan and J. M. Maciejowski, “Soft constraints and exact penalty functions in model predictive control,” 2000.Search in Google Scholar

22. S. Lin, F. Miao, J. Zhang, et al., “ATPC: Adaptive transmission power control for wireless sensor networks,” ACM Trans. Sen. Netw., vol. 12, no. 1, Mar. 2016, ISSN: 1550-4859.10.1145/1182807.1182830Search in Google Scholar

23. T. Morozan, “Optimal stationary control for dynamic systems with Markov perturbations,” Stochastic Analysis and Applications, vol. 1, no. 3, pp. 299–325, 1983.10.1080/07362998308809016Search in Google Scholar

24. M. Cannon, B. Kouvaritakis and X. Wu, “Probabilistic constrained MPC for multiplicative and additive stochastic uncertainty,” IEEE Trans. on Automatic Control, vol. 54, no. 7, pp. 1626–1632, 2009.10.1109/TAC.2009.2017970Search in Google Scholar

25. L. Hewing and M. N. Zeilinger, “Stochastic model predictive control for linear systems using probabilistic reachable sets,” in IEEE Conf. on Decision and Control, 2018, pp. 5182–5188.10.1109/CDC.2018.8619554Search in Google Scholar

26. M. Lorenzen, F. Dabbene, R. Tempo, et al., “Constraint-tightening and stability in stochastic model predictive control,” IEEE Trans. on Automatic Control, vol. 62, no. 7, pp. 3165–3177, 2017.10.1109/TAC.2016.2625048Search in Google Scholar

Received: 2021-02-01
Accepted: 2021-08-24
Published Online: 2021-09-09
Published in Print: 2021-09-27

© 2021 Hahn and Stursberg, published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 5.11.2024 from https://www.degruyter.com/document/doi/10.1515/auto-2021-0033/html
Scroll to top button