Linear quadratic control: Difference between revisions

From Citizendium
Jump to navigation Jump to search
imported>Hendra I. Nurdin
(removed cats)
imported>Hendra I. Nurdin
(Further development)
Line 1: Line 1:
{{subpages}}
{{subpages}}


In [[control engineering]] and [[systems theory (engineering)|systems]] and [[control theory]], '''linear quadratic control''' or LQ control refers to controller design for a deterministic [[linear system|linear]] plant based on the minimization of a quadratic cost [[functional]]. The method is based on the [[state space formalism]] and is a fundamental concept in linear systems and control theory.  
In [[control engineering]] and [[systems theory (engineering)|systems]] and [[control theory]], '''linear quadratic control''' or LQ control refers to controller design for a deterministic [[linear system|linear]] plant based on the minimization of a quadratic cost [[functional]]. The method is founded on the [[state space formalism]] and is a fundamental concept in linear systems and control theory.  


There are two main versions of the method, depending on the setting of the control problem:
There are two main versions of the method, depending on the setting of the control problem:
#Linear quadratic control in discrete time
#Discrete time linear quadratic control  
#Linear quadratic control in continuous time
#Continuous time linear quadratic control  


The objective of LQ control is to find a control signal that minimizes a prescribed quadratic cost functional. In the so-called regulation problem, this functional can be viewed as an abstraction of the "energy" of the overall [[control system]] and minimization of the functional corresponds to minimization of that energy.
LQ control aims to find a control signal that minimizes a prescribed quadratic cost functional. In the so-called optimal regulator problem, this functional can be viewed as an abstraction of the "energy" of the overall [[control system]] and minimization of the functional corresponds to minimization of that energy.
 
==Discrete time linear quadratic control==
 
===Plant model===
In discrete time, the plant (the system to be controlled) is assumed to be linear with ''input'' <math>u_k</math>, ''state'' <math>x_k</math> and ''output'' <math>y_k</math>, and evolves in discrete time ''k''=0,1,... according to the following dynamics:
 
<math>x_{k+1}=Ax_{k}+B u_{k};\, x_0=x</math>
 
<math>y_{k}=Cx_{k}+D u_{k}, </math>
 
where <math>x,x_k \in \mathbb{R}^n</math>, <math>u_k \in \mathbb{R}^m</math> and <math>y_k \in \mathbb{R}^p</math> for all <math>k\geq 0</math> and <math>A,B,C,D</math> are real matrices of the corresponding sizes (e.g., for consistency, <math>A</math> should be of the dimension <math>n \times n</math> while <math>B</math> should be of dimension <math>n \times m</math>). Here <math>x</math> is the ''initial state'' of the plant.
 
===Cost function===
For a finite integer <math>K>0</math>, called the ''control horizon'' or ''horizon'', a real valued cost functional  <math>J</math> of the initial state ''x'' and the control sequence <math>u_0,u_1,\ldots,u_{K-1}</math> up to time <math>K-1</math> is defined as follows: 
 
<math>J^K(x,u_0,u_1,\ldots,u_{K-1})=\sum_{k=0}^{K-1}(x_k^T Q x_k+u_k^T R u_k) + x_K^T \Gamma x_K, \,\, (1)</math>
 
where <math>Q,\Gamma</math> are symmetric matrices (of the corresponding sizes) satisfying <math>Q,\Gamma \geq 0 </math> and <math>R</math> is a symmetric matrix (of the corresponding size) satisfying <math>R > 0 </math>. Note that each term on the right hand side of (1) are non-negative definite quadratic terms and may be interpreted as abstract "energy" terms (e.g., <math>x_k^T Q x_k</math> as the "energy" of <math>x_K</math>). The horizon <math>K</math> represents the terminal time of the control action. The final term <math>x_K^T \Gamma x_K</math> is called the terminal cost and it penalizes the energy of the plant at the final state <math>x_K</math> .     
 
===The LQ regulator problem in discrete time===
The objective of LQ control is to solve the ''optimal regulator problem'':
<blockquote>
'''(Optimal regulator problem)''' For a given horizon ''K'' and initial state ''x'', find a control sequence <math>\tilde u_0,\tilde u_1,\ldots,\tilde u_{K-1} \in \mathbb{R}^m</math> that ''minimizes'' the cost functional <math>J^K</math>, that is,<br><br>
 
<math>J(x,\tilde u_0,\tilde u_1,\ldots,\tilde u_{K-1}) \leq J(x,u_0,u_1,\ldots,u_{K-1}),\,\, (2) </math><br><br>
 
over all possible control sequences <math>u_0,u_1,\ldots,u_{K-1} \in \mathbb{R}^m</math>. </blockquote> 
 
The optimal regulator problem is a type of [[optimal control]] problem and the control sequence <math>\tilde u_0,\tilde u_1,\ldots,\tilde u_{K-1} \in \mathbb{R}^m</math> is an optimal control sequence.


==Related topics==
==Related topics==
[[Linear quadratic Gaussian control]]
[[Linear quadratic Gaussian control]]

Revision as of 06:25, 9 October 2007

This article is developed but not approved.
Main Article
Discussion
Related Articles  [?]
Bibliography  [?]
External Links  [?]
Citable Version  [?]
 
This editable, developed Main Article is subject to a disclaimer.

In control engineering and systems and control theory, linear quadratic control or LQ control refers to controller design for a deterministic linear plant based on the minimization of a quadratic cost functional. The method is founded on the state space formalism and is a fundamental concept in linear systems and control theory.

There are two main versions of the method, depending on the setting of the control problem:

  1. Discrete time linear quadratic control
  2. Continuous time linear quadratic control

LQ control aims to find a control signal that minimizes a prescribed quadratic cost functional. In the so-called optimal regulator problem, this functional can be viewed as an abstraction of the "energy" of the overall control system and minimization of the functional corresponds to minimization of that energy.

Discrete time linear quadratic control

Plant model

In discrete time, the plant (the system to be controlled) is assumed to be linear with input , state and output , and evolves in discrete time k=0,1,... according to the following dynamics:

where , and for all and are real matrices of the corresponding sizes (e.g., for consistency, should be of the dimension while should be of dimension ). Here is the initial state of the plant.

Cost function

For a finite integer , called the control horizon or horizon, a real valued cost functional of the initial state x and the control sequence up to time is defined as follows:

where are symmetric matrices (of the corresponding sizes) satisfying and is a symmetric matrix (of the corresponding size) satisfying . Note that each term on the right hand side of (1) are non-negative definite quadratic terms and may be interpreted as abstract "energy" terms (e.g., as the "energy" of ). The horizon represents the terminal time of the control action. The final term is called the terminal cost and it penalizes the energy of the plant at the final state .

The LQ regulator problem in discrete time

The objective of LQ control is to solve the optimal regulator problem:

(Optimal regulator problem) For a given horizon K and initial state x, find a control sequence that minimizes the cost functional , that is,



over all possible control sequences .

The optimal regulator problem is a type of optimal control problem and the control sequence is an optimal control sequence.

Related topics

Linear quadratic Gaussian control