1. Introduction
1.1. Brief survey
Constrained optimization is the process of optimizing an objective function with respect to some variables in the presence of constraints on those variables. The objective function is either a cost function or energy function, which is to be minimized, or a reward function or utility function, which is to be maximized. Constraints can be either hard constraints, which set conditions for the variables that are required to be satisfied, or soft constraints, which have some variable values that are penalized in the objective function if, and based on the extent that, the conditions on the variables are not satisfied (see, for example [
3,
4,
13,
17,
18] and [
22]).
All control strategies in the most publications, treated as
Static Optimization Methods (SOM), in continuous-time may be represented in the following form
where
is a convex (not obligatory strongly convex) mapping,
is the admissible convex set of arguments and the process
is generated by the simple ordinary differential equation (ODE)
with any initial conditions
. The relation (
2) is referred hereafter to as a
static plant. All known procedures of SOM differ only in designing of control action
(or an optimization algorithm) as a function of the current state
(Markov’s strategy) or more profound available history, namely,
.
Here we will consider more general, and hence, more complex situation when the process
is generated by the
dynamic plant
where the vector function
f in the right-hand side is supposed to be unknown but belonging to some class
of nonlinearities. This problem is more closed to the, so-called,
Extremum Seeking Problem [
1,
12,
14,
23], where the nonlinear dynamics includes the first order derivatives only. So, in [
24], several optimization schemes are considered and there is shown that under appropriate conditions these schemes achieve extremum point from an arbitrarily large domain of initial conditions if the parameters in the controller are appropriately adjusted. This approach was applied in [
15] for two levels plant’s economic optimization. Many advanced process control systems use some form of model predictive control approach [
5,
26]. The paper [
20] describes a new algorithm for extremum seeking using stochastic on-line gradient estimation. The paper [
7] deals with the problem of constrained optimization in dynamic linear time-invariant (LTI) systems characterized by a control vector dimension less than that of the system state vector. The finite-time convergence to a vicinity of order
of the optimal equilibrium point is proved. In [
8] a variable structure convex programming based control for a class of linear uncertain systems with accessible state is presented.
In this paper we consider a class of controlled plants with dynamics governed by a vector system of the second order ordinary differential equations (ODE) with unknown right-hand side. All mechanical Lagrange models belong to this class. The state variables and their velocities are assumed to be measurable. We design a controller minimizing a loss function subjected to a set of constraints to the state of the controlled plant. The designed control action is admitted to be a function of the current sub-gradients of loss function and constraints only, which also supposed to be measurable on-line. The control is designed based on SDM (Subgradient Descent Method) - version [
19,
21] of Integral Sliding Mode (ISM) concept [
9,
25] aimed to minimize "on average" a given convex (not obligatory strongly convex) cost function of the current state under a set of given constraints. An optimization type algorithm is developed and analyzed using ideas of SDM technique [
3]. We prove the reachability of the "
desired regime" (nonstationary analogue of sliding surface) [
9] from the beginning of the process and obtaining an explicit upper bound for the
cost function decrement, that is, the
convergence is proven and the rate of convergence is estimated as
. This paper generalizes the approach, suggested in [
11] for unconstrained dynamic optimization, to the constraint optimization problem realized by an uncertain second order dynamic plant.
1.2. Main contributions
Robust Tracking problem is reformulated as a Constrained Optimization realized by a dynamic plant with unknown (but bounded) right-hand side.
The cost as well as the constraints are admitted to be convex but not obligatory strictly or strongly convex.
Mirror Descent Method (MDM) and ASG – Version of Sliding Mode Control are suggested and realized.
The convergence of the obtained trajectories of controlled uncertain plant to the corresponding admissible zone closed the minimal point is realized.
2. Uncertain plant description and admitted dynamic zone
2.1. Dynamic model
The second order dynamic model (
3) can be represented in the following extended format
Here the extended state variables
are the current coordinates and their velocities at time
Function
is partially continuous in all arguments and admits to be unknown but bounded as
with final positive constants
,
, and
. Hereafter the symbol
means the Euclidean norm.
2.2. Reference trajectory, tracking error dynamics, and admissible
zone
The aim of the controller (which will be exactly formulated below) is to realize the tracking of the state
for the given reference trajectory
. Define the
tracking error as
where
is the continuously differentiable trajectory to be tracked satisfying
In view of that, the error tracking dynamics can be represented as follows
Let us require that the dynamics of should be realized after time within a bounded admissible zone
Let the loss function
be a convex. For example, the following two functions belong to the considered class of the convex loss functions to be optimized:
2.3. Basic assumptions
A1The current states
of the plant (
4) are supposed to be measurable (available) on-line for all
.
A2 The function
, satisfying (
5), is piecewise continuous in all arguments and admits to be unknown.
A3The current state of the reference trajectory are also supposed to be available on-line for any .
A4Here we assume that sub-gradient
1 of the loss function
is available on-line for a current time
,and the set of minimizers
of
on the set
includes the origin
, that is,
A5The admissible set is non empty convex compact, i.e., .
3. Desired dynamics
3.1. Mirror descent method in continuous time
Let us apply mirror descent approach, using the Legendre-Fenchel transformation [
16] as follows. For any
define
so that (see, for instance, [
2,
10])
Define the dynamics for the vector-function
as
Remark 1.
The second differential equation in (12) can be inegrated as follows
Therefore, for all because of convexity and due to (10)–(11).
3.2. Why the dynamics be desired
The following theorem explains why the dynamics may be considered as a desired one.
Theorem 1.
Under Assumptions A1-A5 on the trajectories , generated by (12), for all the following propertry holds
where
Proof. Defining
,
, we have from (
12)
Due to the convexity property for
, we have
and, in view of the relation
it follows
or equivalently,
After integration we get
which implies
Since by (
10) and (
11)
and defining
we get
□
Example 1.
To calculate according (14), it is sufficient to note that the soltion of the problem
4. Robust controller design
4.1. Auxilary sliding variable and its dynamics
Introduce a new auxilary variable (sliding variable)
Notice that the function
is measurable on-line, and that the situation when
corresponds exactly the desired regime (
12), starting from the moment
. Then for
in view of (
8) and the first equation in (
12) we have
4.2. Robust control structure
Since
and taking
we get
which implies
This means that for all
where
Finally, the robust control is
where
Remark 1.
If we wish to get , we need to complete the identity
Since , we may conclude that parameters and initial conditions should be consistent in the sence that
Remark 2.
For the example, for Eucidean r-ball in being the admissible set , from (10)–(11) one has
Notice, that -function (11) is nondifferential in the points of r-sphere of ball, and it is continuous differential in all other points of . The formulas in (21), (24) are presented as their continuous versions on ball -function (11) including the r-sphere.
4.3. Main result
We are ready to formulate the main result.
Theorem 1.
Under Assumptions A1-A5 the robust control (18)-(19) with parameter satisfying (20), provides the property
for all and any regularizing parameter
Proof. Since in view of the relation (
20) of the parameter
and initial conditions
the auxiliary variable
for all
starting from the beginning of the control process. Using the formula (
13) for
we obtain (
25). □
5. Discussion
Equations (15), (20) hold under , at the following cases:
Zero initial conditions , . Thus, for arbitrary (see, as an example, the 1st item in loss function (9)).
Non-zero initial conditions , are collinear oppositely directed vectors. Therefore, and exist (see, as an example, the 1st item in loss function (9)).
Equation (23) holds under non-zero vector η with a sufficiently small and for (see, as an example, the 2nd item in loss function (9)).
6. Conclusion
- The constrained optimization problem is addressed in this study using a second-order differential controlled plant with an unknown (but bounded) right side of the model.
- The desired dynamics in the tracking error variables is designed based on Mirror Descent Method.
- The continuous-time convergence to the set of minimizing points is established, and the associated rate of convergence has been analytically evaluated.
- The robust controller, containing both the continuous (compensating) and the discontinuous , is proposed the ASG-version of Integral Sliding Mode approach.
- The suggested controller, under the special realations of it parameters with the initial conditions, is proved to provide the desired regime from the beginning of the control process.
- This method may has several applications in the development of robust control in mechanical systems, including soft robotics and moving dynamic plants.
Author Contributions
Conceptualization, A.V. and A.P.; methodology, A.V. and A.P.; formal analysis, A.V. and A.P.; writing—original draft preparation, A.V. and A.P.; writing—review and editing, A.V.; supervision, A.P.; project administration, A.P.; funding acquisition, A.V. All authors have read and agreed to the published version of the manuscript.
Funding
A.V. is entitled to a 100 percent discount for publication in this special issue.
Conflicts of Interest
The authors declare no conflict of interest. Declare conflicts of interest or state “The authors declare no conflict of interest.” Authors must identify and declare any personal circumstances or interest that may be perceived as inappropriately influencing the representation or interpretation of reported research results. Any role of the funders in the design of the study; in the collection, analyses or interpretation of data; in the writing of the manuscript; or in the decision to publish the results must be declared in this section. If there is no role, please state “The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results”.
Abbreviations
The following abbreviations are used in this manuscript:
ASG |
Average Sub-Gradient |
SDM |
Subgradient Descent Method |
ISM |
Integral Sliding Mode |
SOM |
Static Optimization Methods |
ODE |
Ordinary Differential Equation |
References
- Ariyur Kartik, B. and Krstic, M., Real-time optimization by extremum-seeking control. John Wiley & Sons, 2003.
- Ben-Tal, A. and Nemirovski, A., The Conjugate Barrier Mirror Descent Method for Non-Smooth Convex Optimization, Preprint of the Faculty of Industr. Eng. Manag., Technion – Israel Inst. Technol., Haifa, 1999.
- Bertsekas, D. P., Constrained Optimization and Lagrange Multiplier Methods, New York: Academic Press. ISBN 0-12-093480-9, 1982.
- Dechter, R., Constraint Processing, Morgan Kaufmann. ISBN 1-55860-890-7, 2003.
- Dehaan, D. and Guay, M., "Extremum-seeking control of state-constrained nonlinear systems", Automatica, 2005, 41(9), 1567-1574.
- Feijer, D. and Paganini. F., "Stability of primal–dual gradient dynamics and applications to network optimization", Automatica, 2010, 46(12), 1974–1981.
- Ferrara, A. and Utkin, V.I., "Sliding Mode Optimization in Dynamic LTI Systems",Journal of Optimization Theory and Applications, 2002, 115(3), 727–740.
- Ferrara, A., A variable structure convex programming based control approach for a class of uncertain linear systems, Systems & Control Letters, 2005, 54(6), 529-538.
- Fridman, L., Poznyak, A. and Bejarano, F. J., Robust Output LQ Optimal Control via Integral Sliding Modes. Birkhäuser, Springer Science and Business Media, New York, 2014.
- Juditsky, A. B., Nazin, A. V., Tsybakov, A. B., Vayatis, N. Recursive aggregation of estimators by the mirror descent algorithm with averaging. // Probl. Inf. Transm., 2005, 41(4), 368–384; translation from Probl. Peredachi Inf., 2005, No. 4, 78–96.
- Poznyak, A.S., Nazin A.V., and Alazki H. Integral Sliding Mode Convex Optimization in Uncertain Lagrangian Systems Driven by PMDC Motors: Averaged Subgradient Approach // IEEE Transactions on Automatic Control. 2021. Vol. 66, No. 9. P. 4267-4273 (1-8).
- Krstic, M. and Wang, H. H., "Stability of extremum seeking feedback for general nonlinear dynamic systems", Automatica, 2000, 36(4), 595-601.
- Prosser, M., Constrained Optimization by Substitution. Basic Mathematics for Economists, New York: Routledge. pp. 338–346. ISBN 0-415-08424-5, 1993.
- Rastrigin, L.A. Systems of extremal control // Nauka, Moscow, 1974 (in Russian).
- Rawlings, J. B., and Amrit, R., Optimizing process economic performance using model predictive control, in Nonlinear Model Predictive Control, Springer, Berlin, Heidelberg, 2009, 119-138.
- Rockafellar, R.T. Convex analysis, Princeton University Press, Princeton, 1970.
- Rossi, F., van Beek, P. and Walsh, T. (eds.), Chapter 1 - Introduction, Foundations of Artificial Intelligence, Handbook of Constraint Programming, Elsevier, 2, pp. 3–12, doi:10.1016/s1574-6526(06)80005-2. [CrossRef]
- Leader, J. J., Numerical Analysis and Scientific Computation, Addison Wesley. ISBN 0-201-73499-0. 2004.
- Nazin, A.V., "Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization", Automation and Remote Control, January 2018, 79(1), 78–88.
- Solis, C.U., Clempner, J.B. and Poznyak, A.S., Extremum seeking by a dynamic plant using mixed integral sliding mode controller with synchronous detection gradient estimation, International journal of Robust and Nonlinear Control, 2018, 29(3), 702-714.
- Simpson-Porco, J. W., Input/output analysis of primal-dual gradient algorithms. In Communication, Control, and Computing (Allerton), 2016, 54th Annual Allerton Conference, IEEE, 2016, 219-224.
- Sun, Wenyu and Yua, Ya-Xiang, Optimization Theory and Methods: Nonlinear Programming, Springer, ISBN 978-1441937650, 2010.
- Tan, Y., Moase, W. H., Manzie, C., Nešić, D., & Mareels, I. M. Y. (2010, July). Extremum seeking from 1922 to 2010. In Control Conference (CCC), IEEE, 29th Chinese,2010, 14-26.
- Tan, Y., Nešić, D. and Mareels, I., "On non-local stability properties of extremum seeking control", Automatica,2006, 42(6), 889-903.
- Utkin, V., Sliding Modes in Control Optimization, Springer Verlag, Berlin, 1992.
- Chunlei, Z. and Ordóñez, R., "Robust and adaptive design of numerical optimization-based extremum seeking control." Automatica 45.3 (2009): 634-646.
1 |
Recall that a vector , satisfying the inequality + for all is called the sub-gradient of the function at the point and is denoted by which is the set of all sub-gradients of F at the point x. If is differentiable at a point x, then . In the minimal point we have . |
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).