×

Optimal control of uncertain quantized linear discrete-time systems. (English) Zbl 1330.93150

Summary: In this paper, the adaptive optimal regulator design for unknown quantized linear discrete-time control systems over fixed finite time is introduced. First, to mitigate the quantization error from input and state quantization, dynamic quantizer with time-varying step-size is utilized wherein it is shown that the quantization error will decrease overtime thus overcoming the drawback of the traditional uniform quantizer. Next, to relax the knowledge of system dynamics and achieve optimality, the adaptive dynamic programming methodology is adopted under Bellman’s principle by using quantized state and input vector. Because of the time-dependency nature of finite horizon, an adaptive online estimator, which learns a newly defined time-varying action-dependent value function, is updated at each time step so that policy and/or value iterations are not needed. Further, an additional error term corresponding to the terminal constraint is defined and minimized along the system trajectory. The proposed design scheme yields a forward-in-time and online scheme, which enjoys great practical merits. Lyapunov analysis is used to show the boundedness of the closed-loop system; whereas when the time horizon is stretched to infinity as in the case of infinite horizon, asymptotic stability of the closed-loop system is demonstrated. Simulation results on a benchmarking batch reactor system are included to verify the theoretical claims. The net result is the design of the optimal adaptive controller for uncertain quantized linear discrete-time systems in a forward-in-time manner.

MSC:

93C55 Discrete-time control/observation systems
93C05 Linear systems in control theory
93C40 Adaptive control/observation systems
90C39 Dynamic programming
49N90 Applications of optimal control and differential games
Full Text: DOI

References:

[1] DelchampsDF. Stabilizing a linear system with quantized state feedback. IEEE Transactions on Automatic Control1990; 35(8):916-924. · Zbl 0719.93067
[2] EliaN, MitterSK. Stabilization of linear systems with limited information. IEEE Transactions on Automatic Control2001; 46(9):1384-1400. · Zbl 1059.93521
[3] BrockettRW, LiberzonD. Quantized feedback stabilization of linear systems. IEEE Transactions on Automatic Control2000; 45(7):1279-1289. · Zbl 0988.93069
[4] LiberzonD. Hybrid feedback stabilization of systems with quantized signals. Automatica2003; 39: 1543-1554. · Zbl 1030.93042
[5] LewisFL, SyrmosVL. Optimal Control (2nd ed). Wiley: Hoboken, NJ, 1995.
[6] SiJ, BartoAG, PowellWB, WunschD. Handbook of Learning and Approximate Dynamic Programming. Wiley: New York, 2004.
[7] WerbosPJ. A menu of designs for reinforcement learning over time. In Neural Nework for Control. MIT Press: Cambridge, 1991; 67-95.
[8] WatkinsC. Learning from delayed rewards. PhD Dissertation, Cambridge University, England, 1989.
[9] BradtkeSJ, YdstieBE. Adaptive linear quadratic control using policy iteration. Proceedings of American Control Conference, Baltimore, MD, 1994; 3475-3479.
[10] ChenZ, JagannathanS. Generalized Hamilton-Jacobi-Bellman formulation based neural network control of affine nonlinear discrete‐time systems. IEEE Transactions on Neural Networks2008; 19(1):90-106.
[11] DierksT, JagannathanS. Online optimal control of affine nonlinear discrete‐time systems with unknown internal dynamics by using time‐based policy update. IEEE Transactions on Neural Networks and Learning Systems2012; 23(7):1118-1129.
[12] BeardRW. Improving the closed‐loop performance of nonlinear systems. PhD Dissertation, Electr. Eng. Dept., Rensselaer Polytechnic Institute, USA, 1995.
[13] HeydariA, BalakrishnanSN. Finite‐horizon control‐constrained nonlinear optimal control using single network adaptive critics. IEEE Transactions on Neural Networks and Learning Systems2013; 24(1):145-157.
[14] WangFY, JinN, LiuD, WeiQ. Adaptive dynamic programming for finite‐horizon optimal control of discrete‐time nonlinear systems with_ϵ‐error bound. IEEE Transactions on Neural Networks2011; 22(1):24-36.
[15] BertsekasDP. Dynamic programming and suboptimal control: a survey from adp to mpc. European Journal of Control2005; 11: 310-334. · Zbl 1293.49056
[16] XuH, JagannathanS, LewisFL. Stochastic optimal control of unknown networked control systems in the presence of random delays and packet losses. Automatica2012; 48: 1017-1030. · Zbl 1244.93177
[17] XuH, JagannathanS. Stochastic optimal controller design for uncertain nonlinear networked control system via neuro dynamic programming. IEEE Transactions on Neural Networks and Learning Systems2013; 24(3):471-484.
[18] NarendraKS, AnnaswamyAM. Stable Adaptive Systems. Prentice‐Hall: New Jersey, 1989. · Zbl 0758.93039
[19] SandbergIW. Notes on uniform approximation of time‐varying systems on finite time intervals. IEEE Transactions on Circuits and Systems-I, Fundamental Theory and Application1998; 45(8):863-865. · Zbl 0952.94027
[20] CybenkoG. Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals, and Systems1989; 2: 303-314. · Zbl 0679.94019
[21] LewisFL, JagannathanS, YesildirekA. Neural Network Control of Robot Manipulators and Nonlinear Systems. CRC Press/Taylor & Francis Group: Philadelphia, PA, 1999.
[22] GreenM, MooreJB. Persistency of excitation in linear systems. Systems and Control Letters1986; 7: 351-360. · Zbl 0607.93062
[23] JagannathanS. Neural Network Control of Nonlinear Discrete‐time Systems. CRC Press/Taylor & Francis Group: Boca Raton, FL, 2006. · Zbl 1123.93010
[24] HeemelsH, TeelAR, WouwN, NesicD. Networked control systems with communication constraints: tradeoffs between transmission intervals, delays and performance. IEEE Transactions on Automatic Control2010; 55(8):1781-1796. · Zbl 1368.93627
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.