Abstract

In this paper, an event-triggered optimal adaptive regulation of an uncertain linear discrete time system is proposed. This scheme solves the optimal control in a forward in- time and online manner by using both dynamic programming and Q learning. First, the time varying action dependent value or the Q-function is estimated online by an adaptive value function estimator (VFE) with event-based state vector and a time dependent basis function. The estimated value function parameters are subsequently used to generate the optimal control gain matrix. Further, aperiodic tuning law for the VFE parameters is proposed not only to estimate the parameters but also handle the terminal constraint. The parameters are tuned only at the event-trigger instants thus reducing computation when compared to the traditional optimal adaptive control. Above all, an adaptive event-trigger condition to decide the event-trigger instants and guarantee stability of the closed-loop system is analytically derived based on the optimal performance criterion via Lyapunov direct method. Nonetheless, the existence of a non-trivial minimum inter-event time is analyzed. Further, it is shown that the parameters converge asymptotically provided the persistency of excitation condition on the regression vector is ensured. Finally, the analytical design is validated with the simulation results.

Department(s)

Electrical and Computer Engineering

Second Department

Computer Science

International Standard Book Number (ISBN)

978-147997746-8

International Standard Serial Number (ISSN)

2576-2370; 0743-1546

Document Type

Article - Conference proceedings

Document Version

Citation

File Type

text

Language(s)

English

Rights

© 2024 Institute of Electrical and Electronics Engineers, All rights reserved.

Publication Date

01 Jan 2014

Share

 
COinS