Show simple item record

dc.contributor.authorAl-Tamimi, Asma Azmien_US
dc.date.accessioned2007-10-08T23:55:04Z
dc.date.available2007-10-08T23:55:04Z
dc.date.issued2007-10-08T23:55:04Z
dc.date.submittedMay 2007en_US
dc.identifier.otherDISS-1694en_US
dc.identifier.urihttp://hdl.handle.net/10106/641
dc.description.abstractIn this work, approximate dynamic programming (ADP) designs based on adaptive critic structures are developed to solve the discrete-time optimal control problems in which the state and action spaces are continuous. This work considers linear discrete-time systems as well as nonlinear discrete-time systems that are affine in the input. This research resulted in forward-in-time reinforcement learning algorithms that converge to the solution of the Generalized Algebraic Riccati Equation (GARE) for linear systems. For the nonlinear case, a forward-in-time reinforcement learning algorithm is presented that converges to the solution of the associated Hamilton-Jacobi Bellman equation (HJB). The results in the linear case can be thought of as a way to solve the GARE of the well-known discrete-time optimal control problem forward in time. Four design algorithms are developed: Heuristic Dynamic programming (HDP), Dual Heuristic dynamic programming (DHP), Action dependent Heuristic Dynamic programming (ADHDP) and Action dependent Dual Heuristic dynamic programming (ADDHP). The significance of these algorithms is that for some of them, particularly the ADHDP algorithm, a priori knowledge of the plant model is not required to solve the dynamic programming problem. Another major outcome of this work is that we introduce a convergent policy iteration scheme based on the HDP algorithm that allows the use of neural networks to arbitrarily approximate for the value function of the discrete-time HJB equation. This online algorithm may be implemented in a way that requires only partial knowledge of the model of the nonlinear dynamical system. The dissertation includes detailed proofs of convergence for the proposed algorithms, HDP, DHP, ADHDP, ADDHP and the nonlinear HDP. Practical numerical examples are provided to show the effectiveness of the developed optimization algorithms. For nonlinear systems, a comparison with methods based on the State-Dependent Riccati Equation (SDRE) is also presented. In all the provided examples, parametric structures like neural networks have been used to find compact representations of the value function and optimal policies for the corresponding optimal control problemsen_US
dc.description.sponsorshipLewis, Franken_US
dc.language.isoENen_US
dc.publisherElectrical Engineeringen_US
dc.titleDiscrete-time Control Algorithms And Adaptive Intelligent Systems Designsen_US
dc.typePh.D.en_US
dc.contributor.committeeChairLewis, Franken_US
dc.degree.departmentElectrical Engineeringen_US
dc.degree.disciplineElectrical Engineeringen_US
dc.degree.grantorUniversity of Texas at Arlingtonen_US
dc.degree.leveldoctoralen_US
dc.degree.namePh.D.en_US
dc.identifier.externalLinkhttps://www.uta.edu/ra/real/editprofile.php?onlyview=1&pid=27
dc.identifier.externalLinkDescriptionLink to Research Profiles


Files in this item

Thumbnail


This item appears in the following Collection(s)

Show simple item record