•  English
    • Persian
    • English
  •   Login
  • Ferdowsi University of Mashhad
  • |
  • Information Center and Central Library
    • Persian
    • English
  • Home
  • Source Types
    • Journal Paper
    • Ebook
    • Conference Paper
    • Standard
    • Protocol
    • Thesis
  • Use Help
View Item 
  •   FUM Digital Library
  • Fum
  • Articles
  • ProfDoc
  • View Item
  •   FUM Digital Library
  • Fum
  • Articles
  • ProfDoc
  • View Item
  • All Fields
  • Title
  • Author
  • Year
  • Publisher
  • Subject
  • Publication Title
  • ISSN
  • DOI
  • ISBN
Advanced Search
JavaScript is disabled for your browser. Some features of this site may not work without it.

Online Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systems

Author:
شعله یاسینی
,
علی کریم پور
,
محمدباقر نقیبی سیستانی
,
حمید رضا مدرّس
,
Sholeh Yasini
,
Ali Karimpour
,
Mohammad Bagher Naghibi Sistani
,
Hamidreza Modares
Year
: 2014
Abstract: Online adaptive optimal control methods based on reinforcement learning algorithms typically need to check for the persistence of excitation condition, which is necessary to be known a priori for convergence of the algorithm. However, this condition is often infeasible to implement or monitor online. This paper proposes an online concurrent reinforcement learning algorithm (CRLA) based on neural networks (NNs) to solve the H1 control problem of partially unknown continuous-time systems, in which the need for persistence of excitation condition is relaxed by using the idea of concurrent learning. First, H1 control problem is formulated as a two-player zero-sum game, and then, online CRLA is employed to obtain the approximation of the optimal value and the Nash equilibrium of the game. The proposed algorithm is implemented on actor–critic–disturbance NN approximator structure to obtain the solution of the Hamilton–Jacobi–Isaacs equation online forward in time. During the implementation of the algorithm, the control input that acts as one player attempts to make the optimal control while the other player, that is, disturbance, tries to make the worstcase possible disturbance. Novel update laws are derived for adaptation of the critic and actor NN weights. The stability of the closed-loop system is guaranteed using Lyapunov technique, and the convergence to the Nash solution of the game is obtained. Simulation results show the effectiveness of the proposed method.
URI: https://libsearch.um.ac.ir:443/fum/handle/fum/3351422
Keyword(s): H1 control,two-player zero-sum games,neural networks,online concurrent reinforcement learning algorithm
Collections :
  • ProfDoc
  • Show Full MetaData Hide Full MetaData
  • Statistics

    Online Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systems

Show full item record

contributor authorشعله یاسینیen
contributor authorعلی کریم پورen
contributor authorمحمدباقر نقیبی سیستانیen
contributor authorحمید رضا مدرّسen
contributor authorSholeh Yasinifa
contributor authorAli Karimpourfa
contributor authorMohammad Bagher Naghibi Sistanifa
contributor authorHamidreza Modaresfa
date accessioned2020-06-06T13:21:13Z
date available2020-06-06T13:21:13Z
date issued2014
identifier urihttps://libsearch.um.ac.ir:443/fum/handle/fum/3351422?locale-attribute=en
description abstractOnline adaptive optimal control methods based on reinforcement learning algorithms typically need to check for the persistence of excitation condition, which is necessary to be known a priori for convergence of the algorithm. However, this condition is often infeasible to implement or monitor online. This paper proposes an online concurrent reinforcement learning algorithm (CRLA) based on neural networks (NNs) to solve the H1 control problem of partially unknown continuous-time systems, in which the need for persistence of excitation condition is relaxed by using the idea of concurrent learning. First, H1 control problem is formulated as a two-player zero-sum game, and then, online CRLA is employed to obtain the approximation of the optimal value and the Nash equilibrium of the game. The proposed algorithm is implemented on actor–critic–disturbance NN approximator structure to obtain the solution of the Hamilton–Jacobi–Isaacs equation online forward in time. During the implementation of the algorithm, the control input that acts as one player attempts to make the optimal control while the other player, that is, disturbance, tries to make the worstcase possible disturbance. Novel update laws are derived for adaptation of the critic and actor NN weights. The stability of the closed-loop system is guaranteed using Lyapunov technique, and the convergence to the Nash solution of the game is obtained. Simulation results show the effectiveness of the proposed method.en
languageEnglish
titleOnline Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systemsen
typeJournal Paper
contenttypeExternal Fulltext
subject keywordsH1 controlen
subject keywordstwo-player zero-sum gamesen
subject keywordsneural networksen
subject keywordsonline concurrent reinforcement learning algorithmen
journal titleInternational Journal of Adaptive Control and Signal Processingfa
pages21-Jan
journal volume0
journal issue0
identifier linkhttps://profdoc.um.ac.ir/paper-abstract-1044410.html
identifier articleid1044410
  • About Us
نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
DSpace software copyright © 2019-2022  DuraSpace