•  Persian
    • Persian
    • English
  •   ورود
  • دانشگاه فردوسی مشهد
  • |
  • مرکز اطلاع‌رسانی و کتابخانه مرکزی
    • Persian
    • English
  • خانه
  • انواع منابع
    • مقاله مجله
    • کتاب الکترونیکی
    • مقاله همایش
    • استاندارد
    • پروتکل
    • پایان‌نامه
  • راهنمای استفاده
View Item 
  •   کتابخانه دیجیتال دانشگاه فردوسی مشهد
  • Fum
  • Articles
  • ProfDoc
  • View Item
  •   کتابخانه دیجیتال دانشگاه فردوسی مشهد
  • Fum
  • Articles
  • ProfDoc
  • View Item
  • همه
  • عنوان
  • نویسنده
  • سال
  • ناشر
  • موضوع
  • عنوان ناشر
  • ISSN
  • شناسه الکترونیک
  • شابک
جستجوی پیشرفته
JavaScript is disabled for your browser. Some features of this site may not work without it.

Online Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systems

نویسنده:
شعله یاسینی
,
علی کریم پور
,
محمدباقر نقیبی سیستانی
,
حمید رضا مدرّس
,
Sholeh Yasini
,
Ali Karimpour
,
Mohammad Bagher Naghibi Sistani
,
Hamidreza Modares
سال
: 2014
چکیده: Online adaptive optimal control methods based on reinforcement learning algorithms typically need to check for the persistence of excitation condition, which is necessary to be known a priori for convergence of the algorithm. However, this condition is often infeasible to implement or monitor online. This paper proposes an online concurrent reinforcement learning algorithm (CRLA) based on neural networks (NNs) to solve the H1 control problem of partially unknown continuous-time systems, in which the need for persistence of excitation condition is relaxed by using the idea of concurrent learning. First, H1 control problem is formulated as a two-player zero-sum game, and then, online CRLA is employed to obtain the approximation of the optimal value and the Nash equilibrium of the game. The proposed algorithm is implemented on actor–critic–disturbance NN approximator structure to obtain the solution of the Hamilton–Jacobi–Isaacs equation online forward in time. During the implementation of the algorithm, the control input that acts as one player attempts to make the optimal control while the other player, that is, disturbance, tries to make the worstcase possible disturbance. Novel update laws are derived for adaptation of the critic and actor NN weights. The stability of the closed-loop system is guaranteed using Lyapunov technique, and the convergence to the Nash solution of the game is obtained. Simulation results show the effectiveness of the proposed method.
یو آر آی: https://libsearch.um.ac.ir:443/fum/handle/fum/3351422
کلیدواژه(گان): H1 control,two-player zero-sum games,neural networks,online concurrent reinforcement learning algorithm
کالکشن :
  • ProfDoc
  • نمایش متادیتا پنهان کردن متادیتا
  • آمار بازدید

    Online Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systems

Show full item record

contributor authorشعله یاسینیen
contributor authorعلی کریم پورen
contributor authorمحمدباقر نقیبی سیستانیen
contributor authorحمید رضا مدرّسen
contributor authorSholeh Yasinifa
contributor authorAli Karimpourfa
contributor authorMohammad Bagher Naghibi Sistanifa
contributor authorHamidreza Modaresfa
date accessioned2020-06-06T13:21:13Z
date available2020-06-06T13:21:13Z
date issued2014
identifier urihttps://libsearch.um.ac.ir:443/fum/handle/fum/3351422
description abstractOnline adaptive optimal control methods based on reinforcement learning algorithms typically need to check for the persistence of excitation condition, which is necessary to be known a priori for convergence of the algorithm. However, this condition is often infeasible to implement or monitor online. This paper proposes an online concurrent reinforcement learning algorithm (CRLA) based on neural networks (NNs) to solve the H1 control problem of partially unknown continuous-time systems, in which the need for persistence of excitation condition is relaxed by using the idea of concurrent learning. First, H1 control problem is formulated as a two-player zero-sum game, and then, online CRLA is employed to obtain the approximation of the optimal value and the Nash equilibrium of the game. The proposed algorithm is implemented on actor–critic–disturbance NN approximator structure to obtain the solution of the Hamilton–Jacobi–Isaacs equation online forward in time. During the implementation of the algorithm, the control input that acts as one player attempts to make the optimal control while the other player, that is, disturbance, tries to make the worstcase possible disturbance. Novel update laws are derived for adaptation of the critic and actor NN weights. The stability of the closed-loop system is guaranteed using Lyapunov technique, and the convergence to the Nash solution of the game is obtained. Simulation results show the effectiveness of the proposed method.en
languageEnglish
titleOnline Concurrent Reinforcement Learning Algorithm to Solve Two-player Zero-sum Games for Partially-unknown Nonlinear Continuous-time Systemsen
typeJournal Paper
contenttypeExternal Fulltext
subject keywordsH1 controlen
subject keywordstwo-player zero-sum gamesen
subject keywordsneural networksen
subject keywordsonline concurrent reinforcement learning algorithmen
journal titleInternational Journal of Adaptive Control and Signal Processingfa
pages21-Jan
journal volume0
journal issue0
identifier linkhttps://profdoc.um.ac.ir/paper-abstract-1044410.html
identifier articleid1044410
  • درباره ما
نرم افزار کتابخانه دیجیتال "دی اسپیس" فارسی شده توسط یابش برای کتابخانه های ایرانی | تماس با یابش
DSpace software copyright © 2019-2022  DuraSpace