Please use this identifier to cite or link to this item:
https://dspace.iiti.ac.in/handle/123456789/16976
Full metadata record
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Shreyas, Sumithra Rudresha R. | en_US |
| dc.contributor.author | Vijesh, Antony V. | en_US |
| dc.date.accessioned | 2025-10-23T12:41:59Z | - |
| dc.date.available | 2025-10-23T12:41:59Z | - |
| dc.date.issued | 2025 | - |
| dc.identifier.citation | Shreyas, S. R. R., & Vijesh, A. v. (2025). A multi-step minimax Q-learning algorithm for two-player zero-sum Markov games. Neurocomputing, 657. https://doi.org/10.1016/j.neucom.2025.131552 | en_US |
| dc.identifier.issn | 18728286 | - |
| dc.identifier.issn | 09252312 | - |
| dc.identifier.other | EID(2-s2.0-105016786440) | - |
| dc.identifier.uri | https://dx.doi.org/10.1016/j.neucom.2025.131552 | - |
| dc.identifier.uri | https://dspace.iiti.ac.in:8080/jspui/handle/123456789/16976 | - |
| dc.description.abstract | An interesting iterative procedure is proposed to solve two-player zero-sum Markov games. Under suitable assumptions, the boundedness of the proposed iterates is obtained theoretically. Using results from stochastic approximation, the almost sure convergence of the proposed multi-step minimax Q-learning is obtained theoretically. More specifically, the proposed algorithm converges to the game theoretic optimal value with probability one, when the model information is not known. Numerical simulations authenticate that the proposed algorithm is effective and easy to implement. © 2025 Elsevier B.V., All rights reserved. | en_US |
| dc.language.iso | en | en_US |
| dc.publisher | Elsevier B.V. | en_US |
| dc.source | Neurocomputing | en_US |
| dc.subject | Minimax Q-learning | en_US |
| dc.subject | Multi-agent Reinforcement Learning | en_US |
| dc.subject | Two-player Zero-sum Markov Games | en_US |
| dc.subject | Approximation Theory | en_US |
| dc.subject | Game Theory | en_US |
| dc.subject | Iterative Methods | en_US |
| dc.subject | Learning Algorithms | en_US |
| dc.subject | Multi Agent Systems | en_US |
| dc.subject | Almost Sure Convergence | en_US |
| dc.subject | Boundedness | en_US |
| dc.subject | Markov Games | en_US |
| dc.subject | Minimax-q Learning | en_US |
| dc.subject | Multi-agent Reinforcement Learning | en_US |
| dc.subject | Multisteps | en_US |
| dc.subject | Q-learning Algorithms | en_US |
| dc.subject | Stochastic Approximations | en_US |
| dc.subject | Two-player Zero-sum Markov Game | en_US |
| dc.subject | Zero Sums | en_US |
| dc.subject | Stochastic Systems | en_US |
| dc.subject | Algorithm | en_US |
| dc.subject | Article | en_US |
| dc.subject | Controlled Study | en_US |
| dc.subject | Game | en_US |
| dc.subject | Human | en_US |
| dc.subject | Human Experiment | en_US |
| dc.subject | Probability | en_US |
| dc.subject | Q Learning | en_US |
| dc.subject | Simulation | en_US |
| dc.title | A multi-step minimax Q-learning algorithm for two-player zero-sum Markov games | en_US |
| dc.type | Journal Article | en_US |
| Appears in Collections: | Department of Mathematics | |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.
Altmetric Badge: