Strategizing against q-learners: a control-theoretical approach
buir.contributor.author | Arslantaş, Yüksel | |
buir.contributor.author | Yüceel, Ege | |
buir.contributor.author | Sayın, Muhammed O. | |
buir.contributor.orcid | Arslantaş, Yüksel|0009-0007-7516-4850 | |
buir.contributor.orcid | Yüceel, Ege|0009-0008-6765-3598 | |
buir.contributor.orcid | Sayın, Muhammed O.|0000-0001-5779-3986 | |
dc.citation.epage | 1738 | |
dc.citation.spage | 1733 | |
dc.citation.volumeNumber | 8 | |
dc.contributor.author | Arslantaş, Yüksel | |
dc.contributor.author | Yüceel, Ege | |
dc.contributor.author | Sayın, Muhammed O. | |
dc.date.accessioned | 2025-02-18T13:04:38Z | |
dc.date.available | 2025-02-18T13:04:38Z | |
dc.date.issued | 2024-06-18 | |
dc.department | Department of Electrical and Electronics Engineering | |
dc.description.abstract | In this letter, we explore the susceptibility of the independent Q-learning algorithms (a classical and widely used multi-agent reinforcement learning method) to strategic manipulation of sophisticated opponents in normal-form games played repeatedly. We quantify how much strategically sophisticated agents can exploit naive Q-learners if they know the opponents' Q-learning algorithm. To this end, we formulate the strategic actors' interactions as a stochastic game (whose state encompasses Q-function estimates of the Q-learners) as if the Q-learning algorithms are the underlying dynamical system. We also present a quantization-based approximation scheme to tackle the continuum state space and analyze its performance for two competing strategic actors and a single strategic actor both analytically and numerically. | |
dc.identifier.doi | 10.1109/LCSYS.2024.3416240 | |
dc.identifier.eissn | 2475-1456 | |
dc.identifier.uri | https://hdl.handle.net/11693/116383 | |
dc.language.iso | English | |
dc.publisher | Institute of Electrical and Electronics Engineers | |
dc.relation.isversionof | https://dx.doi.org/10.1109/LCSYS.2024.3416240 | |
dc.rights | CC BY 4.0 DEED (Attribution 4.0 International) | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/deed.tr | |
dc.source.title | IEEE Control Systems Letters | |
dc.subject | Reinforcement learning | |
dc.subject | Game theory | |
dc.subject | Markov processes | |
dc.title | Strategizing against q-learners: a control-theoretical approach | |
dc.type | Article |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Strategizing_against_q-learners_a_control-theoretical_approach.pdf
- Size:
- 2.34 MB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description: