Strategizing Against Q-Learners: A Control-Theoretical Approach
- 1. Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkiye
Description
In this letter, we explore the susceptibility of the independent Q-learning algorithms (a classical and widely used multi-agent reinforcement learning method) to strategic manipulation of sophisticated opponents in normal-form games played repeatedly. We quantify how much strategically sophisticated agents can exploit naive Q-learners if they know the opponents' Q-learning algorithm. To this end, we formulate the strategic actors' interactions as a stochastic game (whose state encompasses Q-function estimates of the Q-learners) as if the Q-learning algorithms are the underlying dynamical system. We also present a quantization-based approximation scheme to tackle the continuum state space and analyze its performance for two competing strategic actors and a single strategic actor both analytically and numerically.
Files
bib-a8d97835-078a-4cc9-b07b-a03ff6348d9b.txt
Files
(158 Bytes)
| Name | Size | Download all |
|---|---|---|
|
md5:9523a802a582c2f0fcb9112cfa508367
|
158 Bytes | Preview Download |