Q-based Equilibria

04/25/2023
by   Olivier Compte, et al.
0

In dynamic environments, Q-learning is an adaptative rule that provides an estimate (a Q-value) of the continuation value associated with each alternative. A naive policy consists in always choosing the alternative with highest Q-value. We consider a family of Q-based policy rules that may systematically favor some alternatives over others, for example rules that incorporate a leniency bias that favors cooperation. In the spirit of Compte and Postlewaite [2018], we look for equilibrium biases (or Qb-equilibria) within this family of Q-based rules. We examine classic games under various monitoring technologies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/01/2018

Local Equilibria

A local equilibrium is a generalization of a Walrasian, i.e., competitiv...
research
07/11/2019

On Relevant Equilibria in Reachability Games

We study multiplayer reachability games played on a finite directed grap...
research
07/13/2022

Approximate Nash Equilibrium Learning for n-Player Markov Games in Dynamic Pricing

We investigate Nash equilibrium learning in a competitive Markov Game (M...
research
10/02/2018

Relating Metric Distortion and Fairness of Social Choice Rules

One way of evaluating social choice (voting) rules is through a utilitar...
research
10/20/2018

Proportionality Degree of Multiwinner Rules

We study multiwinner elections with approval-based preferences. An insta...
research
05/28/2018

Combinatorial Auctions with Endowment Effect

We study combinatorial auctions with bidders that exhibit endowment effe...

Please sign up or login with your details

Forgot password? Click here to reset