Skip to content

pjschneiCMU/marl-tutorials

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

41 Commits
 
 
 
 
 
 

Repository files navigation

Multi-Agent Reinforcement Learning (MARL)

Warning

Under development!

This repository aims to provide tutorials on multi-agent reinforcement learning (MARL). Before diving into the modeling of multiple agents interacting through the Markov decision process (MDP) framework (introduced by Shapley in 1953), we first cover the basics of game theory and other fundamental concepts.

Note

Currently, the repository has a strong focus on game theory, which may eventually be moved to a separate GitHub repository.

Game Theory Literature

Key Questions for Modeling a Problem via Game Theory

  • How many players are involved?

    The number of players defines the game’s structure (e.g., two-player vs. multi-player) and influences the complexity of the analysis.

  • Are the actions available finite or infinite?

    Determining whether the action space is finite or infinite affects the strategies players can adopt and the mathematical tools used for analysis.

  • Does the game evolve with simultaneous moves or sequential moves?

    Identifying whether players act simultaneously or in sequence helps classify the game (e.g., normal-form vs. extensive-form) and influences strategy formulation.

  • Is the game static or dynamic?

    A static game is played once, while a dynamic game unfolds over multiple periods. Dynamic games often require more complex strategies, like backward induction.

  • Is there complete or incomplete information?

    Assessing whether players have full knowledge of the game's structure and payoffs is vital. Incomplete information often leads to Bayesian or signaling games.

  • Do players have private information (asymmetric information)?

    If some players have private information, the game involves asymmetric information, leading to scenarios like adverse selection or moral hazard.

  • Are players cooperating or competing?

    Understanding whether players work together (cooperative games) or compete (non-cooperative games) influences the modeling approach and solution concepts.

  • What is the payoff structure?

    Analyzing how payoffs are determined—whether zero-sum, positive-sum, or involving externalities—is crucial for understanding incentives and outcomes.

  • Are players rational, and what level of rationality is assumed?

    Assuming players act in their best interest, the level of rationality (e.g., bounded or perfect) assumed in the model can significantly affect predictions.

  • Is the game repeated, and how does repetition affect strategies?

    In repeated games, strategies may evolve based on past interactions, potentially leading to cooperation in repeated settings (e.g., folk theorem).

  • How are equilibria defined, and which equilibrium concepts are applicable?

    Identifying the appropriate equilibrium concept (e.g., Nash, subgame perfect, Bayesian) is essential for predicting outcomes.

Simultaneous-move Games

Normal-form Games (Strategic-form Games)

  • Von Neumann, J., & Morgenstern, O. (1944). Theory of games and economic behavior. Princeton University Press.

This book established the foundations of normal-form games, providing a structure for analyzing strategic interactions where players make decisions simultaneously

Bayesian Games

Harsanyi’s work introduced Bayesian games, which extend normal-form games to situations of incomplete information, allowing players to form beliefs about unknown factors.

Potential Games

  • Monderer, D., & Shapley, L. S. (1996). Potential games. Games and Economic Behavior, 14(1), 124-143.

This paper defines potential games, a class of games where the incentive structure can be captured by a potential function, simplifying the analysis of equilibria.

Robust Games

  • Aghassi, M., & Bertsimas, D. (2006). Robust game theory. Mathematical Programming, 107(1), 231-273.

This work extends game theory to account for uncertainty and robustness, providing tools for decision-making in uncertain environments.

Sequential-move Games

Stackelberg Games

Stackelberg introduced a model where leaders move first and followers react, which is fundamental in analyzing competitive markets and leadership dynamics in sequential games.

Extensive-form Games

  • Von Neumann, J., & Morgenstern, O. (1944). Theory of games and economic behavior. Princeton University Press.

This book also introduced extensive-form games, providing a framework for analyzing games where the order of moves and the information available at each decision point are crucial.

Games Involving Both Simultaneous and Sequential Moves

Stochastic Games (Markov Games)

  • Shapley, L. S. (1953). Stochastic games. Proceedings of the National Academy of Sciences, 39(10), 1095-1100.

Shapley’s work introduced stochastic games, which generalize Markov decision processes to multi-player settings with both simultaneous and sequential moves.

Littman introduced Minimax-Q learning and extended the concept of stochastic games into the realm of multi-agent reinforcement learning, emphasizing their application in AI and dynamic environments.

Special Forms

Evolutionary Game Theory

This paper introduced the concept of evolutionary stable strategies, a cornerstone of evolutionary game theory.

Repeated Games

This work is key in the theory of repeated games, explaining how cooperation can be sustained over time in strategic settings.

Cooperative Games

Shapley introduced the Shapley value, a solution concept in cooperative game theory that is widely used in economics and political science.

Game Theory - Theory Bascis

Equilibrium Concepts

Nash Equilibrium

  • Nash Jr, J. F. (1950). Equilibrium points in n-person games. Proceedings of the National Academy of Sciences, 36(1), 48-49.

This foundational paper introduced the concept of Nash equilibrium and proved that in every finite game with a finite number of players, there exists at least one Nash equilibrium in mixed strategies.

Correlated Equilibrium

This paper introduced the concept of correlated equilibrium, where players can coordinate their strategies based on shared random signals, generalizing Nash equilibrium.

This paper formalizes correlated equilibrium in the context of Bayesian rationality, showing how players' beliefs and strategies can be aligned through common signals.

Coarse Correlated Equilibrium

This paper explores coarse correlated equilibrium, a generalization of correlated equilibrium where players evaluate their expected payoffs before seeing the signal, highlighting strategic considerations in games.

Fixed Point Theorems

This classic result in topology proves that any continuous function from a compact convex set to itself has at least one fixed point, forming the mathematical foundation for proving the existence of Nash equilibria.

Kakutani's theorem extends Brouwer’s theorem to set-valued (multi-valued) functions and is a crucial tool for proving the existence of Nash equilibria in games with mixed strategies.

Multi-Agent Reinforcement Learning Literature

Surveys and Literature Reviews

Algorithms

MADDPG

Cooperative setting

Competitive setting

Learning strategies

  • Bengio, Y., Louradour, J., Collobert, R., & Weston, J. (2009). Curriculum learning. In Proceedings of the 26th International Conference on Machine Learning (ICML) (pp. 41-48).

Learning zero-sum games

TBD

There are multiple GitHub repositories providing literature resources purely focused on recent advancements in MARL research.

About

Multi-Agent Reinforcement Learning

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published