Oven logo

Oven

rlcard1.2.0

Published

A Toolkit for Reinforcement Learning in Card Games

pip install rlcard

Package Downloads

Weekly DownloadsMonthly Downloads

Project URLs

Requires Python

Dependencies

    RLCard: A Toolkit for Reinforcement Learning in Card Games

    Logo

    Testing PyPI version Coverage Status Downloads Downloads License: MIT

    中文文档

    RLCard is a toolkit for Reinforcement Learning (RL) in card games. It supports multiple card environments with easy-to-use interfaces for implementing various reinforcement learning and searching algorithms. The goal of RLCard is to bridge reinforcement learning and imperfect information games. RLCard is developed by DATA Lab at Rice and Texas A&M University, and community contributors.

    Community:

    • Slack: Discuss in our #rlcard-project slack channel.
    • QQ Group: Join our QQ group to discuss. Password: rlcardqqgroup
      • Group 1: 665647450
      • Group 2: 117349516

    News:

    • We have updated the tutorials in Jupyter Notebook to help you walk through RLCard! Please check RLCard Tutorial.
    • All the algorithms can suppport PettingZoo now. Please check here. Thanks the contribtuion from Yifei Cheng.
    • Please follow DouZero, a strong Dou Dizhu AI and the ICML 2021 paper. An online demo is available here. The algorithm is also integrated in RLCard. See Training DMC on Dou Dizhu.
    • Our package is used in PettingZoo. Please check it out!
    • We have released RLCard-Showdown, GUI demo for RLCard. Please check out here!
    • Jupyter Notebook tutorial available! We add some examples in R to call Python interfaces of RLCard with reticulate. See here
    • Thanks for the contribution of @Clarit7 for supporting different number of players in Blackjack. We call for contributions for gradually making the games more configurable. See here for more details.
    • Thanks for the contribution of @Clarit7 for the Blackjack and Limit Hold'em human interface.
    • Now RLCard supports environment local seeding and multiprocessing. Thanks for the testing scripts provided by @weepingwillowben.
    • Human interface of NoLimit Holdem available. The action space of NoLimit Holdem has been abstracted. Thanks for the contribution of @AdrianP-.
    • New game Gin Rummy and human GUI available. Thanks for the contribution of @billh0420.
    • PyTorch implementation available. Thanks for the contribution of @mjudell.

    Contributors

    The following games are mainly developed and maintained by community contributors. Thank you!

    Thank all the contributors!

    daochenzha   hsywhu   CaoYuanpu   billh0420   ruzhwei   adrianpgob   Zhigal   aypee19   Clarit7   lhenry15   ismael-elatifi   mjudell   jkterry1   kaanozdogru   junyuGuo  
    Xixo99   rodrigodelazcano   Michael1015198808   mia1996   kaiks   claude9493   SonSang   rishabhvarshney14   aetheryang   rxng8   nondecidibile   benblack769   zhengsx   andrewnc  

    Cite this work

    If you find this repo useful, you may cite:

    Zha, Daochen, et al. "RLCard: A Platform for Reinforcement Learning in Card Games." IJCAI. 2020.

    @inproceedings{zha2020rlcard,
      title={RLCard: A Platform for Reinforcement Learning in Card Games},
      author={Zha, Daochen and Lai, Kwei-Herng and Huang, Songyi and Cao, Yuanpu and Reddy, Keerthana and Vargas, Juan and Nguyen, Alex and Wei, Ruzhe and Guo, Junyu and Hu, Xia},
      booktitle={IJCAI},
      year={2020}
    }
    

    Installation

    Make sure that you have Python 3.6+ and pip installed. We recommend installing the stable version of rlcard with pip:

    pip3 install rlcard
    

    The default installation will only include the card environments. To use PyTorch implementation of the training algorithms, run

    pip3 install rlcard[torch]
    

    If you are in China and the above command is too slow, you can use the mirror provided by Tsinghua University:

    pip3 install rlcard -i https://pypi.tuna.tsinghua.edu.cn/simple
    

    Alternatively, you can clone the latest version with (if you are in China and Github is slow, you can use the mirror in Gitee):

    git clone https://github.com/datamllab/rlcard.git
    

    or only clone one branch to make it faster:

    git clone -b master --single-branch --depth=1 https://github.com/datamllab/rlcard.git
    

    Then install with

    cd rlcard
    pip3 install -e .
    pip3 install -e .[torch]
    

    We also provide conda installation method:

    conda install -c toubun rlcard
    

    Conda installation only provides the card environments, you need to manually install Pytorch on your demands.

    Examples

    A short example is as below.

    import rlcard
    from rlcard.agents import RandomAgent
    
    env = rlcard.make('blackjack')
    env.set_agents([RandomAgent(num_actions=env.num_actions)])
    
    print(env.num_actions) # 2
    print(env.num_players) # 1
    print(env.state_shape) # [[2]]
    print(env.action_shape) # [None]
    
    trajectories, payoffs = env.run()
    

    RLCard can be flexibly connected to various algorithms. See the following examples:

    Demo

    Run examples/human/leduc_holdem_human.py to play with the pre-trained Leduc Hold'em model. Leduc Hold'em is a simplified version of Texas Hold'em. Rules can be found here.

    >> Leduc Hold'em pre-trained model
    
    >> Start a new game!
    >> Agent 1 chooses raise
    
    =============== Community Card ===============
    ┌─────────┐
    │░░░░░░░░░│
    │░░░░░░░░░│
    │░░░░░░░░░│
    │░░░░░░░░░│
    │░░░░░░░░░│
    │░░░░░░░░░│
    │░░░░░░░░░│
    └─────────┘
    ===============   Your Hand    ===============
    ┌─────────┐
    │J        │
    │         │
    │         │
    │    ♥    │
    │         │
    │         │
    │        J│
    └─────────┘
    ===============     Chips      ===============
    Yours:   +
    Agent 1: +++
    =========== Actions You Can Choose ===========
    0: call, 1: raise, 2: fold
    
    >> You choose action (integer):
    

    We also provide a GUI for easy debugging. Please check here. Some demos:

    doudizhu-replay leduc-replay

    Available Environments

    We provide a complexity estimation for the games on several aspects. InfoSet Number: the number of information sets; InfoSet Size: the average number of states in a single information set; Action Size: the size of the action space. Name: the name that should be passed to rlcard.make to create the game environment. We also provide the link to the documentation and the random example.

    GameInfoSet NumberInfoSet SizeAction SizeNameUsage
    Blackjack (wiki, baike)10^310^110^0blackjackdoc, example
    Leduc Hold’em (paper)10^210^210^0leduc-holdemdoc, example
    Limit Texas Hold'em (wiki, baike)10^1410^310^0limit-holdemdoc, example
    Dou Dizhu (wiki, baike)10^53 ~ 10^8310^2310^4doudizhudoc, example
    Mahjong (wiki, baike)10^12110^4810^2mahjongdoc, example
    No-limit Texas Hold'em (wiki, baike)10^16210^310^4no-limit-holdemdoc, example
    UNO (wiki, baike)10^16310^1010^1unodoc, example
    Gin Rummy (wiki, baike)10^52--gin-rummydoc, example
    Bridge (wiki, baike)--bridgedoc, example

    Supported Algorithms

    Algorithmexamplereference
    Deep Monte-Carlo (DMC)examples/run_dmc.py[paper]
    Deep Q-Learning (DQN)examples/run_rl.py[paper]
    Neural Fictitious Self-Play (NFSP)examples/run_rl.py[paper]
    Counterfactual Regret Minimization (CFR)examples/run_cfr.py[paper]

    Pre-trained and Rule-based Models

    We provide a model zoo to serve as the baselines.

    ModelExplanation
    leduc-holdem-cfrPre-trained CFR (chance sampling) model on Leduc Hold'em
    leduc-holdem-rule-v1Rule-based model for Leduc Hold'em, v1
    leduc-holdem-rule-v2Rule-based model for Leduc Hold'em, v2
    uno-rule-v1Rule-based model for UNO, v1
    limit-holdem-rule-v1Rule-based model for Limit Texas Hold'em, v1
    doudizhu-rule-v1Rule-based model for Dou Dizhu, v1
    gin-rummy-novice-ruleGin Rummy novice rule model

    API Cheat Sheet

    How to create an environment

    You can use the the following interface to make an environment. You may optionally specify some configurations with a dictionary.

    • env = rlcard.make(env_id, config={}): Make an environment. env_id is a string of a environment; config is a dictionary that specifies some environment configurations, which are as follows.
      • seed: Default None. Set a environment local random seed for reproducing the results.
      • allow_step_back: Default False. True if allowing step_back function to traverse backward in the tree.
      • Game specific configurations: These fields start with game_. Currently, we only support game_num_players in Blackjack, .

    Once the environemnt is made, we can access some information of the game.

    • env.num_actions: The number of actions.
    • env.num_players: The number of players.
    • env.state_shape: The shape of the state space of the observations.
    • env.action_shape: The shape of the action features (Dou Dizhu's action can encoded as features)

    What is state in RLCard

    State is a Python dictionary. It consists of observation state['obs'], legal actions state['legal_actions'], raw observation state['raw_obs'] and raw legal actions state['raw_legal_actions'].

    Basic interfaces

    The following interfaces provide a basic usage. It is easy to use but it has assumtions on the agent. The agent must follow agent template.

    • env.set_agents(agents): agents is a list of Agent object. The length of the list should be equal to the number of the players in the game.
    • env.run(is_training=False): Run a complete game and return trajectories and payoffs. The function can be used after the set_agents is called. If is_training is True, it will use step function in the agent to play the game. If is_training is False, eval_step will be called instead.

    Advanced interfaces

    For advanced usage, the following interfaces allow flexible operations on the game tree. These interfaces do not make any assumtions on the agent.

    • env.reset(): Initialize a game. Return the state and the first player ID.
    • env.step(action, raw_action=False): Take one step in the environment. action can be raw action or integer; raw_action should be True if the action is raw action (string).
    • env.step_back(): Available only when allow_step_back is True. Take one step backward. This can be used for algorithms that operate on the game tree, such as CFR (chance sampling).
    • env.is_over(): Return True if the current game is over. Otherewise, return False.
    • env.get_player_id(): Return the Player ID of the current player.
    • env.get_state(player_id): Return the state that corresponds to player_id.
    • env.get_payoffs(): In the end of the game, return a list of payoffs for all the players.
    • env.get_perfect_information(): (Currently only support some of the games) Obtain the perfect information at the current state.

    Library Structure

    The purposes of the main modules are listed as below:

    More Documents

    For more documentation, please refer to the Documents for general introductions. API documents are available at our website.

    Contributing

    Contribution to this project is greatly appreciated! Please create an issue for feedbacks/bugs. If you want to contribute codes, please refer to Contributing Guide. If you have any questions, please contact Daochen Zha with [email protected].

    Acknowledgements

    We would like to thank JJ World Network Technology Co.,LTD for the generous support and all the contributions from the community contributors.