Multi-Agent Arcade Learning Environment Python Interface
Project description
The Arcade Learning Environment
Overview
The Arcade Learning Environment (ALE) is a simple object-oriented framework that allows researchers and hobbyists to develop AI agents for Atari 2600 games. It is built on top of the Atari 2600 emulator Stella and separates the details of emulation from agent design. This video depicts over 50 games currently supported in the ALE.
For an overview of our goals for the ALE read The Arcade Learning Environment: An Evaluation Platform for General Agents. If you use ALE in your research, we ask that you please cite this paper in reference to the environment (BibTeX entry at the end of this document). Also, if you have any questions or comments about the ALE, please contact us through our mailing list.
Feedback and suggestions are welcome and may be addressed to any active member of the ALE team.
Features
- Object-oriented framework with support to add agents and games.
- Emulation core uncoupled from rendering and sound generation modules for fast emulation with minimal library dependencies.
- Automatic extraction of game score and end-of-game signal for more than 50 Atari 2600 games.
- Multi-platform code (compiled and tested under OS X and several Linux distributions, with Cygwin support).
- Communication between agents and emulation core can be accomplished through pipes, allowing for cross-language development (sample Java code included).
- Python development is supported through ctypes.
- Agents programmed in C++ have access to all features in the ALE.
- Visualization tools.
Quick start
Install main dependences:
sudo apt-get install libsdl1.2-dev libsdl-gfx1.2-dev libsdl-image1.2-dev cmake
Compilation:
$ mkdir build && cd build
$ cmake -DUSE_SDL=ON -DUSE_RLGLUE=OFF -DBUILD_EXAMPLES=ON ..
$ make -j 4
To install the Python module:
$ pip install multi-agent-ale-py
Getting the ALE to work on Visual Studio requires a bit of extra wrangling. You may wish to use IslandMan93's Visual Studio port of the ALE.
For more details and installation instructions, see the manual. To ask questions and discuss, please join the ALE-users group.
ALE releases
Releases before v.0.5 are available for download in our previous website. For the latest releases, please check our releases page.
List of command-line parameters
Execute ./ale -help
for more details; alternatively, see documentation
available at http://www.arcadelearningenvironment.org.
-random_seed [n] -- sets the random seed; defaults to the current time
-game_controller [fifo|fifo_named] -- specifies how agents interact
with the ALE; see Java agent documentation for details
-config [file] -- specifies a configuration file, from which additional
parameters are read
-run_length_encoding [false|true] -- determine whether run-length encoding is
used to send data over pipes; irrelevant when an internal agent is
being used
-max_num_frames_per_episode [n] -- sets the maximum number of frames per
episode. Once this number is reached, a new episode will start. Currently
implemented for all agents when using pipes (fifo/fifo_named)
-max_num_frames [n] -- sets the maximum number of frames (independent of how
many episodes are played)
Citing The Arcade Learning Environment
If you use the ALE in your research, we ask that you please cite the following.
M. G. Bellemare, Y. Naddaf, J. Veness and M. Bowling. The Arcade Learning Environment: An Evaluation Platform for General Agents, Journal of Artificial Intelligence Research, Volume 47, pages 253-279, 2013.
In BibTeX format:
@Article{bellemare13arcade,
author = {{Bellemare}, M.~G. and {Naddaf}, Y. and {Veness}, J. and {Bowling}, M.},
title = {The Arcade Learning Environment: An Evaluation Platform for General Agents},
journal = {Journal of Artificial Intelligence Research},
year = "2013",
month = "jun",
volume = "47",
pages = "253--279",
}
If you use the ALE with sticky actions (flag repeat_action_probability
), or if
you use the different game flavours (mode and difficulty switches), we ask you
that you also cite the following:
M. C. Machado, M. G. Bellemare, E. Talvitie, J. Veness, M. J. Hausknecht, M. Bowling. Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents, Journal of Artificial Intelligence Research, Volume 61, pages 523-562, 2018.
In BibTex format:
@Article{machado18arcade,
author = {Marlos C. Machado and Marc G. Bellemare and Erik Talvitie and Joel Veness and Matthew J. Hausknecht and Michael Bowling},
title = {Revisiting the Arcade Learning Environment: Evaluation Protocols and Open Problems for General Agents},
journal = {Journal of Artificial Intelligence Research},
volume = {61},
pages = {523--562},
year = {2018}
}
Contributing, code style
If you would like to make changes to the codebase, please adhere to the following code style conventions.
ALE contains two sets of source files: Files .hxx and .cxx are part of the Stella emulator code. Files .hpp and .cpp are original ALE code. The Stella files are not subject to our conventions, please retain their local style.
The ALE code style conventions are roughly summarised as "clang-format with the following settings: ReflowComments: false, PointerAlignment: Left, KeepEmptyLinesAtTheStartOfBlocks: false, IndentCaseLabels: true, AccessModifierOffset: -1". That is:
- Indent by two spaces; Egyptian braces, no extraneous newlines at the margins of blocks and between top-level declarations.
- Pointer/ref qualifiers go on the left (e.g.
void* p
). - Class member access modifiers are indented by one space.
- Inline comments should be separated from code by two spaces (though this is not currently applied consistently).
- There is no strict line length limit, but keep it reasonable.
- Namespace close braces and
#endif
s should have comments.
The overall format should look reasonably "compact" without being crowded. Use blank lines generously within blocks and long comments to create visual cues for the segmentation of ideas.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for multi-agent-ale-py-0.1.3.dev0.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | f3f832d9ee2a95a6e984a2886b4dabdd857b9009277a4c2bd0a0954fef1acbd7 |
|
MD5 | cdff2e7d13c76b6d7f73bce0d1403ec4 |
|
BLAKE2b-256 | 7a63031d1cc3f9475f44d47eb4229eac2998398477969cec097330d98c0b3785 |
Hashes for multi_agent_ale_py-0.1.3.dev0-cp37-cp37m-macosx_10_15_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 36c648122050c8306178550ce70564fd21d8f97eb9cd33a8205adbb0521f8672 |
|
MD5 | 05fa05cdf3e33e18797aec8787bf5c7c |
|
BLAKE2b-256 | 473ea32d9e9eac93365565fda6fb093f6ef2333a39f8f8735b60dedd6e39f414 |