Skip to main content

A standard API for Multi-Objective Multi-Agent Decision making and a diverse set of reference environments.

Project description

Project Status: Active – The project has reached a stable, usable state and is being actively developed. tests pre-commit Code style: black

MOMAland is an open source Python library for developing and comparing multi-objective multi-agent reinforcement learning algorithms by providing a standard API to communicate between learning algorithms and environments, as well as a standard set of environments compliant with that API. Essentially, the environments follow the standard PettingZoo APIs, but return vectorized rewards as numpy arrays instead of scalar values.

The documentation website is at https://momaland.farama.org/, and we have a public discord server (which we also use to coordinate development work) that you can join here.

Environments

MOMAland includes environments taken from the MOMARL literature, as well as multi-objective version of classical environments, such as SISL or Butterfly. The full list of environments is available at https://momaland.farama.org/environments/all-envs/.

Installation

To install MOMAland, use:

pip install momaland

This does not include dependencies for all components of MOMAland (not everything is required for the basic usage, and some can be problematic to install on certain systems).

  • pip install "momaland[testing]" to install dependencies for API testing.
  • pip install "momaland[learning]" to install dependencies for the supplied learning algorithms.
  • pip install "momaland[all]" for all dependencies for all components.

API

Similar to PettingZoo, the MOMAland API models environments as simple Python env classes. Creating environment instances and interacting with them is very simple - here's an example using the "momultiwalker_stability_v0" environment:

from momaland.envs.momultiwalker_stability import momultiwalker_stability_v0 as _env
import numpy as np

# .env() function will return an AEC environment, as per PZ standard
env = _env.env(render_mode="human")

env.reset(seed=42)
for agent in env.agent_iter():
    # vec_reward is a numpy array
    observation, vec_reward, termination, truncation, info = env.last()

    if termination or truncation:
        action = None
    else:
        action = env.action_space(agent).sample() # this is where you would insert your policy

    env.step(action)
env.close()

# optionally, you can scalarize the reward with weights
# Making the vector reward a scalar reward to shift to single-objective multi-agent (aka PettingZoo)
# We can assign different weights to the objectives of each agent.
weights = {
    "walker_0": np.array([0.7, 0.3]),
    "walker_1": np.array([0.5, 0.5]),
    "walker_2": np.array([0.2, 0.8]),
}
env = LinearizeReward(env, weights)

For details on multi-objective multi-agent RL definitions, see Multi-Objective Multi-Agent Decision Making: A Utility-based Analysis and Survey.

You can also check more examples in this colab notebook! MOMAland Demo in Collab

Learning Algorithms

We provide a set of learning algorithms that are compatible with the MOMAland environments. The learning algorithms are implemented in the learning/ directory. To keep everything as self-contained as possible, each algorithm is implemented as a single-file (close to cleanRL's philosophy).

Nevertheless, we reuse tools provided by other libraries, like multi-objective evaluations and performance indicators from MORL-Baselines.

Here is a list of algorithms that are currently implemented:

Name Single/Multi-policy Reward Utility Observation space Action space Paper
MOMAPPO (OLS) continuous,
discrete
Multi Team Team / Linear Any Any
Scalarized IQL Single Individual Individual / Linear Discrete Discrete
Centralization wrapper Any Team Team / Any Discrete Discrete
Linearization wrapper Single Any Individual / Linear Any Any

Environment Versioning

MOMAland keeps strict versioning for reproducibility reasons. All environments end in a suffix like "_v0". When changes are made to environments that might impact learning results, the number is increased by one to prevent potential confusion.

Development Roadmap

We have a roadmap for future development available here.

Project Maintainers

Project Managers: Florian Felten (@ffelten)

Maintenance for this project is also contributed by the broader Farama team: farama.org/team.

Citing

If you use this repository in your research, please cite:

@inproceedings{TODO}

Development

Setup pre-commit

Clone the repo and run pre-commit install to setup the pre-commit hooks.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

momaland-0.1.0.tar.gz (114.4 kB view details)

Uploaded Source

Built Distribution

momaland-0.1.0-py3-none-any.whl (152.4 kB view details)

Uploaded Python 3

File details

Details for the file momaland-0.1.0.tar.gz.

File metadata

  • Download URL: momaland-0.1.0.tar.gz
  • Upload date:
  • Size: 114.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.3

File hashes

Hashes for momaland-0.1.0.tar.gz
Algorithm Hash digest
SHA256 dcac14dcc66a850c9228938547fe64b8bef876d334e603277b3536618a7e9e75
MD5 09141efa2b4c317140fbd90cc027890e
BLAKE2b-256 abc64bbc737ca56a1b911d31f755e2e031c5bd1061797810cade0ab65b523921

See more details on using hashes here.

File details

Details for the file momaland-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: momaland-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 152.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.3

File hashes

Hashes for momaland-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 e2376d6673c24038961a2f04125a9c25f371acc39005ff3aafbf97d898f4a7bb
MD5 4e56282c0471daf921493088f0fe6163
BLAKE2b-256 3ecb0a4c46b8caf8ee701234adfde99853877a3d19a44d49ffde5a3b82b8c4fc

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page