Skip to content
/ UniLight Public

Multi-Agent Reinforcement Learning for Traffic Signal Control through Universal Communication Method

License

Notifications You must be signed in to change notification settings

zyr17/UniLight

Repository files navigation

UniLight

Source code of "Multi-Agent Reinforcement Learning for Traffic Signal Control through Universal Communication Method", which proposes a universal communication method UniComm between agents.

@inproceedings{DBLP:conf/ijcai/JiangQS0Z22,
  author       = {Qize Jiang and
                  Minhao Qin and
                  Shengmin Shi and
                  Weiwei Sun and
                  Baihua Zheng},
  editor       = {Luc De Raedt},
  title        = {Multi-Agent Reinforcement Learning for Traffic Signal Control through
                  Universal Communication Method},
  booktitle    = {Proceedings of the Thirty-First International Joint Conference on
                  Artificial Intelligence, {IJCAI} 2022, Vienna, Austria, 23-29 July
                  2022},
  pages        = {3854--3860},
  publisher    = {ijcai.org},
  year         = {2022},
  url          = {https://doi.org/10.24963/ijcai.2022/535},
  doi          = {10.24963/ijcai.2022/535},
  timestamp    = {Wed, 27 Jul 2022 16:43:00 +0200},
  biburl       = {https://dblp.org/rec/conf/ijcai/JiangQS0Z22.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}

Usage

We use CityFlow as traffic simulator. Note it is a forked version of the original CityFlow, which added some functions to calculate some statistics in C++ instead of Python.

We suggest using Docker to run our code. Use the docker image: zyr17/unlight. Alternatively, you can build an environment by your self. We run our codes on Python 3.6.5, it should also work on higher version of Python, but the compatibility is not guaranteed.

Docker

git clone [email protected]:zyr17/UniLight.git
cd UniLight && tar zxf data.tgz
docker run -it --rm -v `pwd`:/code zyr17/unilight

# in docker
cd /code

Build

git clone [email protected]:zyr17/CityFlow.git
cd CityFlow && pip install . --upgrade && cd ..
git clone [email protected]:zyr17/UniLight.git
cd UniLight && tar zxf data.tgz && pip install -r requirements.txt

Run

For training:

python main.py --config configs/main/UniLight.yml --cityflow-config configs/cityflow/SH1.yml

For testing:

python main.py --config configs/main/UniLight.yml --cityflow-config configs/cityflow/SH1.yml --preload-model-file ${PATH_TO_MODEL_PT} --test-round 10

Details

Environment

The definition of environment is in envs/CityFlowEnv.py. It runs in another process managed by envs/SubProcVecEnv.py.

Dataset

You can find the newly proposed datasets in data/SH1 and data/SH2, as well as three commonly used JN HZ and NY after extracting the data.tgz tarball.

Agents

The agents and communication mainly defined in rl/agents/NewModelAgent.py, which is wrapped by rl/multiAgents/NewModelMultiAgent.py, and its nn.Module is defined in rl/models/innerModels/NewModel.py.

Arguments and Configs

We parse arguments by utils/argument.py. The config for CityFlow and main process are two separate configs. All CityFlow configs are stored in configs/cityflow, and configs/main for main configs.

Logs

When training starts, all logs will save in logs/${TARGET} folder, including full logs, configs, CityFlow replay files, and saved models. The target foler name is start with current time. There are two exceptions, output of TensorBoard is in runs, and output of wandb is in wandb.

About

Multi-Agent Reinforcement Learning for Traffic Signal Control through Universal Communication Method

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published