- Notifications
You must be signed in to change notification settings - Fork0
eljandoubi/MADDPG-for-Collaboration-and-Competition
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
This repository contains material from thethird Udacity DRL procjet and the coding exerciceDDPG-pendulum.
In this project, we train a MADDPG multi-agent to solve two types of environment.
FirstTennis :
In this environment, two agents control rackets to bounce a ball over a net. If an agent hits the ball over the net, it receives a reward of
The observation space consists of
Second,Soccer :
In this discrete control environment, four agents compete in a
An environment is considered solved, when an average score of +0.5 over 100 consecutive episodes, and for each agent is obtained.
To set up your python environment to run the code in this repository, follow the instructions below.
Create (and activate) a new environment with Python 3.9.
- Linux orMac:
conda create --name drlndsource activate drlnd
- Windows:
conda create --name drlnd activate drlnd
Follow the instructions inPytorch web page to install pytorch and its dependencies (PIL, numpy,...). For Windows and cuda 11.6
conda install pytorch torchvision torchaudio cudatoolkit=11.6 -c pytorch -c conda-forge
Follow the instructions inthis repository to perform a minimal install of OpenAI gym.
- Install thebox2d environment group by following the instructionshere.
pip install gym[box2d]
Follow the instructions inthird Udacity DRL procjet to get the environment.
Clone the repository, and navigate to the
python/
folder. Then, install several dependencies.
git clone https://github.com/eljandoubi/MADDPG-for-Collaboration-and-Competition.gitcd MADDPG-for-Collaboration-and-Competition/pythonpip install.
- Create anIPython kernel for the
drlnd
environment.
python -m ipykernel install --user --name drlnd --display-name"drlnd"
- Before running code in a notebook, change the kernel to match the
drlnd
environment by using the drop-downKernel
menu.
You can train and/or inference Tennis environment:
Run the training and/or inference cell ofTennis.ipynb
.
The pre-trained models with the highest score are stored inTennis_checkpoint
.
Same for Soccer but the current checkpoint isn't the best.
The implementation and resultats are discussed in the report.
About
An implementation of MADDPG multi-agent to solve a Unity environment like Tennis and Soccer.
Topics
Resources
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Uh oh!
There was an error while loading.Please reload this page.