Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

This repository contains an implementation of Deep Deterministic Policy Gradient (DDPG), a reinforcement learning algorithm designed for environments with continuous action spaces. It features actor-critic architecture, experience replay, and exploration strategies, and is tested on environments like MountainCarContinuous. More info on Medium blog!

NotificationsYou must be signed in to change notification settings

sirine-b/DDPG

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 

Repository files navigation

Project Description

This repository contains an implementation ofDeep Deterministic Policy Gradient (DDPG), a Reinforcement Learning algorithm designed for environments with continuous action spaces. It featuresActor-Critic architecture,Experience Replay,Target Networks, andExploration Strategies, and is tested on the likeMountainCarContinuous-v0 environment.Batch Normalisation is also included to stabilise training. For a deeper explanation of DDPG, the theory behind it, how it works and the structure of the code, check out myMedium blog.

Features

  • Actor-Critic Architecture: DDPG uses two networks—one for policy (actor) and one for value estimation (critic).
  • Replay Buffer: Stores past experiences to ensure more stable training.
  • Target Networks: Helps to stabilise learning by slowly updating the target networks.
  • Exploration Noise: ImplementsOrnstein-Uhlenbeck noise to facilitate smooth exploration in continuous action spaces.
  • Batch Normalisation: Used to stabilise training by normalising activations and reducing internal covariate shifts.
  • Testing & Visualisation: Includes functionality to test the agent’s performance and visualise its actions in the MountainCar environment.

Steps to Run

To be able to run this DDPG implementation and experiment with it, please follow the steps described below by copying and pasting the relevant lines onto your command prompt.

1. Clone the repository

git clone https://github.com/sirine-b/DDPG.gitcd DDPG

2. Install the required libraries

pip install -r requirements.txt

3. Open the Jupyter notebook

jupyter notebook DDPG_model.ipynb

4. You're good to go!

You can now run the code yourself and experiment with the training and testing of the DDPG agent by trying out different hyperparameters, environments ... etc

About

This repository contains an implementation of Deep Deterministic Policy Gradient (DDPG), a reinforcement learning algorithm designed for environments with continuous action spaces. It features actor-critic architecture, experience replay, and exploration strategies, and is tested on environments like MountainCarContinuous. More info on Medium blog!

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

[8]ページ先頭

©2009-2025 Movatter.jp