Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

SCPO: Safe Reinforcement Learning with Safety Critic Policy Optimization

License

NotificationsYou must be signed in to change notification settings

SafeRL-Lab/SCPO

Repository files navigation

SCPO is a safe reinforcement learning algorithm. This repo is a fork ofStable Baselines3.

Installation

Note: Stable-Baselines3 supports PyTorch >= 1.11

Prerequisites

SCPO requires Python 3.7+.

Install using pip

Install the Stable Baselines3 package:

pip install -r requirements.txt

We use environments from Bullet-Safety-Gym. Please follow the installation steps fromhttps://github.com/SvenGronauer/Bullet-Safety-Gym.

If you want to run pytorch in gpu mode, please install cuda and pytorch separatelyhttps://pytorch.org/

Training

Example code for training can be found attrain.py. To train models with the best hyperparameters, please checktrain_best_hyper.py.

Running the environment

Checkplay.py.

Alt TextAlt Text
Alt TextAlt Text

Benchmark

Drag Racing   Drag Racing   Drag Racing   Drag Racing   Drag Racing   

Citation

If you find the repository useful, please cite the study

@article{mhamed2023scpo,  title={SCPO: Safe Reinforcement Learning with Safety Critic Policy Optimization},  author={Mhamed, Jaafar and Gu, Shangding},  journal={arXiv preprint arXiv:2311.00880},  year={2023}}

About

SCPO: Safe Reinforcement Learning with Safety Critic Policy Optimization

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages


[8]ページ先頭

©2009-2025 Movatter.jp