Skip to content

Latest commit

 

History

History
24 lines (15 loc) · 808 Bytes

README.md

File metadata and controls

24 lines (15 loc) · 808 Bytes

Soving Continuous Control with Episodic Memory

PyTorch implementation of Episodic Memory Actor-Critic (EMAC).

alt text

TD3 and DDPG architecture parameters were based on official TD3 implementation: link

Usage

For training run:

python train.py --policy EMAC --env Walker2d-v3 --k 2 --alpha 0.1 --beta 0.1 --max_timesteps 200000 --device cuda:0

Results

Paper training curves can be found in curves directory as saved TensorBoard logs in json format. For producing results below run

bash scripts/Walker2d-v3/train_EMAC.sh

results