Skip to content

Commit

Permalink
Fix pre-commit
Browse files Browse the repository at this point in the history
  • Loading branch information
nicolengsy committed Oct 30, 2020
1 parent bbd459a commit 4ee2e66
Showing 1 changed file with 0 additions and 2 deletions.
2 changes: 0 additions & 2 deletions docs/user/algo_teppo.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,6 @@
- :ref:`te_ppo_metaworld_mt1_push`, :ref:`te_ppo_metaworld_mt10`, :ref:`te_ppo_metaworld_mt50`, :ref:`te_ppo_point`
```


Proximal Policy Optimization Algorithms (PPO) is a family of policy gradient methods which alternate between sampling data through interaction with the environment, and optimizing a "surrogate" objective function using stochastic gradient ascent. TEPPO parameterizes the PPO policy via a shared skill embedding space.

## Default Parameters
Expand Down Expand Up @@ -75,4 +74,3 @@ inference_ce_coeff=1e-3
----

*This page was authored by Nicole Shin Ying Ng ([@nicolengsy](https://github.com/nicolengsy)).*

0 comments on commit 4ee2e66

Please sign in to comment.