From bc97e9abee868e22fccf55183e5d5ea612e6297e Mon Sep 17 00:00:00 2001 From: Philip Maas <philip.maas@stud.hs-bochum.de> Date: Wed, 2 Mar 2022 21:07:53 +0000 Subject: [PATCH] Update README.md --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 42341f6..aff562d 100644 --- a/README.md +++ b/README.md @@ -7,7 +7,7 @@ Coming soon # Action Mutation Will get 0 reward, which is basically learning to prevent falling on it's head. The more actions the walker can use, the worse the reward. -This is because the walker tries to generate movement by trembling with it's legs. The covered distance doesn't cover the punishment for doing actions. So after 1600 moves the walker will get a reward arounf -60 +This is because the walker tries to generate movement by trembling with it's legs. The covered distance doesn't cover the punishment for doing actions. So after 1600 moves the walker will get a reward around -60.  ## How it works @@ -61,6 +61,6 @@ We use Windows, Anaconda and Python 3.7 \ # Important Sources Environment: https://github.com/openai/gym/wiki/BipedalWalker-v2 \ -Table of all Environments: https://github.com/openai/gym/wiki/Table-of-environments +Table of all Environments: https://github.com/openai/gym/wiki/Table-of-environments \ OpenAI Website: https://gym.openai.com/envs/BipedalWalker-v2/ \ More on evolution strategies: https://openai.com/blog/evolution-strategies/ -- GitLab