From 5d9f5de52e64c2beb41cea61cece13cf414e90ce Mon Sep 17 00:00:00 2001 From: Philip Maas <philip.maas@stud.hs-bochum.de> Date: Thu, 3 Mar 2022 17:53:27 +0000 Subject: [PATCH] Update README.md --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 3dc7f10..6c0bf4c 100644 --- a/README.md +++ b/README.md @@ -6,7 +6,7 @@ This project tries to solve OpenAI's bipedal walker using three different ways: Coming soon # Action Mutation -Will get 0 reward, which is basically learning to prevent falling on it's head. The more actions the walker can use, the worse the reward. +❌ Will get 0 reward, which is basically learning to prevent falling on it's head. The more actions the walker can use, the worse the reward. This is because the walker tries to generate movement by trembling with it's legs. The covered distance doesn't cover the punishment for doing actions. So after 1600 moves the walker will get a reward around -60.  @@ -28,7 +28,7 @@ This is because the walker tries to generate movement by trembling with it's leg # Evolution Strategies After 1000 episodes, which is about 1h of learning, it will reach ~250 reward.\ -✅ Best score until now: 304/300 in 7000 episodes \ +✅ Best score until now: 304/300 in under 7000 episodes \  ## How it works -- GitLab