From cc34f10b95c1c84fd9f02ef075292544b54d293f Mon Sep 17 00:00:00 2001
From: Philip Maas <philip.maas@stud.hs-bochum.de>
Date: Wed, 2 Mar 2022 21:06:42 +0000
Subject: [PATCH] Update README.md

---
 README.md | 5 ++++-
 1 file changed, 4 insertions(+), 1 deletion(-)

diff --git a/README.md b/README.md
index 1429619..42341f6 100644
--- a/README.md
+++ b/README.md
@@ -6,7 +6,9 @@ This project tries to solve OpenAI's bipedal walker using three different ways:
 Coming soon
 
 # Action Mutation
-Will get 0 reward, which is basically learning to prevent falling on it's head.
+Will get 0 reward, which is basically learning to prevent falling on it's head. The more actions the walker can use, the worse the reward.
+This is because the walker tries to generate movement by trembling with it's legs. The covered distance doesn't cover the punishment for doing actions. So after 1600 moves the walker will get a reward arounf -60
+![Reward](./MutateActions/5_50_50_0.2.png)
 
 ## How it works
 1. Generate a population with a starting number randomized actions (we don't need enough actions to solve the problem right now)
@@ -16,6 +18,7 @@ Will get 0 reward, which is basically learning to prevent falling on it's head.
 4. Mutate all children and increment their number of actions
 
 ## Hyperparameters
+
 | Parameter         | Description                                                 | Interval  |
 |-------------------|-------------------------------------------------------------|-----------|
 | `POP_SIZE`        | Size of population.                                         | [0;∞[     |
-- 
GitLab