From 51884555000d11e4ccf72c5d60eb2481002d8ae0 Mon Sep 17 00:00:00 2001
From: Philip Maas <philip.maas@stud.hs-bochum.de>
Date: Thu, 3 Mar 2022 08:35:57 +0000
Subject: [PATCH] Update README.md

---
 README.md | 31 +++++++++++++++----------------
 1 file changed, 15 insertions(+), 16 deletions(-)

diff --git a/README.md b/README.md
index aff562d..29dc69b 100644
--- a/README.md
+++ b/README.md
@@ -19,13 +19,12 @@ This is because the walker tries to generate movement by trembling with it's leg
 
 ## Hyperparameters
 
-| Parameter         | Description                                                 | Interval  |
-|-------------------|-------------------------------------------------------------|-----------|
-| `POP_SIZE`        | Size of population.                                         | [0;∞[     |
-| `MUTATION_FACTOR` | Percentage of weights that will be mutated for each mutant. | [0;1]     |
-| `ACTIONS_START`   | Number of actions in the first generation.                  | [0;1600]  |
-| `INCREASE BY`     | Incrementation of steps for each episode.                   | [0;∞[     |
-| `MAX_STEPS`       | Number of steps that are played in one episode.             | [0; 1600] |
+| Parameter         | Description                                                 | Interval  | Our Choice |
+|-------------------|-------------------------------------------------------------|-----------|------------|
+| `POP_SIZE`        | Size of population.                                         | [0;∞[     | 50         |          
+| `MUTATION_FACTOR` | Percentage of actions that will be mutated for each walker. | [0;1]     | 0.2        |
+| `BRAIN_SIZE`      | Number of actions in the first generation.                  | [0;1600]  | 50         |
+| `INCREASE BY`     | Incrementation of steps for each episode.                   | [0;∞[     | 5          |
 
 # Evolution Strategies
 After 1000 episodes, which is about 1h of learning, it will reach ~250 reward.\
@@ -39,15 +38,15 @@ Best score until now: 292/300 in 7000 episodes \
 4. The better the reward, the higher the chance to pass weights to next gen
 
 ## Hyperparameters
-| Parameter         | Description                                                 | Interval  |
-|-------------------|-------------------------------------------------------------|-----------|
-| `HIDDEN_LAYER`    | Size of hidden layer.                                       | [1;∞[     |
-| `BIAS`            | Add a bias neuron to the input layer.                       | {0,1}     |
-| `POP_SIZE`        | Size of population.                                         | [0;∞[     |
-| `MUTATION_FACTOR` | Percentage of weights that will be mutated for each mutant. | [0;1]     |
-| `LEARNING_RATE`   | This is the rate of learning.                               | [0;1]     |
-| `GENS`            | Number of generations.                                      | [0;∞[     |
-| `MAX_STEPS`       | Number of steps that are played in one episode.             | [0; 1600] |
+| Parameter         | Description                                                 | Interval  | Our Choice |
+|-------------------|-------------------------------------------------------------|-----------|------------|
+| `HIDDEN_LAYER`    | Size of hidden layer.                                       | [1;∞[     | 12         |
+| `BIAS`            | Add a bias neuron to the input layer.                       | {0,1}     | 0          |
+| `POP_SIZE`        | Size of population.                                         | [0;∞[     | 50         |
+| `MUTATION_FACTOR` | Percentage of weights that will be mutated for each mutant. | [0;1]     | 0.1        |
+| `LEARNING_RATE`   | This is the rate of learning.                               | [0;1]     | 0.03       |
+| `GENS`            | Number of generations.                                      | [0;∞[     | 2000       |
+| `MAX_STEPS`       | Number of steps that are played in one episode.             | [0; 1600] | 300        |
 
 
 # Installation
-- 
GitLab