diff --git a/README.md b/README.md index bfe13a96ddf8060b8ddc21416e8618e91d681ff8..ad53e68985d8994be42a92e99bcd6bd8f4ed0ac8 100644 --- a/README.md +++ b/README.md @@ -87,6 +87,21 @@ Also: Decay alpha and sigma to 0.05 after 1000 gens and 0.01 after 5000 gens for | `GENS` | Number of generations. | [0;∞[ | 2000 | | `MAX_STEPS` | Number of steps that are played in one episode. | [0; 1600] | 300 | +## ES Transfer: Solving the LunarLanderContinuous-v2 +✅ After 30 minutes of learning it will reach >200 reward in 100 consecutive episodes. + +Rewards of fully learned agent in 100 episodes:\ + + +| Parameter | Description | Interval | Our Choice | +|-------------------|-------------------------------------------------------------|-----------|------------| +| `HIDDEN_LAYER` | Size of hidden layer. | [1;∞[ | 4 | +| `BIAS` | Add a bias neuron to the input layer. | {0,1} | 1 | +| `POP_SIZE` | Size of population. | [0;∞[ | 50 | +| `MUTATION_FACTOR` | Percentage of weights that will be mutated for each mutant. | [0;1] | 0.1 | +| `LEARNING_RATE` | This is the rate of learning. | [0;1] | 0.1 | +| `GENS` | Number of generations. | [0;∞[ | 500 | +| `MAX_STEPS` | Number of steps that are played in one episode. | [0; 1000] | 1000 | # Installation We use Windows, Anaconda and Python 3.7 \ @@ -95,8 +110,6 @@ We use Windows, Anaconda and Python 3.7 \ `conda install swig`\ `pip install -r requirements.txt` - - # Important Sources Environment: https://github.com/openai/gym/wiki/BipedalWalker-v2 \ Table of all Environments: https://github.com/openai/gym/wiki/Table-of-environments \