Skip to content
GitLab
Explore
Sign in
Register
Primary navigation
Search or go to…
Project
Bipedal Walker Evo
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Package registry
Container registry
Model registry
Operate
Environments
Terraform modules
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Philip Maas
Bipedal Walker Evo
Commits
dd6e4b74
Commit
dd6e4b74
authored
3 years ago
by
Philip Maas
Browse files
Options
Downloads
Patches
Plain Diff
Update README.md
parent
e12a37ee
No related branches found
No related tags found
No related merge requests found
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
README.md
+29
-4
29 additions, 4 deletions
README.md
with
29 additions
and
4 deletions
README.md
+
29
−
4
View file @
dd6e4b74
# Bipedal Walker Evo
This project tries to solve OpenAI's bipedal walker with an evolutionary strategy.
\
This project tries to solve OpenAI's bipedal walker using three different ways: Q-Learning, Mutation of Actions and Evolution Strategies.
\
# Q-Learning
Coming soon
# Action Mutation
Will get 0 reward, which is basically learning to prevent falling on it's head.
## How it works
1.
Generate a population with a starting number randomized actions (we don't need enough actions to solve the problem right now)
2.
Let the population play the game reward every walker of the generation accordingly
3.
The best walker survives without mutating
3.
The better the reward the higher the chance to pass actions to next generation. Each child has a single parent, no crossover.
4.
Mutate all children and increment their number of actions
## Hyperparameters
| Parameter | Description | Interval |
|-------------------|-------------------------------------------------------------|-----------|
|
`POP_SIZE`
| Size of population. | [0;∞[ |
|
`MUTATION_FACTOR`
| Percentage of weights that will be mutated for each mutant. | [0;1] |
|
`ACTIONS_START`
| Number of actions in the first generation. | [0;1600] |
|
`INCREASE BY`
| Incrementation of steps for each episode. | [0;∞[ |
|
`MAX_STEPS`
| Number of steps that are played in one episode. | [0; 1600] |
# Evolution Strategies
After 1000 episodes, which is about 1h of learning, it will reach ~250 reward.
\
Best score until now: 292/300
Best score until now: 292/300 in 7000 episodes
\

## How it works
1.
Generate a randomly weighted neural net
...
...
@@ -22,7 +47,7 @@ Best score until now: 292/300
|
`MAX_STEPS`
| Number of steps that are played in one episode. | [0; 1600] |
#
# Installation
# Installation
We use Windows, Anaconda and Python 3.7
\
`conda create -n evo_neuro python=3.7`
\
`conda activate evo_neuro`
\
...
...
@@ -31,7 +56,7 @@ We use Windows, Anaconda and Python 3.7 \
#
#
Sources
#
Important
Sources
Environment: https://github.com/openai/gym/wiki/BipedalWalker-v2
\
Table of all Environments: https://github.com/openai/gym/wiki/Table-of-environments
OpenAI Website: https://gym.openai.com/envs/BipedalWalker-v2/
\
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment