Skip to content

Deep Reinforcement Learning for Continuous Control

Notifications You must be signed in to change notification settings

meiermark/rl-control

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Project 2: Continuous Control

Introduction

This project uses the Reacher environment.

Trained Agent

In this environment, a double-jointed arm can move to target locations. A reward of +0.1 is provided for each step that the agent's hand is in the goal location. Thus, the goal of your agent is to maintain its position at the target location for as many time steps as possible.

The observation space consists of 33 variables corresponding to position, rotation, velocity, and angular velocities of the arm. Each action is a vector with four numbers, corresponding to torque applicable to two joints. Every entry in the action vector is a number between -1 and 1.

Requirements

  1. Install anaconda click here

  2. Create (and activate) a new environment with Python 3.6.

    • Linux or Mac:
    conda create --name drl python=3.6
    source activate drl
    • Windows:
    conda create --name drl python=3.6 
    activate drl
  3. Follow the instructions in this repository to perform a minimal install of OpenAI gym.

    • Next, install the classic control environment group by following the instructions here.
    • Then, install the box2d environment group by following the instructions here.
  4. Clone the repository (if you haven't already!). Then, install several dependencies.

git clone https://github.com/meiermark/rl-navigation.git
cd rl-navigation/ml-agents
pip install .
  1. Create an IPython kernel for the drl environment.
python -m ipykernel install --user --name drl --display-name "drl"
  1. Before running code in the notebook, change the kernel to match the drl environment by using the drop-down Kernel menu.

Different Setups

  • First version contains a single agent. The corresponding docu and code can be found in single_agent
  • Second version contains 20 agents. The corresponding docu and code can be found in multiple_agents

Please follow the README.md in single_agent and multiple_agents.