Skip to content

jakegrigsby/puppersim-1

 
 

Repository files navigation

Pupper Sim

Simulation and Reinforcement Learning for DJI Pupper v2 Robot

System setup

Operating system requirements

  • Mac
  • Linux
  • Windows (untested, not recommended)

Mac-only setup

Install xcode command line tools.

xcode-select --install

If you already have the tools installed you'll get an error saying so, which you can ignore.

Conda setup

Install miniconda, then

conda create --name rl_pupper python=3.7
conda activate rl_pupper
pip install ray arspb

Getting the code ready

git clone https://github.com/jietan/puppersim.git
cd puppersim
pip install -e .

Then to verify the installation, run

python3 puppersim/pupper_example.py

You should see the PyBullet GUI pop up and see Pupper doing an exercise.

Click for instructions if pupper_example.py is running slowly

Stop pupper_example.py. Then run

python3 puppersim/pupper_minimal_server.py

then in a new terminal tab/window

python3 puppersim/pupper_example.py --render=False

This runs the visualizer GUI and simulator as two separate processes.


Training

From the outer puppersim folder run:

python3 puppersim/pupper_ars_train.py --rollout_length=200

Depending on your computer specs, each training iteration will take around 1 - 5 seconds.

Troubleshooting

Click to expand
  • Pybullet hangs when starting training. Possible issue: You have multiple suspended pybullet clients. Solution: Restart your computer.

Protocol for saving policies

Click to expand

If you want to save a policy, create a folder within puppersim/data with the type of gait and date, eg pretrained_trot_1_22_22. From the data folder, copy the following files into the folder you just made.

  • The .npz policy file you want, e.g. lin_policy_plus_latest.npz
  • log.txt
  • params.json

From puppersim/config also copy the .gin file you used to train the robot, e.g. pupper_pmtg.gin file into the folder you just made. When you run a policy on the robot, make sure your pupper_robot_*_.gin file matches the pupper_pmtg.gin file you saved.

Then add a README.md in the folder with a brief description of what you did, including your motivation for saving this policy.


Test an ARS policy

You can visualize the policy during or after training with the following command. From the outer puppersim folder run:

python3 puppersim/pupper_ars_run_policy.py  --expert_policy_file  data/lin_policy_plus_latest.npz  --json_file data/params.json --render

If you specified non-default locations for the expert policy and json files when running the training command, you should specify the correct locations in this command.

Deployment

Prerequisites

Linux

Set up Avahi (once per computer)

sudo apt install avahi-*

Run the following, you should see Pupper's IP address

avahi-resolve-host-name raspberrypi.local -4

Setup the zero password login for your pupper (once per computer) (original raspberry pi password: raspberry)

ssh-keygen
cat ~/.ssh/id_rsa.pub | ssh pi@`avahi-resolve-host-name raspberrypi.local -4 | awk '{print $2}'` 'mkdir .ssh/ && cat >> .ssh/authorized_keys'
Mac

Setup the zero password login for your pupper (only once per computer) (original raspberry pi password: raspberry)

Once per computer, run

ssh-keygen
cat ~/.ssh/id_rsa.pub | ssh [email protected] 'mkdir -p .ssh/ && cat >> .ssh/authorized_keys'

Run pretrained policy on Pupper

  • Turn on the Pupper robot, wait for it to complete the calibration motion.
  • Connect your laptop with the Pupper using an USB-C cable
  • Run the following command on your laptop:
./deploy_to_robot.sh python3 puppersim/puppersim/pupper_ars_run_policy.py --expert_policy_file=puppersim/data/lin_policy_plus_latest.npz --json_file=puppersim/data/params.json --run_on_robot

Simulating the heuristic controller

Click to expand Navigate to the outer puppersim folder and run
python3 puppersim/pupper_server.py

Clone the the heuristic controller:

git clone https://github.com/stanfordroboticsclub/StanfordQuadruped.git
cd StanfordQuadruped
git checkout dji

In a separate terminal, navigate to StanfordQuadruped and run

python3 run_djipupper_sim.py

Keyboard controls:

  • wasd --> moves robot forward/back and left/right
  • arrow keys --> turns robot left/right
  • q --> activates/deactivates robot
  • e --> starts/stops trotting gait
  • ijkl --> tilts and raises robot

About

Simulation for DJI Pupper v2 robot

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 60.1%
  • Python 39.7%
  • Shell 0.2%