Skip to content

A messy repository containing various notebooks and code snippets documenting my robotics/RL progress

Notifications You must be signed in to change notification settings

AlexanderJernstrom/robotics-learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Robotics learning

This is a quite messy repo with a bunch of code snippets and notebooks related to tracking my progress in robotics and RL.

Projects

Currently contains two small little projects outlined below

Quadrotor Control

Altitude control of a quadrotor in Mujoco.

  • /rl/quadrotor_control/lqr.py - Main LQR code
  • /rl/quadrotor_control/control.ipynb - Simulation and control results

Twitter post Blog post

"Robot" arm

An exetremely simple two joint robot arm simulated in Mujoco. Uses a custom built inverse kinematics solver (uses optimization) to move the arm to a certain target. Implemented in Jax and MJX. Kind of works but still a bit janky and could be made much more dynamic. E.g relies on hand calculating Jacobian which could be done using rotation matrices (on the roadmap). Can be found in /rl/robotarm with the following files:

  • /rl/robotarm/mjx.ipynb - Notebook responsible for running the simulation in Mujoco
  • /rl/robotarm/robot_arm.xml- Mujoco XML file describing the robot arm
  • /rl/robotarm/lib.py - Various functions related to the inverse kinematics of the arm

A little demo of how it works (hardcoded version without inverse kinematics): Twitter post

TODO:

  • Simple description of the math involved and how it was done

Multiarm bandit

A very simple simulation of the multiarm bandit where a greedy and an $$\epsilon$$-greedy method were compared. This exercise is taken from the book "Reinforcement Learning: An Introduction" by Richard Sutton and Andrew Barto from the section on Multi-arm bandits. My first real Reinforcement learning project, something is wrong since $$\epsilon$$ method should perform much better, also on the roadmap to fix. Most of the code lies in /rl/multiarmbandit.ipynb.

Currently learning

  • Interested in Model Predictive Control and how that can be used. Experimenting with how it can be used for commaai's controls challenge
  • Looking into how I can get my simple robot arm into the real world using an Arduino

About

A messy repository containing various notebooks and code snippets documenting my robotics/RL progress

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published