In this paper:
Contents:
Modular Controller Design
most popular conventional approach to controlling physical legged sys.
breaks the control problem down into smaller submodules. Each module is based on template dynamics or heuristics and generates reference values for the next module.
Template-dynamics-based control module
approximates the robot as a point mass with a massless limb to compute the next foothold position
Given the foothold positions, the next module computes a parameterized trajectory for the foot to follow. The last module tracks the trajectory with a simple PID controller
Drawbacks:
limited detail in the modeling constrains the model’s accuracy ⇒ limiting the operational state domain of each module. (i.e. slow acceleration, fixed upright pose of the body, and limited velocity of the limbs)
design of modular controllers is extremely laborious. Imagine this arduous work for every new robot or even for every new maneuver
Trajectory Optimization(TO)
able to mitigate the aforementioned problems.
two modules: planning and tracking
a series of approximations are employed to reduce complexity.
<aside> 💡 Need to find more about TO…
</aside>
Reinforcement Learning
overcome the limitations of prior model-based approaches by learning effective controllers directly from experience.
Drawbacks:
Direct application of learning methods to physical legged systems is complicated and thus often applied to simple models.
TO + DRL
improve simulation fidelity analytically
improve simulation fidelity in a data-driven way
Command input: forward velocity, lateral velocity, and yaw rate.
Policy network: maps the observation of the current state → joint state history to the joint position targets
Actuator network: maps joint state history → joint position targets to 12 joint torque values
Rigid-body simulator: outputs the next state of the robot, given the joint torques and the current state as input