TLDR: This paper introduces a novel method for creating simple, understandable, and accurate dynamic models for jumping quadruped robots. It combines a linear autoencoder to reduce complex robot movements into a low-dimensional space with SINDy (Sparse Identification of Nonlinear Dynamics) to discover interpretable symbolic equations for these movements. The approach, validated in simulations and on real robots, outperforms traditional models and shows robust performance, paving the way for more efficient and transparent robot control.
Robots that can move like animals, especially quadruped robots, are becoming increasingly important for tasks like search and rescue or exploring dangerous areas. However, controlling these robots, particularly during complex movements like jumping, is a significant challenge. This is because their full-body dynamics are very intricate and require a lot of computing power to model accurately in real-time.
Traditional methods for modeling robot movements, such as the Spring-Loaded Inverted Pendulum (SLIP) model, simplify these complex dynamics to make them easier to compute. While these simplified models are efficient, they often lose accuracy, especially when the robot interacts with the ground in multiple ways or when its limbs move in complex, non-linear patterns. On the other hand, advanced machine learning techniques can achieve high accuracy but often result in “black box” models that are difficult for humans to understand or interpret.
A new research paper introduces a clever way to overcome these limitations by combining two powerful techniques: Sparse Identification of Nonlinear Dynamics (SINDy) and linear autoencoders. This novel approach aims to create simplified, yet accurate and understandable, dynamic models specifically for jumping quadruped robots. You can read the full paper here: Symbolic Learning of Interpretable Reduced-Order Models for Jumping Quadruped Robots.
How the New Approach Works
The core idea is to represent the robot’s complex, high-dimensional movements in a much simpler, low-dimensional “latent space.” Imagine taking a very detailed, multi-layered map and compressing it into a simpler, more abstract representation that still captures the most important features. This is achieved using a “linear autoencoder,” which is a type of neural network designed to reduce data dimensionality while keeping a clear, direct relationship with the original data. This linearity is crucial because it helps maintain the interpretability of the model.
Once the robot’s state (its position, speed, etc.) is mapped into this simplified latent space, the SINDy algorithm takes over. SINDy is a method that can “discover” the underlying governing equations of a system from data. It does this by searching for a sparse (meaning, having very few terms) combination of simple mathematical functions (like polynomials or trigonometric functions) that best describe the robot’s dynamics in the latent space. The “sparse” nature of the solution is key to making the resulting equations easy to understand and interpret.
The training process for this model is also innovative. Jumping is a “hybrid” motion, meaning it involves distinct phases like being in contact with the ground (takeoff and landing) and being airborne (flight). To handle this, the researchers developed a sequential training strategy. First, the autoencoder learns to reconstruct the robot’s configuration during the contact phase. Then, with the encoder fixed, the decoder is fine-tuned using data from all jumping phases. Finally, the SINDy algorithm learns the symbolic dynamics for each phase, ensuring the overall model is consistent and physically accurate across the entire jump.
Validation and Results
The effectiveness of this new methodology was rigorously tested using both simulated and real-world data. In simulations, the researchers used Unitree Go1 robots for synchronized jumps (where all legs leave and land at the same time) and Unitree A1 robots for “froggy” jumps (which involve a more complex sequence of leg contacts). The model was also validated on an actual Go1 robot, demonstrating its practical applicability even with noisy real-world data.
The results were promising. For the Go1 robot, a two-dimensional latent model showed that the learned dimensions primarily corresponded to the robot’s body position in the horizontal (x) and vertical (z) directions. When compared to the traditional actuated Spring-Loaded Inverted Pendulum (aSLIP) model, the new learned model showed superior accuracy, especially in predicting the robot’s horizontal movement during a jump.
Further analysis revealed that using 4 to 6 latent dimensions provided the best balance between model accuracy and complexity. Adding more dimensions beyond this range offered only marginal improvements in accuracy but made the model significantly more complex. Even with the more complex “froggy” jump of the A1 robot, which involves a noticeable angular rotation, the four-dimensional model successfully captured these dynamics and most joint positions.
Crucially, the model demonstrated robust performance when transferring from simulation to the real world. Despite the inherent noise and complexities of real-world data, the model, initially trained on simulated data and then fine-tuned with limited real-world samples, effectively replicated actual jumps. This highlights its ability to retain core dynamic features and generalize well across different environments.
Also Read:
- Enhancing Robot Planning with Causal Structure Learning
- AI-Powered Navigation: Enhancing Robot Safety in Pedestrian Environments
Future Outlook
This research marks a significant step towards creating more interpretable and efficient models for complex robot behaviors. The researchers plan to expand this methodology to other dynamic movements like walking and running, and to more complex robotic systems, including humanoids, which involve even more rapid contact changes and higher-dimensional structures.


