TLDR: HOTA (Hamiltonian Optimal Transport Advection) is a new method that uses a Hamiltonian-Jacobi-Bellman framework to solve dynamic optimal transport problems. It’s designed to efficiently optimize data trajectories without needing explicit density modeling, even with non-smooth costs. HOTA outperforms existing methods in matching target distributions and finding optimal paths, and it scales well to high-dimensional data.
In the realm of artificial intelligence and machine learning, understanding and guiding the movement of data, often represented as probability distributions, is a fundamental challenge. This is where the concept of Optimal Transport (OT) comes into play. Traditionally, OT has been a powerful tool for comparing distributions and finding the most cost-effective way to transform one into another. However, classical OT often falls short when dealing with complex data landscapes, as it tends to produce simple, straight-line paths that don’t account for the intricate geometry of the underlying data manifold.
This limitation led to the development of Dynamical Optimal Transport, which considers the evolution of probability distributions over time. This dynamic approach allows for more nuanced control over trajectories, incorporating factors like velocity, acceleration, and energy along the paths. It’s particularly useful for spaces with non-trivial geometries, such as those with curvature or obstacles, or where movement is influenced by various “potentials.” A key problem within this dynamic framework is the Generalized Schrödinger Bridge (GSB) problem, which aims to find optimal stochastic trajectories between two distributions while considering additional costs or rewards along the path.
Existing methods for solving GSB, particularly those based on Hamilton-Jacobi-Bellman (HJB) equations, have faced significant hurdles. These include unstable optimization, high-variance gradients, poor sample efficiency in high dimensions, and a lack of strict terminal distribution matching. Furthermore, they often require smooth, differentiable cost functions, limiting their applicability to real-world scenarios where costs might be non-smooth or even discontinuous.
Addressing these challenges, researchers Nazar Buzun, Daniil Shlenskii, Maxim Bobrin, and Dmitry V. Dylov have introduced a novel approach called Hamiltonian Optimal Transport Advection, or HOTA. HOTA offers a robust and efficient solution to the dual dynamical OT problem by explicitly leveraging Kantorovich potentials within an HJB-based framework. This innovative method effectively bypasses the need for explicit density modeling, a common requirement in many generative models, and remarkably, it performs well even when the cost functions are non-smooth.
HOTA’s core contributions are significant. Firstly, it provides a Hamiltonian dual reformulation of dynamic OT, linking Kantorovich potentials with an HJB value function. This results in a “density-free” objective, which simplifies the learning process and offers performance gains over previous methods. Secondly, HOTA is designed to be robust to complex geometries and can handle non-smooth cost functions, thanks to its explicit incorporation of potential terms. Lastly, empirical evaluations demonstrate that HOTA achieves state-of-the-art results across a diverse set of tasks. It consistently outperforms existing dynamic OT solvers in terms of both “feasibility” (how accurately it matches the target distribution) and “optimality” (the efficiency and cost-effectiveness of the generated trajectories).
The method’s implementation involves approximating the value function using a neural network and optimizing it based on two criteria: a potential matching loss (Lpot) that ensures the target distribution is met, and an HJB residual loss (Lhjb) that enforces the HJB equation. A clever adaptive gradient balancing mechanism and a replay buffer are employed to maintain numerical stability and improve training efficiency, especially in high-dimensional spaces.
Experiments showcased HOTA’s superior performance on various 2D datasets, including those with challenging non-differentiable potentials like BabyMaze, Slit, and Box. It also demonstrated impressive scalability to high-dimensional spaces, such as N-dimensional unit spheres, maintaining stable performance as dimensionality increased. Furthermore, HOTA was successfully applied to a high-dimensional opinion depolarization task, outperforming baselines in controlling complex stochastic processes.
Also Read:
- Tackling High-Dimensional Control Games with PINN-Based Policy Iteration
- RAD: Enhancing Decision-Making in Offline Reinforcement Learning Through Dynamic Trajectory Retrieval
In conclusion, HOTA represents a significant advancement in optimal transport, offering a powerful and versatile framework for guiding probability flows. Its ability to handle complex geometries and non-smooth costs, combined with its robust empirical performance, positions it as a leading solution for dynamic optimal transport problems. For more in-depth details, you can refer to the full research paper here.


