TLDR: This research introduces a novel partial column generation strategy that uses Graph Neural Networks (GNNs) to accelerate the solution of the team formation and routing problem. The GNN model predicts which pricing problems are likely to yield beneficial solutions, allowing the optimization algorithm to focus computational effort more efficiently. Computational experiments demonstrate that this GNN-based approach significantly outperforms traditional methods, especially on challenging instances, by finding more optimal solutions and improving solution quality under tight time limits. While effective, the study notes potential performance degradation under very long runtimes, suggesting areas for future refinement.
Optimizing complex operations like assigning skilled workers to teams and planning their routes for tasks is a significant challenge in various industries, including airports, healthcare, and maintenance. This intricate problem, known as the team formation and routing problem, requires efficient solutions, especially in dynamic environments where plans need frequent adjustments.
Traditional methods often rely on exact solution techniques like column generation. Column generation is a powerful mathematical approach used to solve large optimization problems by breaking them down into a smaller ‘master problem’ and one or more ‘pricing problems’. The pricing problems identify potential improvements to the current solution, which are then added back to the master problem in an iterative process until an optimal solution is found.
While effective, column generation can be computationally intensive, particularly when dealing with multiple pricing problems. Solving all pricing problems in every iteration can consume significant time and resources, especially if many of them do not yield useful improvements. This is where the concept of ‘partial column generation’ comes into play – a strategy designed to accelerate the process by selectively solving only a subset of the pricing problems.
A recent research paper, titled “Partial Column Generation with Graph Neural Networks for Team Formation and Routing,” introduces a novel approach to partial column generation. Authored by Giacomo Dall’Olio, Rainer Kolisch, and Yaoxin Wu, this paper proposes integrating machine learning, specifically Graph Neural Networks (GNNs), into the column generation routine to make smarter decisions about which pricing problems to solve. You can find the full research paper here: RESEARCH_PAPER_URL.
A Smarter Way to Select Pricing Problems
The core idea behind this new strategy is to predict which pricing problems are most likely to generate ‘negative columns’ – that is, solutions that can actually improve the overall optimization. By focusing computational effort only on these promising problems, the algorithm can significantly reduce the time spent on less fruitful calculations.
The researchers developed a machine learning model tailored for the team formation and routing problem that uses Graph Neural Networks for these predictions. GNNs are particularly well-suited for this task because they can directly process data structured as graphs, which naturally aligns with how pricing problems (often elementary shortest path problems with resource constraints) are represented. Furthermore, GNNs can handle varying input sizes, a crucial feature since the complexity of pricing problem instances can differ.
The GNN model takes a graph representation of a single pricing problem instance and estimates the probability that its optimal value will be negative. A higher estimated probability means the pricing problem is more likely to yield a useful column. This prediction guides the selection process, ensuring that only the most promising pricing problems are solved.
How the GNN Works
To feed the pricing problem instances into the GNN, the researchers devised a unique ‘partially bipartite graph’ representation. This graph effectively captures all relevant information, including task-specific parameters, time-dependent factors, and stochastic travel times between locations. The graph has two main types of nodes: ‘transportation nodes’ representing tasks and ‘supplementary nodes’ representing discrete points in time. Edges connect tasks to other tasks (transportation graph) and tasks to time points (supplementary edges), allowing the model to understand both spatial and temporal relationships.
The GNN’s architecture involves several layers that process and transform this graph data. An initial ’embedding layer’ converts raw features into a richer numerical representation. ‘Convolutional layers’ then allow information to flow between neighboring nodes, mimicking how information would spread in a real-world network. ‘Bipartite convolutional layers’ specifically handle the message passing between the transportation and supplementary nodes, integrating temporal and task-specific information. Finally, a ‘pooling layer’ aggregates all this processed information into a single representation, which is then fed to a ‘final layer’ that outputs the probability prediction.
The model is trained using supervised learning, where it learns from a large dataset of pricing problem instances that have been solved exactly, with their outcomes (negative or non-negative optimal value) serving as labels. To address imbalances in the training data, the researchers employed undersampling techniques, ensuring the model learns effectively from both easy and hard-to-classify instances.
Performance and Impact
Computational experiments demonstrated that this GNN-based partial column generation strategy significantly enhances the solution method. When compared to traditional partial column generation approaches and even full column generation, the GNN strategy consistently outperformed them, especially on challenging instances solved under tight time limits. The algorithm was able to solve more instances, find more optimal solutions, and achieve higher quality suboptimal solutions with lower optimality gaps.
For instance, on hard instances characterized by low worker availability, the GNN strategy showed an even more pronounced positive impact, increasing the share of solved and optimally solved instances and substantially decreasing optimality gaps. While the GNN strategy does introduce a computational overhead (around 5.43% for feature collection and prediction), the benefits in terms of solution quality and speed far outweigh this cost.
However, the study also noted a limitation: the GNN strategy’s performance can degrade under extended time limits. This is attributed to the increasing complexity and heterogeneity of pricing problems encountered in the later stages of the optimization process, as well as a potential lack of training data for such prolonged scenarios. Future research aims to address this by developing metrics to determine when to switch to other strategies or by refining the deep learning model for better long-term performance.
Also Read:
- Teaching Neural Networks to Solve Knapsack: A Two-Phase Algorithmic Approach
- Smart Control: How AI Teams Learn Safely with a Hierarchical Approach
Conclusion
This research marks a significant step forward in solving complex team formation and routing problems. By intelligently integrating Graph Neural Networks into partial column generation, the proposed strategy offers a powerful tool for accelerating optimization algorithms, leading to faster and higher-quality solutions in real-world applications like airport operations. The findings also validate the effectiveness of the partially bipartite graph representation and the architectural choices for the deep neural model, paving the way for similar ML-based optimization approaches in other challenging combinatorial problems, such as rich vehicle routing problems.


