Ensemble Methods for Route Choice
Recently published:
Wang, Haotian, Moylan, E. and Levinson, D. (2023) Ensemble Methods for Route Choice . Transportation Research part C. Volume 167, October 2024, 104803 [doi]
Abstract
Understanding travellers’ route preferences allows for the calculation of traffic flow on network segments and helps in assessing facility requirements, costs, and the impact of network modifications. Most research employs logit-based choice methods to model the route choices of individuals, but machine learning models are gaining increasing interest. However, all of these methods typically rely on a single ‘best’ model for predictions, which may be sensitive to measurement errors in the training data. Moreover, predictions from discarded models might still provide insights into route choices. The ensemble approach combines outcomes from multiple models using various pattern recognition methods, assumptions, and/or data sets to deliver improved predictions. When configured correctly, ensemble models offer greater prediction accuracy and account for uncertainties. To examine the advantages of ensemble techniques, a data set from the I-35 W Bridge Collapse study in 2008, and another from the 2011 Travel Behavior Inventory (TBI), both in Minneapolis–St. Paul (The Twin Cities) are used to train a set of route choice models and combine them with ensemble techniques. The analysis considered travellers’ socio-demographics and trip attributes. The trained models are applied to two datasets, the Longitudinal Employer-Household Dynamics (LEHD) commute trips and TBI morning peak trips, for validation. Predictions are also compared with the loop detector records on freeway links. Traditional Multinomial Logit and Path-Size Logit models, along with machine learning methods such as Decision Tree, Random Forest, Extra Tree, AdaBoost, Support Vector Machine, and Neural Network, serve as the foundation for this study. Ensemble rules are tested in both case studies, including hard voting, soft voting, ranked choice voting, and stacking. Based on the results, heterogeneous ensembles using soft voting outperform the base models and other ensemble rules on testing sets.