Abstract for presentation (Poster or Podium) with a Paper in the Conference Proceedings
Intelligent Transportation Systems
Mustafa Attallah, B.Sc.
Doctoral Student
Saint Louis University
Saint Louis, Missouri, United States
Jalil Kianfar, PhD, PE
Associate Professor
Saint Louis University
Saint Louis, MO, United States
Jalil Kianfar, PhD, PE
Saint Louis University
Saint Louis, Missouri, United States
Machine learning models are increasingly being utilized to develop predictive models for Transportation Systems Management and Operations (TSMO) applications. These models are often assessed based on a global performance metric that evaluates the model's performance when the entire testing dataset is presented to the model. A TSMO application is expected to perform reliably and consistently in various situations and roadway conditions. The reliability and consistency of the model predictions for various scenarios are critical to the success of transportation agencies’ efforts to address mobility and safety issues. Systematic bias might be influencing the model when a model's performance is inconsistent for different scenarios. This paper investigates the systematic bias that the traffic management center may face when applying machine learning methods to predict incident clearance time. Additionally, dimension-reduction techniques are employed as mitigation techniques in the model development process. This paper investigates the impact of two common dimension-reduction methods, important feature selection and principal component analysis, on algorithmic bias.
 In a case study, this paper investigates the systematic bias of RF, BRNN, KNN, SVR, NNET, and XGB in incident clearance time prediction. Incident data from three interstate corridors in Missouri, USA, were utilized to develop and evaluate the models. Repeated k-fold validation was used to prepare thirty pairs of training and testing sets to demonstrate and assess the learners' performance variations due to data splits. The results indicated that the six learners suffer from systematic bias. The analysis of the impact of dimension-reduction algorithms revealed that the important feature selection method did not significantly mitigate the systematic bias. On the other hand, the principal component analysis method significantly addressed this bias for all learners, with poor-performing learners gaining the most improvements. In addition to contributing to reducing bias, the principal component analysis significantly reduced the learners' overall error metrics.