Skip to content

AI4ALL-Ignite-S25-C18-GE/BridgeSense

Repository files navigation

BridgeSense | A Structural Integrity Forecast Engine

Problem Statement

This project's goal is to predict if a bridge in Georgia will receive a 'Poor' condition rating, using available NBI features such as traffic, environmental factors, and construction material. Given that a poor rating indicates serious problems that may lead to failure in the near future, this prediction provides critical insight that can be used proactively to take active measures and avoid collapse, rather than waiting for the scheduled times where bridges are inspected.

Key Results

  • Examined over 500,000 data points of bridges in Georgia from 1900 to 2022.
  • Identified three biases
    • Inspection Frequency Bias: Depending on how often bridges were inspected due to various factors such as location and economics, results may skew toward bridges that get inspected more often.
    • Geographic Bias: There will be some unpredictable factors that alter the result, which decreases the validity of the study, namely the focus of bridges in one location, Georgia. Specific features of Georgia may play a large role in our results.
    • Algorithmic Bias: It may ignore outliers, misjudge whether or not it is a linear relationship, and may ignore multiple factors that influence the outcome.
  • Evaluated four distinct modeling approaches: Random Forest, XGBoost, CatBoost, and a soft-voting Ensemble of all three.
    • XGBoost was selected as the final champion model. While the Ensemble model achieved a slightly higher F1-score for the "Poor" class, XGBoost was chosen for its significantly superior 88% Recall
  • Final XGBoost SHAP analysis
    • Class 0: Poor, Class 1: Fair, Class 2: Good
    • Bridge Age and bridge roadway width curb to curb were identified as the most significant predictors of structural condition SHAP summary

Methodologies

  • Feature Engineering: Automated feature selection with RFECV; created interaction and non-linear features

  • Class Imbalance Handling: Applied SMOTE and class weighting

  • Hyperparameter Tuning: Optimized models using Optuna (Bayesian Optimization)

  • Model Interpretability: Explained model predictions using SHAP visualizations

  • Model Deployment: Deployed the final model as an interactive web app using Streamlit, hosted on Render → Live Demo

Data Sources

Kaggle NBI Datasets (https://www.fhwa.dot.gov/bridge/nbi/ascii.cfm)

Technologies Used

  • Core Stack: Python, Pandas, NumPy, Scikit-learn

  • ML Libraries: XGBoost, CatBoost, Imbalanced-learn, Optuna, SHAP

  • Deployment & Tools: Colab, Streamlit, Render, Joblib, VS Code, Git, Git LFS

  • Visualization: Matplotlib, Seaborn

Authors

This project was completed in collaboration with:

About

AI4ALL SU25

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages