AR and Computer Vision Revolutionize Bridge Inspections

In a paper published in the journal Sensors, researchers introduced a novel bridge inspection methodology using computer vision and augmented reality (AR) to improve fatigue crack detection.

Study: AR and Computer Vision Revolutionize Bridge Inspections. Image Credit: S.Phoophinyo/Shutterstock
Study: AR and Computer Vision Revolutionize Bridge Inspections. Image Credit: S.Phoophinyo/Shutterstock

The method employs a computer vision algorithm to analyze structural surface motion from AR headset videos, tracking feature points and measuring distance variations for efficient crack detection without camera motion compensation. Results are displayed as holograms on the bridge surface in an AR environment, aiding inspectors with enhanced visualization and decision-making. This approach shows significant potential for real-world applications validated through laboratory tests.

Background

Past steel bridge fatigue crack inspection work has primarily relied on human vision, guided by the Federal Highway Administration (FHWA) and state transportation guidelines. However, visual inspections show significant limitations, with inspectors only achieving 50% detection rates for 26 mm cracks and 90% for 113 mm cracks.

More reliable detection methods are crucial given that 42% of United States (U.S.) bridges are over 50 years old and vulnerable to fatigue cracks. Image processing and artificial intelligence (AI) based techniques often struggle to differentiate real cracks from similar features like corrosion marks. AR offers a promising solution, enhancing crack detection with additional visual information and interactive tools. Although AR devices have limited computational power, they can use wireless connections for real-time data processing.

Efficient AR-Based Crack Detection

The proposed human-centered bridge inspection process begins by creating a local area network (LAN) using a Wi-Fi router, connecting an AR headset, such as the HoloLens 2 (HL2), to a server computer that runs the crack detection algorithm and stores data. Using the AR headset's camera, the bridge inspector records a short video of a fatigue-crack-prone region. This video is automatically transmitted to the server via the LAN, where it is processed for crack detection. Alternatively, a cellular network can connect to a remote server for data processing and storage.

An ROI is selected from the first video frame on the server, and distinctive feature points are tracked throughout the video. These features are grouped into local circular regions (LCRs) to calculate distances between feature point pairs. Significant distance variations, indicating fatigue cracks, are identified by analyzing the distance history of feature point pairs within the LCRs. This method enhances computational efficiency and enables near-real-time results by eliminating the need for global motion compensation (GMC).

A range of threshold values is applied to determine the optimal threshold value for crack detection, producing various candidate detection results. These results are converted into holograms and overlaid onto the real-world view of the structure via the AR device. An interactive virtual menu allows the inspector to toggle through threshold values and select the optimal one, enhancing crack detection and aiding in informed decision-making during the inspection.

The proposed human-centered bridge inspection process empowered by AR and computer vision.

Enhanced Crack Detection

The proposed crack detection algorithm, adapted from previous work, identifies crack-induced surface motion under fatigue loading by analyzing discontinuities caused by the opening and closing of cracks in recorded video. Feature points with high-intensity gradients are detected using the Shi–Tomasi algorithm and tracked across video frames using the Kanade–Lucas–Tomasi (KLT) tracker.

Unlike methods requiring GMC, this approach tracks distance changes between feature point pairs, significantly enhancing computational efficiency and enabling near-real-time detection.

By evaluating LCRs around feature points, the algorithm identifies significant distance variations indicating cracks, even amidst global camera movements. The results are presented as holograms in an AR environment, allowing inspectors to select optimal threshold values for accurate crack detection interactively. They are visualized as clusters of highlighted feature points. This method improves computational efficiency and enhances crack detection accuracy, facilitating more informed decision-making in field applications.

Quantifying Crack Detection

The study presents a comprehensive methodology for quantifying crack detection outcomes using vision-based approaches. It introduces a novel distance-based algorithm that tracks feature points to identify crack-induced surface motion in recorded videos.

By employing density-based spatial clustering of applications (DBSCAN) for clustering, the method accurately segments crack regions based on detected feature points. Evaluation metrics such as intersection over union (IOU) assess detection accuracy against ground truth boundaries. Results demonstrate improved performance over previous displacement-based methods, showcasing higher IOU scores and more precise crack localization in both 2D and 3D video scenarios.

Fatigue Load Impact

The study conducted a parametric investigation using vision-based approaches into the influence of fatigue load levels on crack detection accuracy. Videos captured from two distinct viewpoints were analyzed across 10 fatigue load cases, ranging from 0.9 to 22.2 kN.

Results indicated varying IOU scores, with higher scores correlating to increased load levels, demonstrating enhanced capability in crack detection as load intensity rose. The method proved consistent in detecting cracks with openings exceeding 0.5 mm, showcasing robust performance across different load scenarios in 2D and 3D video analyses. Additionally, integrating an AR environment facilitated efficient inspection, allowing inspectors to visualize and assess crack detection results in near-real-time, thereby enhancing inspection efficiency and effectiveness.

Conclusion

In summary, the paper introduced an advanced methodology for human-centered bridge fatigue crack inspection by integrating computer vision and AR. The approach utilized video feature tracking from a moving camera to detect fatigue cracks in near-real-time, empowering AR-enabled decision-making during inspections. It analyzed differential surface motion through feature points, successfully detecting and localizing fatigue cracks without requiring GMC.

Experimental validation on various specimens confirmed superior accuracy and computational efficiency compared to traditional methods, supported by IOU metric quantification and parametric studies on fatigue load impact. The developed AR environment, featuring cloud database integration, hologram generation, and interactive tools, demonstrated its potential for enhancing field inspection capabilities.

Journal reference:

Article Revisions

  • Jun 18 2024 - Addition of graphic from the journal paper. "The proposed human-centered bridge inspection process empowered by AR and computer vision."
Silpaja Chandrasekar

Written by

Silpaja Chandrasekar

Dr. Silpaja Chandrasekar has a Ph.D. in Computer Science from Anna University, Chennai. Her research expertise lies in analyzing traffic parameters under challenging environmental conditions. Additionally, she has gained valuable exposure to diverse research areas, such as detection, tracking, classification, medical image analysis, cancer cell detection, chemistry, and Hamiltonian walks.

Citations

Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    Chandrasekar, Silpaja. (2024, June 17). AR and Computer Vision Revolutionize Bridge Inspections. AZoAi. Retrieved on November 18, 2024 from https://www.azoai.com/news/20240617/AR-and-Computer-Vision-Revolutionize-Bridge-Inspections.aspx.

  • MLA

    Chandrasekar, Silpaja. "AR and Computer Vision Revolutionize Bridge Inspections". AZoAi. 18 November 2024. <https://www.azoai.com/news/20240617/AR-and-Computer-Vision-Revolutionize-Bridge-Inspections.aspx>.

  • Chicago

    Chandrasekar, Silpaja. "AR and Computer Vision Revolutionize Bridge Inspections". AZoAi. https://www.azoai.com/news/20240617/AR-and-Computer-Vision-Revolutionize-Bridge-Inspections.aspx. (accessed November 18, 2024).

  • Harvard

    Chandrasekar, Silpaja. 2024. AR and Computer Vision Revolutionize Bridge Inspections. AZoAi, viewed 18 November 2024, https://www.azoai.com/news/20240617/AR-and-Computer-Vision-Revolutionize-Bridge-Inspections.aspx.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
AI Framework Transforms Scene Representation with Precise, Editable 3D and 4D Visuals