Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Benchmarking Robustness in Object Detection: Autonomous Driving when Winter is Coming (1907.07484v2)

Published 17 Jul 2019 in cs.CV, cs.LG, and stat.ML

Abstract: The ability to detect objects regardless of image distortions or weather conditions is crucial for real-world applications of deep learning like autonomous driving. We here provide an easy-to-use benchmark to assess how object detection models perform when image quality degrades. The three resulting benchmark datasets, termed Pascal-C, Coco-C and Cityscapes-C, contain a large variety of image corruptions. We show that a range of standard object detection models suffer a severe performance loss on corrupted images (down to 30--60\% of the original performance). However, a simple data augmentation trick---stylizing the training images---leads to a substantial increase in robustness across corruption type, severity and dataset. We envision our comprehensive benchmark to track future progress towards building robust object detection models. Benchmark, code and data are publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Claudio Michaelis (8 papers)
  2. Benjamin Mitzkus (3 papers)
  3. Robert Geirhos (28 papers)
  4. Evgenia Rusak (10 papers)
  5. Oliver Bringmann (34 papers)
  6. Alexander S. Ecker (23 papers)
  7. Matthias Bethge (103 papers)
  8. Wieland Brendel (55 papers)
Citations (401)

Summary

  • The paper reveals that popular object detectors lose 30–60% performance when facing image corruptions like blur and noise.
  • It introduces three benchmark datasets to evaluate model vulnerabilities under varying weather and distortion conditions.
  • A simple style-based data augmentation technique significantly boosts detector resilience without needing extra labels or architecture changes.

Benchmarking Robustness in Object Detection: Addressing Image Corruptions in Autonomous Driving

The paper by Michaelis et al. investigates the robustness of object detection algorithms when exposed to image degradations, a scenario commonly encountered in real-world tasks like autonomous driving. The authors propose a benchmark designed to evaluate how effectively state-of-the-art models can handle adverse conditions such as blurring, noise, and weather anomalies. This paper identifies critical deficits in current object detection models and introduces data augmentation techniques to address these vulnerabilities.

The researchers introduce three benchmark datasets, PASCAL-C, COCO-C, and Cityscapes-C, which incorporate a variety of image corruptions across multiple severity levels. These corrupted datasets serve as a testing ground to measure the robustness of object detection models. The authors find that popular models like Faster R-CNN, Mask R-CNN, and others experience a performance reduction to 30–60% of their original capability when faced with image distortions. This finding highlights a significant robustness gap in existing algorithms.

A key contribution of this paper is demonstrating that the robustness of object detection models can be substantially improved using a simple data augmentation technique: stylizing training images. By integrating style-transferred versions of training images, the authors achieve notable gains in model resilience across diverse corruption types and severity levels. For instance, this augmentation approach helps models retain a higher proportion of their performance when exposed to unseen corruptions, thus bridging some of the critical existing gaps in object detection systems. This technique proves advantageous as it requires no additional labeling or architectural changes.

Importantly, the paper tackles the practical implications of robustness in autonomous vehicle applications, emphasizing the importance of model reliability in dynamic environmental conditions. For autonomous driving, the capability to accurately detect and recognize objects despite adverse weather is vital for safety and operational effectiveness. The insights obtained from this research suggest that robustness improvements could mitigate current limitations that obstruct the widespread deployment of autonomous cars in various weather environments.

Theoretical implications of this work include offering a structured method to quantify model robustness, providing a clear benchmark against which future progress can be measured. Moreover, this benchmark is openly shared, allowing for community-driven advancements in the development of robust object detection systems.

Looking forward, the research identifies potential areas for further improvement, such as exploring architectural modifications, refining data augmentation strategies, and enhancing training losses to improve robustness. The paper's benchmarks can now serve as a litmus test for new methods aiming to tackle these challenges.

In summary, the work by Michaelis et al. contributes significantly to the field of object detection by providing tools and methodologies for better understanding and improving robustness under commonly experienced corruptions, with significant implications for AI applications in safety-critical systems like autonomous vehicles.