YOLO - You only look once 10647 times
Authors: Christian Limberg, Andrew Melnik, Helge Ritter, Helmut Prendinger
Abstract
In this work, we explore the You Only Look Once (YOLO) single-stage object detection architecture and compare it to the simultaneous classification of 10647 fixed region proposals.
We use two different approaches to demonstrate that each of YOLO's grid cells is attentive to a specific sub-region of previous layers. This finding makes YOLO's method comparable to local region proposals.
Such insight reduces the conceptual gap between YOLO-like single-stage object detection models, R-CNN-like two-stage region proposal based models, and ResNet-like image classification models.
This page shows interactive exploration tools and exported media for a better visual understanding of the YOLO information processing streams.