%0 Conference Proceedings %T Temporal Coherence for Active Learning in Videos %A Javad Zolfaghari Bengar %A Abel Gonzalez-Garcia %A Gabriel Villalonga %A Bogdan Raducanu %A Hamed H. Aghdam %A Mikhail Mozerov %A Antonio Lopez %A Joost Van de Weijer %B IEEE International Conference on Computer Vision Workshops %D 2019 %F Javad Zolfaghari Bengar2019 %O LAMP; ADAS; 600.124; 602.200; 600.118; 600.120; 600.141 %O exported from refbase (http://refbase.cvc.uab.es/show.php?record=3294), last updated on Tue, 08 Feb 2022 12:09:19 +0100 %X Autonomous driving systems require huge amounts of data to train. Manual annotation of this data is time-consuming and prohibitively expensive since it involves human resources. Therefore, active learning emerged as an alternative to ease this effort and to make data annotation more manageable. In this paper, we introduce a novel active learning approach for object detection in videos by exploiting temporal coherence. Our active learning criterion is based on the estimated number of errors in terms of false positives and false negatives. The detections obtained by the object detector are used to define the nodes of a graph and tracked forward and backward to temporally link the nodes. Minimizing an energy function defined on this graphical model provides estimates of both false positives and false negatives. Additionally, we introduce a synthetic video dataset, called SYNTHIA-AL, specially designed to evaluate active learning for video object detection in road scenes. Finally, we show that our approach outperforms active learning baselines tested on two datasets. %U https://ieeexplore.ieee.org/document/9022609 %U http://refbase.cvc.uab.es/files/ZGV2019.pdf %U http://dx.doi.org/10.1109/ICCVW.2019.00120 %P 914-923