|
Javier Varona, Jordi Gonzalez, Xavier Roca, & Juan J. Villanueva. (2003). Appearance Tracking for Video Surveillance.
|
|
|
Gemma Sanchez, Ernest Valveny, Josep Llados, Joan Mas, & N. Lozano. (2004). A platform to extract knowledge from graphic documents. Application to an architectural sketch understanding scenario.
|
|
|
Philippe Dosch, & Josep Llados. (2004). Vectorial Signatures for Symbol Discrimination.
|
|
|
Gemma Sanchez, & Josep Llados. (2004). Syntactic models to represent perceptually regular repetitive patterns in graphic documents.
|
|
|
E. Barakova, Maya Dimitrova, T. Lorents, & Petia Radeva. (2004). The Web as an “Autobiographical Agent”.
|
|
|
Ernest Valveny, & Philippe Dosch. (2004). Symbol Recognition Contest: A Synthesis.
|
|
|
Ernest Valveny, & Philippe Dosch. (2004). Performance Evaluation of Symbol Recognition. In A. D.(E.) S. Marinai (Ed.), Document Analysis Systems (Vol. 3163, 354–365).
|
|
|
Oriol Ramos Terrades, & Ernest Valveny. (2004). Indexing Technical Symbols Using Ridgelets Transform.
|
|
|
Josep Llados, & Young-Bin Kwon. (2004). Graphics Recognition. Recent Advances and Perspectives.
|
|
|
Franck Davoine, & Fadi Dornaika. (2005). Head and facial animation tracking using appearance-adaptive models and particle filters. In V. Pavlovic and T.S. Huang (editors), Real–Time Vision for Human–Computer Interaction.
|
|
|
Fadi Dornaika, & Angel Sappa. (2007). SFM for Planar Scenes: a Direct and Robust Approach. In book chapter: Informatics in Control, Automation and Robotics II, Ed. J. Filipe, J. Ferrier, J. Cetto and M. Carvalho, pp. 129–136. (best papers ICINCO 2005).
|
|
|
Misael Rosales, Petia Radeva, J. Mauri, & Oriol Pujol. (2004). Simulation Model of Intravascular Ultrasound Images.
|
|
|
Josep Llados. (2007). Advances in Graphics Recognition. In Digital Document Processing, Major Directions and Recent Advances, Advances in Pattern Recognition, B.B. Chaudhuri, ed., 281–304.
|
|
|
David Vazquez, Antonio Lopez, Daniel Ponsa, & David Geronimo. (2013). Interactive Training of Human Detectors. In Multiodal Interaction in Image and Video Applications (Vol. 48, pp. 169–182). Springer Berlin Heidelberg.
Abstract: Image based human detection remains as a challenging problem. Most promising detectors rely on classifiers trained with labelled samples. However, labelling is a manual labor intensive step. To overcome this problem we propose to collect images of pedestrians from a virtual city, i.e., with automatic labels, and train a pedestrian detector with them, which works fine when such virtual-world data are similar to testing one, i.e., real-world pedestrians in urban areas. When testing data is acquired in different conditions than training one, e.g., human detection in personal photo albums, dataset shift appears. In previous work, we cast this problem as one of domain adaptation and solve it with an active learning procedure. In this work, we focus on the same problem but evaluating a different set of faster to compute features, i.e., Haar, EOH and their combination. In particular, we train a classifier with virtual-world data, using such features and Real AdaBoost as learning machine. This classifier is applied to real-world training images. Then, a human oracle interactively corrects the wrong detections, i.e., few miss detections are manually annotated and some false ones are pointed out too. A low amount of manual annotation is fixed as restriction. Real- and virtual-world difficult samples are combined within what we call cool world and we retrain the classifier with this data. Our experiments show that this adapted classifier is equivalent to the one trained with only real-world data but requiring 90% less manual annotations.
Keywords: Pedestrian Detection; Virtual World; AdaBoost; Domain Adaptation
|
|
|
Murad Al Haj, Andrew Bagdanov, Jordi Gonzalez, & Xavier Roca. (2009). Robust and Efficient Multipose Face Detection Using Skin Color Segmentation. In 4th Iberian Conference on Pattern Recognition and Image Analysis (Vol. 5524). LNCS. Springer Berlin Heidelberg.
Abstract: In this paper we describe an efficient technique for detecting faces in arbitrary images and video sequences. The approach is based on segmentation of images or video frames into skin-colored blobs using a pixel-based heuristic. Scale and translation invariant features are then computed from these segmented blobs which are used to perform statistical discrimination between face and non-face classes. We train and evaluate our method on a standard, publicly available database of face images and analyze its performance over a range of statistical pattern classifiers. The generalization of our approach is illustrated by testing on an independent sequence of frames containing many faces and non-faces. These experiments indicate that our proposed approach obtains false positive rates comparable to more complex, state-of-the-art techniques, and that it generalizes better to new data. Furthermore, the use of skin blobs and invariant features requires fewer training samples since significantly fewer non-face candidate regions must be considered when compared to AdaBoost-based approaches.
|
|