toggle visibility Search & Display Options

Select All    Deselect All
 |   | 
Details
  Records Links
Author Muhammad Anwer Rao; David Vazquez; Antonio Lopez edit   pdf
doi  isbn
openurl 
  Title Opponent Colors for Human Detection Type Conference Article
  Year 2011 Publication 5th Iberian Conference on Pattern Recognition and Image Analysis Abbreviated Journal  
  Volume 6669 Issue Pages (down) 363-370  
  Keywords Pedestrian Detection; Color; Part Based Models  
  Abstract Human detection is a key component in fields such as advanced driving assistance and video surveillance. However, even detecting non-occluded standing humans remains a challenge of intensive research. Finding good features to build human models for further detection is probably one of the most important issues to face. Currently, shape, texture and motion features have deserve extensive attention in the literature. However, color-based features, which are important in other domains (e.g., image categorization), have received much less attention. In fact, the use of RGB color space has become a kind of choice by default. The focus has been put in developing first and second order features on top of RGB space (e.g., HOG and co-occurrence matrices, resp.). In this paper we evaluate the opponent colors (OPP) space as a biologically inspired alternative for human detection. In particular, by feeding OPP space in the baseline framework of Dalal et al. for human detection (based on RGB, HOG and linear SVM), we will obtain better detection performance than by using RGB space. This is a relevant result since, up to the best of our knowledge, OPP space has not been previously used for human detection. This suggests that in the future it could be worth to compute co-occurrence matrices, self-similarity features, etc., also on top of OPP space, i.e., as we have done with HOG in this paper.  
  Address Las Palmas de Gran Canaria. Spain  
  Corporate Author Thesis  
  Publisher Springer Place of Publication Berlin Heidelberg Editor J. Vitria; J.M. Sanches; M. Hernandez  
  Language English Summary Language English Original Title Opponent Colors for Human Detection  
  Series Editor Series Title Lecture Notes on Computer Science Abbreviated Series Title LNCS  
  Series Volume Series Issue Edition  
  ISSN 0302-9743 ISBN 978-3-642-21256-7 Medium  
  Area Expedition Conference IbPRIA  
  Notes ADAS Approved no  
  Call Number ADAS @ adas @ RVL2011a Serial 1666  
Permanent link to this record
 

 
Author Alejandro Gonzalez Alzate; Gabriel Villalonga; Jiaolong Xu; David Vazquez; Jaume Amores; Antonio Lopez edit   pdf
url  doi
openurl 
  Title Multiview Random Forest of Local Experts Combining RGB and LIDAR data for Pedestrian Detection Type Conference Article
  Year 2015 Publication IEEE Intelligent Vehicles Symposium IV2015 Abbreviated Journal  
  Volume Issue Pages (down) 356-361  
  Keywords Pedestrian Detection  
  Abstract Despite recent significant advances, pedestrian detection continues to be an extremely challenging problem in real scenarios. In order to develop a detector that successfully operates under these conditions, it becomes critical to leverage upon multiple cues, multiple imaging modalities and a strong multi-view classifier that accounts for different pedestrian views and poses. In this paper we provide an extensive evaluation that gives insight into how each of these aspects (multi-cue, multimodality and strong multi-view classifier) affect performance both individually and when integrated together. In the multimodality component we explore the fusion of RGB and depth maps obtained by high-definition LIDAR, a type of modality that is only recently starting to receive attention. As our analysis reveals, although all the aforementioned aspects significantly help in improving the performance, the fusion of visible spectrum and depth information allows to boost the accuracy by a much larger margin. The resulting detector not only ranks among the top best performers in the challenging KITTI benchmark, but it is built upon very simple blocks that are easy to implement and computationally efficient. These simple blocks can be easily replaced with more sophisticated ones recently proposed, such as the use of convolutional neural networks for feature representation, to further improve the accuracy.  
  Address Seoul; Corea; June 2015  
  Corporate Author Thesis  
  Publisher Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN ISBN Medium  
  Area ACDC Expedition Conference IV  
  Notes ADAS; 600.076; 600.057; 600.054 Approved no  
  Call Number ADAS @ adas @ GVX2015 Serial 2625  
Permanent link to this record
 

 
Author Fadi Dornaika; Angel Sappa edit  openurl
  Title Improving Appearance-Based 3D Face Tracking Using Sparse Stereo Data Type Conference Article
  Year 2007 Publication Advances in Computer Graphics and Computer Vision, Abbreviated Journal  
  Volume Issue Pages (down) 354–366  
  Keywords  
  Abstract  
  Address  
  Corporate Author Thesis  
  Publisher Springer Verlag Place of Publication Editor J. Braz, A. Ranchordas, H. Araujo and J. Jorge,  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN ISBN Medium  
  Area Expedition Conference VISAPP  
  Notes ADAS Approved no  
  Call Number ADAS @ adas @ DoS2007d Serial 1046  
Permanent link to this record
 

 
Author Gioacchino Vino; Angel Sappa edit  doi
isbn  openurl
  Title Revisiting Harris Corner Detector Algorithm: a Gradual Thresholding Approach Type Conference Article
  Year 2013 Publication 10th International Conference on Image Analysis and Recognition Abbreviated Journal  
  Volume 7950 Issue Pages (down) 354-363  
  Keywords  
  Abstract This paper presents an adaptive thresholding approach intended to increase the number of detected corners, while reducing the amount of those ones corresponding to noisy data. The proposed approach works by using the classical Harris corner detector algorithm and overcome the difficulty in finding a general threshold that work well for all the images in a given data set by proposing a novel adaptive thresholding scheme. Initially, two thresholds are used to discern between strong corners and flat regions. Then, a region based criteria is used to discriminate between weak corners and noisy points in the midway interval. Experimental results show that the proposed approach has a better capability to reject false corners and, at the same time, to detect weak ones. Comparisons with the state of the art are provided showing the validity of the proposed approach.  
  Address Póvoa de Varzim; Portugal; June 2013  
  Corporate Author Thesis  
  Publisher Springer Berlin Heidelberg Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title LNCS  
  Series Volume Series Issue Edition  
  ISSN 0302-9743 ISBN 978-3-642-39093-7 Medium  
  Area Expedition Conference ICIAR  
  Notes ADAS; 600.055 Approved no  
  Call Number Admin @ si @ ViS2013 Serial 2562  
Permanent link to this record
 

 
Author Patricia Marquez; Debora Gil; Aura Hernandez-Sabate; Daniel Kondermann edit   pdf
url  doi
isbn  openurl
  Title When Is A Confidence Measure Good Enough? Type Conference Article
  Year 2013 Publication 9th International Conference on Computer Vision Systems Abbreviated Journal  
  Volume 7963 Issue Pages (down) 344-353  
  Keywords Optical flow, confidence measure, performance evaluation  
  Abstract Confidence estimation has recently become a hot topic in image processing and computer vision.Yet, several definitions exist of the term “confidence” which are sometimes used interchangeably. This is a position paper, in which we aim to give an overview on existing definitions,
thereby clarifying the meaning of the used terms to facilitate further research in this field. Based on these clarifications, we develop a theory to compare confidence measures with respect to their quality.
 
  Address St Petersburg; Russia; July 2013  
  Corporate Author Thesis  
  Publisher Springer Link Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title LNCS  
  Series Volume Series Issue Edition  
  ISSN 0302-9743 ISBN 978-3-642-39401-0 Medium  
  Area Expedition Conference ICVS  
  Notes IAM;ADAS; 600.044; 600.057; 600.060; 601.145 Approved no  
  Call Number IAM @ iam @ MGH2013a Serial 2218  
Permanent link to this record
 

 
Author Fahad Shahbaz Khan; Muhammad Anwer Rao; Joost Van de Weijer; Michael Felsberg; J.Laaksonen edit  url
doi  isbn
openurl 
  Title Deep semantic pyramids for human attributes and action recognition Type Conference Article
  Year 2015 Publication Image Analysis, Proceedings of 19th Scandinavian Conference , SCIA 2015 Abbreviated Journal  
  Volume 9127 Issue Pages (down) 341-353  
  Keywords Action recognition; Human attributes; Semantic pyramids  
  Abstract Describing persons and their actions is a challenging problem due to variations in pose, scale and viewpoint in real-world images. Recently, semantic pyramids approach [1] for pose normalization has shown to provide excellent results for gender and action recognition. The performance of semantic pyramids approach relies on robust image description and is therefore limited due to the use of shallow local features. In the context of object recognition [2] and object detection [3], convolutional neural networks (CNNs) or deep features have shown to improve the performance over the conventional shallow features.
We propose deep semantic pyramids for human attributes and action recognition. The method works by constructing spatial pyramids based on CNNs of different part locations. These pyramids are then combined to obtain a single semantic representation. We validate our approach on the Berkeley and 27 Human Attributes datasets for attributes classification. For action recognition, we perform experiments on two challenging datasets: Willow and PASCAL VOC 2010. The proposed deep semantic pyramids provide a significant gain of 17.2%, 13.9%, 24.3% and 22.6% compared to the standard shallow semantic pyramids on Berkeley, 27 Human Attributes, Willow and PASCAL VOC 2010 datasets respectively. Our results also show that deep semantic pyramids outperform conventional CNNs based on the full bounding box of the person. Finally, we compare our approach with state-of-the-art methods and show a gain in performance compared to best methods in literature.
 
  Address Denmark; Copenhagen; June 2015  
  Corporate Author Thesis  
  Publisher Springer International Publishing Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN 0302-9743 ISBN 978-3-319-19664-0 Medium  
  Area Expedition Conference SCIA  
  Notes LAMP; 600.068; 600.079;ADAS Approved no  
  Call Number Admin @ si @ KRW2015b Serial 2672  
Permanent link to this record
 

 
Author Alex Goldhoorn; Arnau Ramisa; Ramon Lopez de Mantaras; Ricardo Toledo edit  isbn
openurl 
  Title Using the Average Landmark Vector Method for Robot Homing Type Conference Article
  Year 2007 Publication Artificial Intelligence Research and Development, Proceedings of the 10th International Conference of the ACIA Abbreviated Journal  
  Volume 163 Issue Pages (down) 331–338  
  Keywords  
  Abstract  
  Address  
  Corporate Author Thesis  
  Publisher Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN ISBN 978–1–58603–798–7 Medium  
  Area Expedition Conference CCIA’07  
  Notes RV;ADAS Approved no  
  Call Number Admin @ si @ GRL2007 Serial 899  
Permanent link to this record
 

 
Author Angel Sappa; Niki Aifanti; Sotiris Malassiotis; Michael G. Strintzis edit  openurl
  Title Monocular 3D Human Body Reconstruction Towards Depth Augmentation of Television Sequences Type Conference Article
  Year 2003 Publication IEEE International Conference on Image Processing, Barcelona, Spain, September 2003 Abbreviated Journal  
  Volume Issue Pages (down) 325-328  
  Keywords  
  Abstract  
  Address Barcelona  
  Corporate Author Thesis  
  Publisher Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN ISBN Medium  
  Area Expedition Conference  
  Notes ADAS Approved no  
  Call Number ADAS @ adas @ SAM2003 Serial 418  
Permanent link to this record
 

 
Author Diego Cheda; Daniel Ponsa; Antonio Lopez edit   pdf
url  openurl
  Title Monocular Depth-based Background Estimation Type Conference Article
  Year 2012 Publication 7th International Conference on Computer Vision Theory and Applications Abbreviated Journal  
  Volume Issue Pages (down) 323-328  
  Keywords  
  Abstract In this paper, we address the problem of reconstructing the background of a scene from a video sequence with occluding objects. The images are taken by hand-held cameras. Our method composes the background by selecting the appropriate pixels from previously aligned input images. To do that, we minimize a cost function that penalizes the deviations from the following assumptions: background represents objects whose distance to the camera is maximal, and background objects are stationary. Distance information is roughly obtained by a supervised learning approach that allows us to distinguish between close and distant image regions. Moving foreground objects are filtered out by using stationariness and motion boundary constancy measurements. The cost function is minimized by a graph cuts method. We demonstrate the applicability of our approach to recover an occlusion-free background in a set of sequences.  
  Address Roma  
  Corporate Author Thesis  
  Publisher Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title  
  Series Volume Series Issue Edition  
  ISSN ISBN Medium  
  Area Expedition Conference VISAPP  
  Notes ADAS Approved no  
  Call Number Admin @ si @ CPL2012b; ADAS @ adas @ cpl2012e Serial 2012  
Permanent link to this record
 

 
Author Dennis G.Romero; Anselmo Frizera; Angel Sappa; Boris X. Vintimilla; Teodiano F.Bastos edit   pdf
url  doi
isbn  openurl
  Title A predictive model for human activity recognition by observing actions and context Type Conference Article
  Year 2015 Publication Advanced Concepts for Intelligent Vision Systems, Proceedings of 16th International Conference, ACIVS 2015 Abbreviated Journal  
  Volume 9386 Issue Pages (down) 323-333  
  Keywords  
  Abstract This paper presents a novel model to estimate human activities — a human activity is defined by a set of human actions. The proposed approach is based on the usage of Recurrent Neural Networks (RNN) and Bayesian inference through the continuous monitoring of human actions and its surrounding environment. In the current work human activities are inferred considering not only visual analysis but also additional resources; external sources of information, such as context information, are incorporated to contribute to the activity estimation. The novelty of the proposed approach lies in the way the information is encoded, so that it can be later associated according to a predefined semantic structure. Hence, a pattern representing a given activity can be defined by a set of actions, plus contextual information or other kind of information that could be relevant to describe the activity. Experimental results with real data are provided showing the validity of the proposed approach.  
  Address Catania; Italy; October 2015  
  Corporate Author Thesis  
  Publisher Springer International Publishing Place of Publication Editor  
  Language Summary Language Original Title  
  Series Editor Series Title Abbreviated Series Title LNCS  
  Series Volume Series Issue Edition  
  ISSN 0302-9743 ISBN 978-3-319-25902-4 Medium  
  Area Expedition Conference ACIVS  
  Notes ADAS; 600.076 Approved no  
  Call Number Admin @ si @ RFS2015 Serial 2661  
Permanent link to this record
Select All    Deselect All
 |   | 
Details

Save Citations:
Export Records: