Home | [131–140] << 141 142 143 144 145 146 147 148 149 150 >> [151–160] |
Records | |||||
---|---|---|---|---|---|
Author | Ivan Huerta; Michael Holte; Thomas B. Moeslund; Jordi Gonzalez | ||||
Title | Chromatic shadow detection and tracking for moving foreground segmentation | Type | Journal Article | ||
Year | 2015 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 41 | Issue | Pages | 42-53 | |
Keywords | Detecting moving objects; Chromatic shadow detection; Temporal local gradient; Spatial and Temporal brightness and angle distortions; Shadow tracking | ||||
Abstract | Advanced segmentation techniques in the surveillance domain deal with shadows to avoid distortions when detecting moving objects. Most approaches for shadow detection are still typically restricted to penumbra shadows and cannot cope well with umbra shadows. Consequently, umbra shadow regions are usually detected as part of moving objects, thus aecting the performance of the nal detection. In this paper we address the detection of both penumbra and umbra shadow regions. First, a novel bottom-up approach is presented based on gradient and colour models, which successfully discriminates between chromatic moving cast shadow regions and those regions detected as moving objects. In essence, those regions corresponding to potential shadows are detected based on edge partitioning and colour statistics. Subsequently (i) temporal similarities between textures and (ii) spatial similarities between chrominance angle and brightness distortions are analysed for each potential shadow region for detecting the umbra shadow regions. Our second contribution renes even further the segmentation results: a tracking-based top-down approach increases the performance of our bottom-up chromatic shadow detection algorithm by properly correcting non-detected shadows.
To do so, a combination of motion lters in a data association framework exploits the temporal consistency between objects and shadows to increase the shadow detection rate. Experimental results exceed current state-of-the- art in shadow accuracy for multiple well-known surveillance image databases which contain dierent shadowed materials and illumination conditions. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ISE; 600.078; 600.063 | Approved | no | ||
Call Number | Admin @ si @ HHM2015 | Serial | 2703 | ||
Permanent link to this record | |||||
Author | Antonio Lopez; Gabriel Villalonga; Laura Sellart; German Ros; David Vazquez; Jiaolong Xu; Javier Marin; Azadeh S. Mozafari | ||||
Title | Training my car to see using virtual worlds | Type | Journal Article | ||
Year | 2017 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 38 | Issue | Pages | 102-118 | |
Keywords | |||||
Abstract | Computer vision technologies are at the core of different advanced driver assistance systems (ADAS) and will play a key role in oncoming autonomous vehicles too. One of the main challenges for such technologies is to perceive the driving environment, i.e. to detect and track relevant driving information in a reliable manner (e.g. pedestrians in the vehicle route, free space to drive through). Nowadays it is clear that machine learning techniques are essential for developing such a visual perception for driving. In particular, the standard working pipeline consists of collecting data (i.e. on-board images), manually annotating the data (e.g. drawing bounding boxes around pedestrians), learning a discriminative data representation taking advantage of such annotations (e.g. a deformable part-based model, a deep convolutional neural network), and then assessing the reliability of such representation with the acquired data. In the last two decades most of the research efforts focused on representation learning (first, designing descriptors and learning classifiers; later doing it end-to-end). Hence, collecting data and, especially, annotating it, is essential for learning good representations. While this has been the case from the very beginning, only after the disruptive appearance of deep convolutional neural networks that it became a serious issue due to their data hungry nature. In this context, the problem is that manual data annotation is a tiresome work prone to errors. Accordingly, in the late 00’s we initiated a research line consisting of training visual models using photo-realistic computer graphics, especially focusing on assisted and autonomous driving. In this paper, we summarize such a work and show how it has become a new tendency with increasing acceptance. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS; 600.118 | Approved | no | ||
Call Number | Admin @ si @ LVS2017 | Serial | 2985 | ||
Permanent link to this record | |||||
Author | Pau Rodriguez; Miguel Angel Bautista; Sergio Escalera; Jordi Gonzalez | ||||
Title | Beyond Oneshot Encoding: lower dimensional target embedding | Type | Journal Article | ||
Year | 2018 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 75 | Issue | Pages | 21-31 | |
Keywords | Error correcting output codes; Output embeddings; Deep learning; Computer vision | ||||
Abstract | Target encoding plays a central role when learning Convolutional Neural Networks. In this realm, one-hot encoding is the most prevalent strategy due to its simplicity. However, this so widespread encoding schema assumes a flat label space, thus ignoring rich relationships existing among labels that can be exploited during training. In large-scale datasets, data does not span the full label space, but instead lies in a low-dimensional output manifold. Following this observation, we embed the targets into a low-dimensional space, drastically improving convergence speed while preserving accuracy. Our contribution is two fold: (i) We show that random projections of the label space are a valid tool to find such lower dimensional embeddings, boosting dramatically convergence rates at zero computational cost; and (ii) we propose a normalized eigenrepresentation of the class manifold that encodes the targets with minimal information loss, improving the accuracy of random projections encoding while enjoying the same convergence rates. Experiments on CIFAR-100, CUB200-2011, Imagenet, and MIT Places demonstrate that the proposed approach drastically improves convergence speed while reaching very competitive accuracy rates. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ISE; HuPBA; 600.098; 602.133; 602.121; 600.119 | Approved | no | ||
Call Number | Admin @ si @ RBE2018 | Serial | 3120 | ||
Permanent link to this record | |||||
Author | Julio C. S. Jacques Junior; Xavier Baro; Sergio Escalera | ||||
Title | Exploiting feature representations through similarity learning, post-ranking and ranking aggregation for person re-identification | Type | Journal Article | ||
Year | 2018 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 79 | Issue | Pages | 76-85 | |
Keywords | |||||
Abstract | Person re-identification has received special attention by the human analysis community in the last few years. To address the challenges in this field, many researchers have proposed different strategies, which basically exploit either cross-view invariant features or cross-view robust metrics. In this work, we propose to exploit a post-ranking approach and combine different feature representations through ranking aggregation. Spatial information, which potentially benefits the person matching, is represented using a 2D body model, from which color and texture information are extracted and combined. We also consider background/foreground information, automatically extracted via Deep Decompositional Network, and the usage of Convolutional Neural Network (CNN) features. To describe the matching between images we use the polynomial feature map, also taking into account local and global information. The Discriminant Context Information Analysis based post-ranking approach is used to improve initial ranking lists. Finally, the Stuart ranking aggregation method is employed to combine complementary ranking lists obtained from different feature representations. Experimental results demonstrated that we improve the state-of-the-art on VIPeR and PRID450s datasets, achieving 67.21% and 75.64% on top-1 rank recognition rate, respectively, as well as obtaining competitive results on CUHK01 dataset. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | HuPBA; 602.143 | Approved | no | ||
Call Number | Admin @ si @ JBE2018 | Serial | 3138 | ||
Permanent link to this record | |||||
Author | Meysam Madadi; Sergio Escalera; Alex Carruesco Llorens; Carlos Andujar; Xavier Baro; Jordi Gonzalez | ||||
Title | Top-down model fitting for hand pose recovery in sequences of depth images | Type | Journal Article | ||
Year | 2018 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 79 | Issue | Pages | 63-75 | |
Keywords | |||||
Abstract | State-of-the-art approaches on hand pose estimation from depth images have reported promising results under quite controlled considerations. In this paper we propose a two-step pipeline for recovering the hand pose from a sequence of depth images. The pipeline has been designed to deal with images taken from any viewpoint and exhibiting a high degree of finger occlusion. In a first step we initialize the hand pose using a part-based model, fitting a set of hand components in the depth images. In a second step we consider temporal data and estimate the parameters of a trained bilinear model consisting of shape and trajectory bases. We evaluate our approach on a new created synthetic hand dataset along with NYU and MSRA real datasets. Results demonstrate that the proposed method outperforms the most recent pose recovering approaches, including those based on CNNs. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | HUPBA; 600.098 | Approved | no | ||
Call Number | Admin @ si @ MEC2018 | Serial | 3203 | ||
Permanent link to this record | |||||
Author | Jorge Charco; Angel Sappa; Boris X. Vintimilla; Henry Velesaca | ||||
Title | Camera pose estimation in multi-view environments: From virtual scenarios to the real world | Type | Journal Article | ||
Year | 2021 | Publication | Image and Vision Computing | Abbreviated Journal | IVC |
Volume | 110 | Issue | Pages | 104182 | |
Keywords | |||||
Abstract | This paper presents a domain adaptation strategy to efficiently train network architectures for estimating the relative camera pose in multi-view scenarios. The network architectures are fed by a pair of simultaneously acquired images, hence in order to improve the accuracy of the solutions, and due to the lack of large datasets with pairs of overlapped images, a domain adaptation strategy is proposed. The domain adaptation strategy consists on transferring the knowledge learned from synthetic images to real-world scenarios. For this, the networks are firstly trained using pairs of synthetic images, which are captured at the same time by a pair of cameras in a virtual environment; and then, the learned weights of the networks are transferred to the real-world case, where the networks are retrained with a few real images. Different virtual 3D scenarios are generated to evaluate the relationship between the accuracy on the result and the similarity between virtual and real scenarios—similarity on both geometry of the objects contained in the scene as well as relative pose between camera and objects in the scene. Experimental results and comparisons are provided showing that the accuracy of all the evaluated networks for estimating the camera pose improves when the proposed domain adaptation strategy is used, highlighting the importance on the similarity between virtual-real scenarios. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | MSIAU; 600.130; 600.122 | Approved | no | ||
Call Number | Admin @ si @ CSV2021 | Serial | 3577 | ||
Permanent link to this record | |||||
Author | Aitor Alvarez-Gila; Adrian Galdran; Estibaliz Garrote; Joost Van de Weijer | ||||
Title | Self-supervised blur detection from synthetically blurred scenes | Type | Journal Article | ||
Year | 2019 | Publication | Image and Vision Computing | Abbreviated Journal | IMAVIS |
Volume | 92 | Issue | Pages | 103804 | |
Keywords | |||||
Abstract | Blur detection aims at segmenting the blurred areas of a given image. Recent deep learning-based methods approach this problem by learning an end-to-end mapping between the blurred input and a binary mask representing the localization of its blurred areas. Nevertheless, the effectiveness of such deep models is limited due to the scarcity of datasets annotated in terms of blur segmentation, as blur annotation is labor intensive. In this work, we bypass the need for such annotated datasets for end-to-end learning, and instead rely on object proposals and a model for blur generation in order to produce a dataset of synthetically blurred images. This allows us to perform self-supervised learning over the generated image and ground truth blur mask pairs using CNNs, defining a framework that can be employed in purely self-supervised, weakly supervised or semi-supervised configurations. Interestingly, experimental results of such setups over the largest blur segmentation datasets available show that this approach achieves state of the art results in blur segmentation, even without ever observing any real blurred image. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | LAMP; 600.109; 600.120 | Approved | no | ||
Call Number | Admin @ si @ AGG2019 | Serial | 3301 | ||
Permanent link to this record | |||||
Author | Maria Vanrell; Jordi Vitria | ||||
Title | Optimal 3x3 decomposable disks for morphological transformations | Type | Journal | ||
Year | 1997 | Publication | Image and Vision Computing, 15(2): 845–854 | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | OR;CIC;MV | Approved | no | ||
Call Number | BCNPCL @ bcnpcl @ VaV1997c | Serial | 543 | ||
Permanent link to this record | |||||
Author | Fadi Dornaika; J. Ahlberg | ||||
Title | Fitting 3D face models for tracking and active appearance model training | Type | Journal | ||
Year | 2006 | Publication | Image and Vision Computing, 24(9): 1010–1024 | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | Approved | no | |||
Call Number | Admin @ si @ DoA2006 | Serial | 733 | ||
Permanent link to this record | |||||
Author | Joan Serrat; Ferran Diego; Jose Manuel Alvarez; Felipe Lumbreras | ||||
Title | Alignment of Videos Recorded from Moving Vehicles | Type | Conference Article | ||
Year | 2007 | Publication | in 14th International Conference on Image Analysis and Processing, | Abbreviated Journal | |
Volume | Issue | Pages | 512–517 | ||
Keywords | |||||
Abstract | |||||
Address | Modena (Italia) | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ SDA2007 | Serial | 879 | ||
Permanent link to this record | |||||
Author | M. Bressan; Jordi Vitria | ||||
Title | Independent Modes of Variation in Point Distribution Models | Type | Miscellaneous | ||
Year | 2001 | Publication | In C. Arcelli, L.P. Cordella, G. Sanniti di Baja (Eds.): Visual Form 2001 4tth International Workshop on Visual Visual Form 2001 4tth International Workshop on Visual Form, IWVF4, Proceedings, LNCS 2059, Springer Verlag, 123 | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | Capri, Italia | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | OR;MV | Approved | no | ||
Call Number | BCNPCL @ bcnpcl @ BVi2001 | Serial | 80 | ||
Permanent link to this record | |||||
Author | David Masip; Jordi Vitria | ||||
Title | On the Nearest Neighbor Approach for Gender Recognition | Type | Miscellaneous | ||
Year | 2003 | Publication | In I. Aguilo, Ll. Valverde M.T. Escrig, editors. Artificial Intelligence Research and Development. IOS PRESS pp.178–188 | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | Amsterdam | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | OR;MV | Approved | no | ||
Call Number | BCNPCL @ bcnpcl @ MaV2003b | Serial | 387 | ||
Permanent link to this record | |||||
Author | Antonio Hernandez; Carlos Primo; Sergio Escalera | ||||
Title | Automatic user interaction correction via Multi-label Graph cuts | Type | Conference Article | ||
Year | 2011 | Publication | In ICCV 2011 1st IEEE International Workshop on Human Interaction in Computer Vision HICV | Abbreviated Journal | |
Volume | Issue | Pages | 1276-1281 | ||
Keywords | |||||
Abstract | Most applications in image segmentation requires from user interaction in order to achieve accurate results. However, user wants to achieve the desired segmentation accuracy reducing effort of manual labelling. In this work, we extend standard multi-label α-expansion Graph Cut algorithm so that it analyzes the interaction of the user in order to modify the object model and improve final segmentation of objects. The approach is inspired in the fact that fast user interactions may introduce some pixel errors confusing object and background. Our results with different degrees of user interaction and input errors show high performance of the proposed approach on a multi-label human limb segmentation problem compared with classical α-expansion algorithm. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-1-4673-0062-9 | Medium | ||
Area | Expedition | Conference | HICV | ||
Notes | MILAB; HuPBA | Approved | no | ||
Call Number | Admin @ si @ HPE2011 | Serial | 1892 | ||
Permanent link to this record | |||||
Author | G.D. Evangelidis; Ferran Diego; Joan Serrat; Antonio Lopez | ||||
Title | Slice Matching for Accurate Spatio-Temporal Alignment | Type | Conference Article | ||
Year | 2011 | Publication | In ICCV Workshop on Visual Surveillance | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | video alignment | ||||
Abstract | Video synchronization and alignment is a rather recent topic in computer vision. It usually deals with the problem of aligning sequences recorded simultaneously by static, jointly- or independently-moving cameras. In this paper, we investigate the more difficult problem of matching videos captured at different times from independently-moving cameras, whose trajectories are approximately coincident or parallel. To this end, we propose a novel method that pixel-wise aligns videos and allows thus to automatically highlight their differences. This primarily aims at visual surveillance but the method can be adopted as is by other related video applications, like object transfer (augmented reality) or high dynamic range video. We build upon a slice matching scheme to first synchronize the sequences, while we develop a spatio-temporal alignment scheme to spatially register corresponding frames and refine the temporal mapping. We investigate the performance of the proposed method on videos recorded from vehicles driven along different types of roads and compare with related previous works. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | VS | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ EDS2011; ADAS @ adas @ eds2011a | Serial | 1861 | ||
Permanent link to this record | |||||
Author | Francesco Ciompi; A. Palaioroutas; M. Loeve; Oriol Pujol; Petia Radeva; H. Tiddens; M. de Bruijne | ||||
Title | Lung Tissue Classification in Severe Advanced Cystic Fibrosis from CT Scans | Type | Conference Article | ||
Year | 2011 | Publication | In MICCAI 2011 4th International Workshop on Pulmonary Image Analysis | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | Toronto, Canada | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | PIA | ||
Notes | MILAB;HuPBA | Approved | no | ||
Call Number | Admin @ si @ CPL2011 | Serial | 1798 | ||
Permanent link to this record |