Records |
Author |
Arjan Gijsenij; Theo Gevers; Joost Van de Weijer |
Title |
Physics-based Edge Evaluation for Improved Color Constancy |
Type |
Conference Article |
Year |
2009 |
Publication |
22nd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
581 – 588 |
Keywords |
|
Abstract |
Edge-based color constancy makes use of image derivatives to estimate the illuminant. However, different edge types exist in real-world images such as shadow, geometry, material and highlight edges. These different edge types may have a distinctive influence on the performance of the illuminant estimation. |
Address |
Miami, USA |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-3992-8 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
CAT;ISE |
Approved |
no |
Call Number |
CAT @ cat @ GGW2009 |
Serial |
1197 |
Permanent link to this record |
|
|
|
Author |
Sergio Escalera; R. M. Martinez; Jordi Vitria; Petia Radeva; Maria Teresa Anguera |
Title |
Dominance Detection in Face-to-face Conversations |
Type |
Conference Article |
Year |
2009 |
Publication |
2nd IEEE Workshop on CVPR for Human communicative Behavior analysis |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
97–102 |
Keywords |
|
Abstract |
Dominance is referred to the level of influence a person has in a conversation. Dominance is an important research area in social psychology, but the problem of its automatic estimation is a very recent topic in the contexts of social and wearable computing. In this paper, we focus on dominance detection from visual cues. We estimate the correlation among observers by categorizing the dominant people in a set of face-to-face conversations. Different dominance indicators from gestural communication are defined, manually annotated, and compared to the observers opinion. Moreover, the considered indicators are automatically extracted from video sequences and learnt by using binary classifiers. Results from the three analysis shows a high correlation and allows the categorization of dominant people in public discussion video sequences. |
Address |
Miami, USA |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
2160-7508 |
ISBN |
978-1-4244-3994-2 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
HuPBA; OR; MILAB;MV |
Approved |
no |
Call Number |
BCNPCL @ bcnpcl @ EMV2009 |
Serial |
1227 |
Permanent link to this record |
|
|
|
Author |
Anjan Dutta; Zeynep Akata |
Title |
Semantically Tied Paired Cycle Consistency for Zero-Shot Sketch-based Image Retrieval |
Type |
Conference Article |
Year |
2019 |
Publication |
32nd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
5089-5098 |
Keywords |
|
Abstract |
Zero-shot sketch-based image retrieval (SBIR) is an emerging task in computer vision, allowing to retrieve natural images relevant to sketch queries that might not been seen in the training phase. Existing works either require aligned sketch-image pairs or inefficient memory fusion layer for mapping the visual information to a semantic space. In this work, we propose a semantically aligned paired cycle-consistent generative (SEM-PCYC) model for zero-shot SBIR, where each branch maps the visual information to a common semantic space via an adversarial training. Each of these branches maintains a cycle consistency that only requires supervision at category levels, and avoids the need of highly-priced aligned sketch-image pairs. A classification criteria on the generators' outputs ensures the visual to semantic space mapping to be discriminating. Furthermore, we propose to combine textual and hierarchical side information via a feature selection auto-encoder that selects discriminating side information within a same end-to-end model. Our results demonstrate a significant boost in zero-shot SBIR performance over the state-of-the-art on the challenging Sketchy and TU-Berlin datasets. |
Address |
Long beach; California; USA; June 2019 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
DAG; 600.141; 600.121 |
Approved |
no |
Call Number |
Admin @ si @ DuA2019 |
Serial |
3268 |
Permanent link to this record |
|
|
|
Author |
Mario Rojas; David Masip; A. Todorov; Jordi Vitria |
Title |
Automatic Point-based Facial Trait Judgments Evaluation |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
2715–2720 |
Keywords |
|
Abstract |
Humans constantly evaluate the personalities of other people using their faces. Facial trait judgments have been studied in the psychological field, and have been determined to influence important social outcomes of our lives, such as elections outcomes and social relationships. Recent work on textual descriptions of faces has shown that trait judgments are highly correlated. Further, behavioral studies suggest that two orthogonal dimensions, valence and dominance, can describe the basis of the human judgments from faces. In this paper, we used a corpus of behavioral data of judgments on different trait dimensions to automatically learn a trait predictor from facial pixel images. We study whether trait evaluations performed by humans can be learned using machine learning classifiers, and used later in automatic evaluations of new facial images. The experiments performed using local point-based descriptors show promising results in the evaluation of the main traits. |
Address |
San Francisco CA, USA |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
OR;MV |
Approved |
no |
Call Number |
BCNPCL @ bcnpcl @ RMT2010 |
Serial |
1282 |
Permanent link to this record |
|
|
|
Author |
Josep M. Gonfaus; Xavier Boix; Joost Van de Weijer; Andrew Bagdanov; Joan Serrat; Jordi Gonzalez |
Title |
Harmony Potentials for Joint Classification and Segmentation |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
3280–3287 |
Keywords |
|
Abstract |
Hierarchical conditional random fields have been successfully applied to object segmentation. One reason is their ability to incorporate contextual information at different scales. However, these models do not allow multiple labels to be assigned to a single node. At higher scales in the image, this yields an oversimplified model, since multiple classes can be reasonable expected to appear within one region. This simplified model especially limits the impact that observations at larger scales may have on the CRF model. Neglecting the information at larger scales is undesirable since class-label estimates based on these scales are more reliable than at smaller, noisier scales. To address this problem, we propose a new potential, called harmony potential, which can encode any possible combination of class labels. We propose an effective sampling strategy that renders tractable the underlying optimization problem. Results show that our approach obtains state-of-the-art results on two challenging datasets: Pascal VOC 2009 and MSRC-21. |
Address |
San Francisco CA, USA |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS;CIC;ISE |
Approved |
no |
Call Number |
ADAS @ adas @ GBW2010 |
Serial |
1296 |
Permanent link to this record |
|
|
|
Author |
Jose Manuel Alvarez; Theo Gevers; Antonio Lopez |
Title |
3D Scene Priors for Road Detection |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
57–64 |
Keywords |
road detection |
Abstract |
Vision-based road detection is important in different areas of computer vision such as autonomous driving, car collision warning and pedestrian crossing detection. However, current vision-based road detection methods are usually based on low-level features and they assume structured roads, road homogeneity, and uniform lighting conditions. Therefore, in this paper, contextual 3D information is used in addition to low-level cues. Low-level photometric invariant cues are derived from the appearance of roads. Contextual cues used include horizon lines, vanishing points, 3D scene layout and 3D road stages. Moreover, temporal road cues are included. All these cues are sensitive to different imaging conditions and hence are considered as weak cues. Therefore, they are combined to improve the overall performance of the algorithm. To this end, the low-level, contextual and temporal cues are combined in a Bayesian framework to classify road sequences. Large scale experiments on road sequences show that the road detection method is robust to varying imaging conditions, road types, and scenarios (tunnels, urban and highway). Further, using the combined cues outperforms all other individual cues. Finally, the proposed method provides highest road detection accuracy when compared to state-of-the-art methods. |
Address |
San Francisco; CA; USA; June 2010 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS;ISE |
Approved |
no |
Call Number |
ADAS @ adas @ AGL2010a |
Serial |
1302 |
Permanent link to this record |
|
|
|
Author |
Mohammad Rouhani; Angel Sappa |
Title |
Relaxing the 3L Algorithm for an Accurate Implicit Polynomial Fitting |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
3066-3072 |
Keywords |
|
Abstract |
This paper presents a novel method to increase the accuracy of linear fitting of implicit polynomials. The proposed method is based on the 3L algorithm philosophy. The novelty lies on the relaxation of the additional constraints, already imposed by the 3L algorithm. Hence, the accuracy of the final solution is increased due to the proper adjustment of the expected values in the aforementioned additional constraints. Although iterative, the proposed approach solves the fitting problem within a linear framework, which is independent of the threshold tuning. Experimental results, both in 2D and 3D, showing improvements in the accuracy of the fitting are presented. Comparisons with both state of the art algorithms and a geometric based one (non-linear fitting), which is used as a ground truth, are provided. |
Address |
San Francisco; CA; USA; June 2010 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS |
Approved |
no |
Call Number |
ADAS @ adas @ RoS2010a |
Serial |
1303 |
Permanent link to this record |
|
|
|
Author |
Javier Marin; David Vazquez; David Geronimo; Antonio Lopez |
Title |
Learning Appearance in Virtual Scenarios for Pedestrian Detection |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
137–144 |
Keywords |
Pedestrian Detection; Domain Adaptation |
Abstract |
Detecting pedestrians in images is a key functionality to avoid vehicle-to-pedestrian collisions. The most promising detectors rely on appearance-based pedestrian classifiers trained with labelled samples. This paper addresses the following question: can a pedestrian appearance model learnt in virtual scenarios work successfully for pedestrian detection in real images? (Fig. 1). Our experiments suggest a positive answer, which is a new and relevant conclusion for research in pedestrian detection. More specifically, we record training sequences in virtual scenarios and then appearance-based pedestrian classifiers are learnt using HOG and linear SVM. We test such classifiers in a publicly available dataset provided by Daimler AG for pedestrian detection benchmarking. This dataset contains real world images acquired from a moving car. The obtained result is compared with the one given by a classifier learnt using samples coming from real images. The comparison reveals that, although virtual samples were not specially selected, both virtual and real based training give rise to classifiers of similar performance. |
Address |
San Francisco; CA; USA; June 2010 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
English |
Summary Language |
English |
Original Title |
Learning Appearance in Virtual Scenarios for Pedestrian Detection |
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS |
Approved |
no |
Call Number |
ADAS @ adas @ MVG2010 |
Serial |
1304 |
Permanent link to this record |
|
|
|
Author |
David Aldavert; Arnau Ramisa; Ramon Lopez de Mantaras; Ricardo Toledo |
Title |
Fast and Robust Object Segmentation with the Integral Linear Classifier |
Type |
Conference Article |
Year |
2010 |
Publication |
23rd IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
1046–1053 |
Keywords |
|
Abstract |
We propose an efficient method, built on the popular Bag of Features approach, that obtains robust multiclass pixel-level object segmentation of an image in less than 500ms, with results comparable or better than most state of the art methods. We introduce the Integral Linear Classifier (ILC), that can readily obtain the classification score for any image sub-window with only 6 additions and 1 product by fusing the accumulation and classification steps in a single operation. In order to design a method as efficient as possible, our building blocks are carefully selected from the quickest in the state of the art. More precisely, we evaluate the performance of three popular local descriptors, that can be very efficiently computed using integral images, and two fast quantization methods: the Hierarchical K-Means, and the Extremely Randomized Forest. Finally, we explore the utility of adding spatial bins to the Bag of Features histograms and that of cascade classifiers to improve the obtained segmentation. Our method is compared to the state of the art in the difficult Graz-02 and PASCAL 2007 Segmentation Challenge datasets. |
Address |
San Francisco; CA; USA; June 2010 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4244-6984-0 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS |
Approved |
no |
Call Number |
Admin @ si @ ARL2010a |
Serial |
1311 |
Permanent link to this record |
|
|
|
Author |
Ivo Everts; Jan van Gemert; Theo Gevers |
Title |
Evaluation of Color STIPs for Human Action Recognition |
Type |
Conference Article |
Year |
2013 |
Publication |
IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
2850-2857 |
Keywords |
|
Abstract |
This paper is concerned with recognizing realistic human actions in videos based on spatio-temporal interest points (STIPs). Existing STIP-based action recognition approaches operate on intensity representations of the image data. Because of this, these approaches are sensitive to disturbing photometric phenomena such as highlights and shadows. Moreover, valuable information is neglected by discarding chromaticity from the photometric representation. These issues are addressed by Color STIPs. Color STIPs are multi-channel reformulations of existing intensity-based STIP detectors and descriptors, for which we consider a number of chromatic representations derived from the opponent color space. This enhanced modeling of appearance improves the quality of subsequent STIP detection and description. Color STIPs are shown to substantially outperform their intensity-based counterparts on the challenging UCF~sports, UCF11 and UCF50 action recognition benchmarks. Moreover, the results show that color STIPs are currently the single best low-level feature choice for STIP-based approaches to human action recognition. |
Address |
Portland; oregon; June 2013 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ALTRES;ISE |
Approved |
no |
Call Number |
Admin @ si @ EGG2013 |
Serial |
2364 |
Permanent link to this record |
|
|
|
Author |
Naila Murray; Maria Vanrell; Xavier Otazu; C. Alejandro Parraga |
Title |
Saliency Estimation Using a Non-Parametric Low-Level Vision Model |
Type |
Conference Article |
Year |
2011 |
Publication |
IEEE conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
433-440 |
Keywords |
Gaussian mixture model;ad hoc parameter selection;center-surround inhibition windows;center-surround mechanism;color appearance model;convolution;eye-fixation data;human vision;innate spatial pooling mechanism;inverse wavelet transform;low-level visual front-end;nonparametric low-level vision model;saliency estimation;saliency map;scale integration;scale-weighted center-surround response;scale-weighting function;visual task;Gaussian processes;biology;biology computing;colour vision;computer vision;visual perception;wavelet transforms |
Abstract |
Many successful models for predicting attention in a scene involve three main steps: convolution with a set of filters, a center-surround mechanism and spatial pooling to construct a saliency map. However, integrating spatial information and justifying the choice of various parameter values remain open problems. In this paper we show that an efficient model of color appearance in human vision, which contains a principled selection of parameters as well as an innate spatial pooling mechanism, can be generalized to obtain a saliency model that outperforms state-of-the-art models. Scale integration is achieved by an inverse wavelet transform over the set of scale-weighted center-surround responses. The scale-weighting function (termed ECSF) has been optimized to better replicate psychophysical data on color appearance, and the appropriate sizes of the center-surround inhibition windows have been determined by training a Gaussian Mixture Model on eye-fixation data, thus avoiding ad-hoc parameter selection. Additionally, we conclude that the extension of a color appearance model to saliency estimation adds to the evidence for a common low-level visual front-end for different visual tasks. |
Address |
Colorado Springs |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4577-0394-2 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
CIC |
Approved |
no |
Call Number |
Admin @ si @ MVO2011 |
Serial |
1757 |
Permanent link to this record |
|
|
|
Author |
Marco Pedersoli; Andrea Vedaldi; Jordi Gonzalez |
Title |
A Coarse-to-fine Approach for fast Deformable Object Detection |
Type |
Conference Article |
Year |
2011 |
Publication |
IEEE conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
1353-1360 |
Keywords |
|
Abstract |
|
Address |
Colorado Springs; USA |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ISE |
Approved |
no |
Call Number |
Admin @ si @ PVG2011 |
Serial |
1764 |
Permanent link to this record |
|
|
|
Author |
Miguel Oliveira; Angel Sappa; V.Santos |
Title |
Unsupervised Local Color Correction for Coarsely Registered Images |
Type |
Conference Article |
Year |
2011 |
Publication |
IEEE conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
201-208 |
Keywords |
|
Abstract |
The current paper proposes a new parametric local color correction technique. Initially, several color transfer functions are computed from the output of the mean shift color segmentation algorithm. Secondly, color influence maps are calculated. Finally, the contribution of every color transfer function is merged using the weights from the color influence maps. The proposed approach is compared with both global and local color correction approaches. Results show that our method outperforms the technique ranked first in a recent performance evaluation on this topic. Moreover, the proposed approach is computed in about one tenth of the time. |
Address |
Colorado Springs |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4577-0394-2 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
ADAS |
Approved |
no |
Call Number |
Admin @ si @ OSS2011; ADAS @ adas @ |
Serial |
1766 |
Permanent link to this record |
|
|
|
Author |
Albert Gordo; Florent Perronnin |
Title |
Asymmetric Distances for Binary Embeddings |
Type |
Conference Article |
Year |
2011 |
Publication |
IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
729 - 736 |
Keywords |
|
Abstract |
In large-scale query-by-example retrieval, embedding image signatures in a binary space offers two benefits: data compression and search efficiency. While most embedding algorithms binarize both query and database signatures, it has been noted that this is not strictly a requirement. Indeed, asymmetric schemes which binarize the database signatures but not the query still enjoy the same two benefits but may provide superior accuracy. In this work, we propose two general asymmetric distances which are applicable to a wide variety of embedding techniques including Locality Sensitive Hashing (LSH), Locality Sensitive Binary Codes (LSBC), Spectral Hashing (SH) and Semi-Supervised Hashing (SSH). We experiment on four public benchmarks containing up to 1M images and show that the proposed asymmetric distances consistently lead to large improvements over the symmetric Hamming distance for all binary embedding techniques. We also propose a novel simple binary embedding technique – PCA Embedding (PCAE) – which is shown to yield competitive results with respect to more complex algorithms such as SH and SSH. |
Address |
Providence, RI |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
978-1-4577-0394-2 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
DAG |
Approved |
no |
Call Number |
Admin @ si @ GoP2011; IAM @ iam @ GoP2011 |
Serial |
1817 |
Permanent link to this record |
|
|
|
Author |
Antonio Hernandez; Nadezhda Zlateva; Alexander Marinov; Miguel Reyes; Petia Radeva; Dimo Dimov; Sergio Escalera |
Title |
Graph Cuts Optimization for Multi-Limb Human Segmentation in Depth Maps |
Type |
Conference Article |
Year |
2012 |
Publication |
25th IEEE Conference on Computer Vision and Pattern Recognition |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
726-732 |
Keywords |
|
Abstract |
We present a generic framework for object segmentation using depth maps based on Random Forest and Graph-cuts theory, and apply it to the segmentation of human limbs in depth maps. First, from a set of random depth features, Random Forest is used to infer a set of label probabilities for each data sample. This vector of probabilities is used as unary term in α-β swap Graph-cuts algorithm. Moreover, depth of spatio-temporal neighboring data points are used as boundary potentials. Results on a new multi-label human depth data set show high performance in terms of segmentation overlapping of the novel methodology compared to classical approaches. |
Address |
Portland; Oregon; June 2013 |
Corporate Author |
|
Thesis |
|
Publisher |
IEEE Xplore |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1063-6919 |
ISBN |
978-1-4673-1226-4 |
Medium |
|
Area |
|
Expedition |
|
Conference ![sorted by Conference field, descending order (down)](img/sort_desc.gif) |
CVPR |
Notes |
MILAB;HuPBA |
Approved |
no |
Call Number |
Admin @ si @ HZM2012b |
Serial |
2046 |
Permanent link to this record |