Home | << 1 2 3 4 5 6 7 8 9 10 >> [11–13] |
![]() |
Records | Links | |||||
---|---|---|---|---|---|---|
Author | Christophe Rigaud; Dimosthenis Karatzas; Joost Van de Weijer; Jean-Christophe Burie; Jean-Marc Ogier |
![]() ![]() ![]() |
||||
Title | Automatic text localisation in scanned comic books | Type | Conference Article | |||
Year | 2013 | Publication | Proceedings of the International Conference on Computer Vision Theory and Applications | Abbreviated Journal | ||
Volume | Issue | Pages | 814-819 | |||
Keywords | Text localization; comics; text/graphic separation; complex background; unstructured document | |||||
Abstract | Comic books constitute an important cultural heritage asset in many countries. Digitization combined with subsequent document understanding enable direct content-based search as opposed to metadata only search (e.g. album title or author name). Few studies have been done in this direction. In this work we detail a novel approach for the automatic text localization in scanned comics book pages, an essential step towards a fully automatic comics book understanding. We focus on speech text as it is semantically important and represents the majority of the text present in comics. The approach is compared with existing methods of text localization found in the literature and results are presented. | |||||
Address | Barcelona; February 2013 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | VISAPP | |||
Notes ![]() |
DAG; CIC; 600.056 | Approved | no | |||
Call Number | Admin @ si @ RKW2013b | Serial | 2261 | |||
Permanent link to this record | ||||||
Author | Christophe Rigaud; Dimosthenis Karatzas; Joost Van de Weijer; Jean-Christophe Burie; Jean-Marc Ogier |
![]() ![]() ![]() ![]() |
||||
Title | An active contour model for speech balloon detection in comics | Type | Conference Article | |||
Year | 2013 | Publication | 12th International Conference on Document Analysis and Recognition | Abbreviated Journal | ||
Volume | Issue | Pages | 1240-1244 | |||
Keywords | ||||||
Abstract | Comic books constitute an important cultural heritage asset in many countries. Digitization combined with subsequent comic book understanding would enable a variety of new applications, including content-based retrieval and content retargeting. Document understanding in this domain is challenging as comics are semi-structured documents, combining semantically important graphical and textual parts. Few studies have been done in this direction. In this work we detail a novel approach for closed and non-closed speech balloon localization in scanned comic book pages, an essential step towards a fully automatic comic book understanding. The approach is compared with existing methods for closed balloon localization found in the literature and results are presented. | |||||
Address | washington; USA; August 2013 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 1520-5363 | ISBN | Medium | |||
Area | Expedition | Conference | ICDAR | |||
Notes ![]() |
DAG; CIC; 600.056 | Approved | no | |||
Call Number | Admin @ si @ RKW2013a | Serial | 2260 | |||
Permanent link to this record | ||||||
Author | Alicia Fornes; Josep Llados; Gemma Sanchez; Xavier Otazu; Horst Bunke |
![]() ![]() ![]() |
||||
Title | A Combination of Features for Symbol-Independent Writer Identification in Old Music Scores | Type | Journal Article | |||
Year | 2010 | Publication | International Journal on Document Analysis and Recognition | Abbreviated Journal | IJDAR | |
Volume | 13 | Issue | 4 | Pages | 243-259 | |
Keywords | ||||||
Abstract | The aim of writer identification is determining the writer of a piece of handwriting from a set of writers. In this paper, we present an architecture for writer identification in old handwritten music scores. Even though an important amount of music compositions contain handwritten text, the aim of our work is to use only music notation to determine the author. The main contribution is therefore the use of features extracted from graphical alphabets. Our proposal consists in combining the identification results of two different approaches, based on line and textural features. The steps of the ensemble architecture are the following. First of all, the music sheet is preprocessed for removing the staff lines. Then, music lines and texture images are generated for computing line features and textural features. Finally, the classification results are combined for identifying the writer. The proposed method has been tested on a database of old music scores from the seventeenth to nineteenth centuries, achieving a recognition rate of about 92% with 20 writers. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Springer-Verlag | Place of Publication | Editor | |||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 1433-2833 | ISBN | Medium | |||
Area | Expedition | Conference | ||||
Notes ![]() |
DAG; CAT;CIC | Approved | no | |||
Call Number | FLS2010b | Serial | 1319 | |||
Permanent link to this record | ||||||
Author | Alicia Fornes; Xavier Otazu; Josep Llados |
![]() ![]() ![]() ![]() |
||||
Title | Show through cancellation and image enhancement by multiresolution contrast processing | Type | Conference Article | |||
Year | 2013 | Publication | 12th International Conference on Document Analysis and Recognition | Abbreviated Journal | ||
Volume | Issue | Pages | 200-204 | |||
Keywords | ||||||
Abstract | Historical documents suffer from different types of degradation and noise such as background variation, uneven illumination or dark spots. In case of double-sided documents, another common problem is that the back side of the document usually interferes with the front side because of the transparency of the document or ink bleeding. This effect is called the show through phenomenon. Many methods are developed to solve these problems, and in the case of show-through, by scanning and matching both the front and back sides of the document. In contrast, our approach is designed to use only one side of the scanned document. We hypothesize that show-trough are low contrast components, while foreground components are high contrast ones. A Multiresolution Contrast (MC) decomposition is presented in order to estimate the contrast of features at different spatial scales. We cancel the show-through phenomenon by thresholding these low contrast components. This decomposition is also able to enhance the image removing shadowed areas by weighting spatial scales. Results show that the enhanced images improve the readability of the documents, allowing scholars both to recover unreadable words and to solve ambiguities. | |||||
Address | Washington; USA; August 2013 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 1520-5363 | ISBN | Medium | |||
Area | Expedition | Conference | ICDAR | |||
Notes ![]() |
DAG; 602.006; 600.045; 600.061; 600.052;CIC | Approved | no | |||
Call Number | Admin @ si @ FOL2013 | Serial | 2241 | |||
Permanent link to this record | ||||||
Author | Arjan Gijsenij; Theo Gevers; Joost Van de Weijer |
![]() ![]() ![]() ![]() ![]() |
||||
Title | Improving Color Constancy by Photometric Edge Weighting | Type | Journal Article | |||
Year | 2012 | Publication | IEEE Transaction on Pattern Analysis and Machine Intelligence | Abbreviated Journal | TPAMI | |
Volume | 34 | Issue | 5 | Pages | 918-929 | |
Keywords | ||||||
Abstract | : Edge-based color constancy methods make use of image derivatives to estimate the illuminant. However, different edge types exist in real-world images such as material, shadow and highlight edges. These different edge types may have a distinctive influence on the performance of the illuminant estimation. Therefore, in this paper, an extensive analysis is provided of different edge types on the performance of edge-based color constancy methods. First, an edge-based taxonomy is presented classifying edge types based on their photometric properties (e.g. material, shadow-geometry and highlights). Then, a performance evaluation of edge-based color constancy is provided using these different edge types. From this performance evaluation it is derived that specular and shadow edge types are more valuable than material edges for the estimation of the illuminant. To this end, the (iterative) weighted Grey-Edge algorithm is proposed in which these edge types are more emphasized for the estimation of the illuminant. Images that are recorded under controlled circumstances demonstrate that the proposed iterative weighted Grey-Edge algorithm based on highlights reduces the median angular error with approximately $25\%$. In an uncontrolled environment, improvements in angular error up to $11\%$ are obtained with respect to regular edge-based color constancy. | |||||
Address | Los Alamitos; CA; USA; | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 0162-8828 | ISBN | Medium | |||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC;ISE | Approved | no | |||
Call Number | Admin @ si @ GGW2012 | Serial | 1850 | |||
Permanent link to this record | ||||||
Author | Albert Gordo |
![]() ![]() |
||||
Title | A Cyclic Page Layout Descriptor for Document Classification & Retrieval | Type | Report | |||
Year | 2009 | Publication | CVC Technical Report | Abbreviated Journal | ||
Volume | 128 | Issue | Pages | |||
Keywords | ||||||
Abstract | ||||||
Address | ||||||
Corporate Author | Computer Vision Center | Thesis | Master's thesis | |||
Publisher | Place of Publication | Bellaterra, Barcelona | Editor | |||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC;DAG | Approved | no | |||
Call Number | Admin @ si @ Gor2009 | Serial | 2387 | |||
Permanent link to this record | ||||||
Author | Maria del Camp Davesa |
![]() ![]() |
||||
Title | Human action categorization in image sequences | Type | Report | |||
Year | 2011 | Publication | CVC Technical Report | Abbreviated Journal | ||
Volume | 169 | Issue | Pages | |||
Keywords | ||||||
Abstract | ||||||
Address | Bellaterra (Spain) | |||||
Corporate Author | Computer Vision Center | Thesis | Master's thesis | |||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ||||
Notes ![]() |
CiC;CIC | Approved | no | |||
Call Number | Admin @ si @ Dav2011 | Serial | 1934 | |||
Permanent link to this record | ||||||
Author | Marcos V Conde; Javier Vazquez; Michael S Brown; Radu TImofte |
![]() ![]() ![]() ![]() |
||||
Title | NILUT: Conditional Neural Implicit 3D Lookup Tables for Image Enhancement | Type | Conference Article | |||
Year | 2024 | Publication | 38th AAAI Conference on Artificial Intelligence | Abbreviated Journal | ||
Volume | Issue | Pages | ||||
Keywords | ||||||
Abstract | 3D lookup tables (3D LUTs) are a key component for image enhancement. Modern image signal processors (ISPs) have dedicated support for these as part of the camera rendering pipeline. Cameras typically provide multiple options for picture styles, where each style is usually obtained by applying a unique handcrafted 3D LUT. Current approaches for learning and applying 3D LUTs are notably fast, yet not so memory-efficient, as storing multiple 3D LUTs is required. For this reason and other implementation limitations, their use on mobile devices is less popular. In this work, we propose a Neural Implicit LUT (NILUT), an implicitly defined continuous 3D color transformation parameterized by a neural network. We show that NILUTs are capable of accurately emulating real 3D LUTs. Moreover, a NILUT can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly. Our novel approach is memory-efficient, controllable and can complement previous methods, including learned ISPs. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | AAAI | |||
Notes ![]() |
CIC; MACO | Approved | no | |||
Call Number | Admin @ si @ CVB2024 | Serial | 3872 | |||
Permanent link to this record | ||||||
Author | Danna Xue; Javier Vazquez; Luis Herranz; Yang Zhang; Michael S Brown |
![]() ![]() ![]() |
||||
Title | Integrating High-Level Features for Consistent Palette-based Multi-image Recoloring | Type | Journal Article | |||
Year | 2023 | Publication | Computer Graphics Forum | Abbreviated Journal | CGF | |
Volume | Issue | Pages | ||||
Keywords | ||||||
Abstract | Achieving visually consistent colors across multiple images is important when images are used in photo albums, websites, and brochures. Unfortunately, only a handful of methods address multi-image color consistency compared to one-to-one color transfer techniques. Furthermore, existing methods do not incorporate high-level features that can assist graphic designers in their work. To address these limitations, we introduce a framework that builds upon a previous palette-based color consistency method and incorporates three high-level features: white balance, saliency, and color naming. We show how these features overcome the limitations of the prior multi-consistency workflow and showcase the user-friendly nature of our framework. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; MACO | Approved | no | |||
Call Number | Admin @ si @ XVH2023 | Serial | 3883 | |||
Permanent link to this record | ||||||
Author | Danna Xue; Luis Herranz; Javier Vazquez; Yanning Zhang |
![]() ![]() ![]() ![]() |
||||
Title | Burst Perception-Distortion Tradeoff: Analysis and Evaluation | Type | Conference Article | |||
Year | 2023 | Publication | IEEE International Conference on Acoustics, Speech and Signal Processing | Abbreviated Journal | ||
Volume | Issue | Pages | ||||
Keywords | ||||||
Abstract | Burst image restoration attempts to effectively utilize the complementary cues appearing in sequential images to produce a high-quality image. Most current methods use all the available images to obtain the reconstructed image. However, using more images for burst restoration is not always the best option regarding reconstruction quality and efficiency, as the images acquired by handheld imaging devices suffer from degradation and misalignment caused by the camera noise and shake. In this paper, we extend the perception-distortion tradeoff theory by introducing multiple-frame information. We propose the area of the unattainable region as a new metric for perception-distortion tradeoff evaluation and comparison. Based on this metric, we analyse the performance of burst restoration from the perspective of the perception-distortion tradeoff under both aligned bursts and misaligned bursts situations. Our analysis reveals the importance of inter-frame alignment for burst restoration and shows that the optimal burst length for the restoration model depends both on the degree of degradation and misalignment. | |||||
Address | Rodhes Islands; Greece; June 2023 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ICASSP | |||
Notes ![]() |
CIC; MACO | Approved | no | |||
Call Number | Admin @ si @ XHV2023 | Serial | 3909 | |||
Permanent link to this record | ||||||
Author | Fahad Shahbaz Khan; Joost Van de Weijer; Andrew Bagdanov; Michael Felsberg |
![]() ![]() ![]() ![]() |
||||
Title | Scale Coding Bag-of-Words for Action Recognition | Type | Conference Article | |||
Year | 2014 | Publication | 22nd International Conference on Pattern Recognition | Abbreviated Journal | ||
Volume | Issue | Pages | 1514-1519 | |||
Keywords | ||||||
Abstract | Recognizing human actions in still images is a challenging problem in computer vision due to significant amount of scale, illumination and pose variation. Given the bounding box of a person both at training and test time, the task is to classify the action associated with each bounding box in an image.
Most state-of-the-art methods use the bag-of-words paradigm for action recognition. The bag-of-words framework employing a dense multi-scale grid sampling strategy is the de facto standard for feature detection. This results in a scale invariant image representation where all the features at multiple-scales are binned in a single histogram. We argue that such a scale invariant strategy is sub-optimal since it ignores the multi-scale information available with each bounding box of a person. This paper investigates alternative approaches to scale coding for action recognition in still images. We encode multi-scale information explicitly in three different histograms for small, medium and large scale visual-words. Our first approach exploits multi-scale information with respect to the image size. In our second approach, we encode multi-scale information relative to the size of the bounding box of a person instance. In each approach, the multi-scale histograms are then concatenated into a single representation for action classification. We validate our approaches on the Willow dataset which contains seven action categories: interacting with computer, photography, playing music, riding bike, riding horse, running and walking. Our results clearly suggest that the proposed scale coding approaches outperform the conventional scale invariant technique. Moreover, we show that our approach obtains promising results compared to more complex state-of-the-art methods. |
|||||
Address | Stockholm; August 2014 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ICPR | |||
Notes ![]() |
CIC; LAMP; 601.240; 600.074; 600.079 | Approved | no | |||
Call Number | Admin @ si @ KWB2014 | Serial | 2450 | |||
Permanent link to this record | ||||||
Author | Fahad Shahbaz Khan; Joost Van de Weijer; Muhammad Anwer Rao; Michael Felsberg; Carlo Gatta |
![]() ![]() ![]() ![]() |
||||
Title | Semantic Pyramids for Gender and Action Recognition | Type | Journal Article | |||
Year | 2014 | Publication | IEEE Transactions on Image Processing | Abbreviated Journal | TIP | |
Volume | 23 | Issue | 8 | Pages | 3633-3645 | |
Keywords | ||||||
Abstract | Person description is a challenging problem in computer vision. We investigated two major aspects of person description: 1) gender and 2) action recognition in still images. Most state-of-the-art approaches for gender and action recognition rely on the description of a single body part, such as face or full-body. However, relying on a single body part is suboptimal due to significant variations in scale, viewpoint, and pose in real-world images. This paper proposes a semantic pyramid approach for pose normalization. Our approach is fully automatic and based on combining information from full-body, upper-body, and face regions for gender and action recognition in still images. The proposed approach does not require any annotations for upper-body and face of a person. Instead, we rely on pretrained state-of-the-art upper-body and face detectors to automatically extract semantic information of a person. Given multiple bounding boxes from each body part detector, we then propose a simple method to select the best candidate bounding box, which is used for feature extraction. Finally, the extracted features from the full-body, upper-body, and face regions are combined into a single representation for classification. To validate the proposed approach for gender recognition, experiments are performed on three large data sets namely: 1) human attribute; 2) head-shoulder; and 3) proxemics. For action recognition, we perform experiments on four data sets most used for benchmarking action recognition in still images: 1) Sports; 2) Willow; 3) PASCAL VOC 2010; and 4) Stanford-40. Our experiments clearly demonstrate that the proposed approach, despite its simplicity, outperforms state-of-the-art methods for gender and action recognition. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 1057-7149 | ISBN | Medium | |||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; LAMP; 601.160; 600.074; 600.079;MILAB | Approved | no | |||
Call Number | Admin @ si @ KWR2014 | Serial | 2507 | |||
Permanent link to this record | ||||||
Author | Domicele Jonauskaite; Lucia Camenzind; C. Alejandro Parraga; Cecile N Diouf; Mathieu Mercapide Ducommun; Lauriane Müller; Melanie Norberg; Christine Mohr |
![]() ![]() ![]() ![]() |
||||
Title | Colour-emotion associations in individuals with red-green colour blindness | Type | Journal Article | |||
Year | 2021 | Publication | PeerJ | Abbreviated Journal | ||
Volume | 9 | Issue | Pages | e11180 | ||
Keywords | Affect; Chromotherapy; Colour cognition; Colour vision deficiency; Cross-modal correspondences; Daltonism; Deuteranopia; Dichromatic; Emotion; Protanopia. | |||||
Abstract | Colours and emotions are associated in languages and traditions. Some of us may convey sadness by saying feeling blue or by wearing black clothes at funerals. The first example is a conceptual experience of colour and the second example is an immediate perceptual experience of colour. To investigate whether one or the other type of experience more strongly drives colour-emotion associations, we tested 64 congenitally red-green colour-blind men and 66 non-colour-blind men. All participants associated 12 colours, presented as terms or patches, with 20 emotion concepts, and rated intensities of the associated emotions. We found that colour-blind and non-colour-blind men associated similar emotions with colours, irrespective of whether colours were conveyed via terms (r = .82) or patches (r = .80). The colour-emotion associations and the emotion intensities were not modulated by participants' severity of colour blindness. Hinting at some additional, although minor, role of actual colour perception, the consistencies in associations for colour terms and patches were higher in non-colour-blind than colour-blind men. Together, these results suggest that colour-emotion associations in adults do not require immediate perceptual colour experiences, as conceptual experiences are sufficient. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; LAMP; 600.120; 600.128 | Approved | no | |||
Call Number | Admin @ si @ JCP2021 | Serial | 3564 | |||
Permanent link to this record | ||||||
Author | Shida Beigpour; Christian Riess; Joost Van de Weijer; Elli Angelopoulou |
![]() ![]() ![]() ![]() |
||||
Title | Multi-Illuminant Estimation with Conditional Random Fields | Type | Journal Article | |||
Year | 2014 | Publication | IEEE Transactions on Image Processing | Abbreviated Journal | TIP | |
Volume | 23 | Issue | 1 | Pages | 83-95 | |
Keywords | color constancy; CRF; multi-illuminant | |||||
Abstract | Most existing color constancy algorithms assume uniform illumination. However, in real-world scenes, this is not often the case. Thus, we propose a novel framework for estimating the colors of multiple illuminants and their spatial distribution in the scene. We formulate this problem as an energy minimization task within a conditional random field over a set of local illuminant estimates. In order to quantitatively evaluate the proposed method, we created a novel data set of two-dominant-illuminant images comprised of laboratory, indoor, and outdoor scenes. Unlike prior work, our database includes accurate pixel-wise ground truth illuminant information. The performance of our method is evaluated on multiple data sets. Experimental results show that our framework clearly outperforms single illuminant estimators as well as a recently proposed multi-illuminant estimation approach. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 1057-7149 | ISBN | Medium | |||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; LAMP; 600.074; 600.079 | Approved | no | |||
Call Number | Admin @ si @ BRW2014 | Serial | 2451 | |||
Permanent link to this record | ||||||
Author | M. Danelljan; Fahad Shahbaz Khan; Michael Felsberg; Joost Van de Weijer |
![]() ![]() ![]() ![]() |
||||
Title | Adaptive color attributes for real-time visual tracking | Type | Conference Article | |||
Year | 2014 | Publication | 27th IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | ||
Volume | Issue | Pages | 1090 - 1097 | |||
Keywords | ||||||
Abstract | Visual tracking is a challenging problem in computer vision. Most state-of-the-art visual trackers either rely on luminance information or use simple color representations for image description. Contrary to visual tracking, for object
recognition and detection, sophisticated color features when combined with luminance have shown to provide excellent performance. Due to the complexity of the tracking problem, the desired color feature should be computationally efficient, and possess a certain amount of photometric invariance while maintaining high discriminative power. This paper investigates the contribution of color in a tracking-by-detection framework. Our results suggest that color attributes provides superior performance for visual tracking. We further propose an adaptive low-dimensional variant of color attributes. Both quantitative and attributebased evaluations are performed on 41 challenging benchmark color sequences. The proposed approach improves the baseline intensity-based tracker by 24% in median distance precision. Furthermore, we show that our approach outperforms state-of-the-art tracking methods while running at more than 100 frames per second. |
|||||
Address | Nottingham; UK; September 2014 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | CVPR | |||
Notes ![]() |
CIC; LAMP; 600.074; 600.079 | Approved | no | |||
Call Number | Admin @ si @ DKF2014 | Serial | 2509 | |||
Permanent link to this record | ||||||
Author | Fahad Shahbaz Khan; Shida Beigpour; Joost Van de Weijer; Michael Felsberg |
![]() ![]() ![]() |
||||
Title | Painting-91: A Large Scale Database for Computational Painting Categorization | Type | Journal Article | |||
Year | 2014 | Publication | Machine Vision and Applications | Abbreviated Journal | MVAP | |
Volume | 25 | Issue | 6 | Pages | 1385-1397 | |
Keywords | ||||||
Abstract | Computer analysis of visual art, especially paintings, is an interesting cross-disciplinary research domain. Most of the research in the analysis of paintings involve medium to small range datasets with own specific settings. Interestingly, significant progress has been made in the field of object and scene recognition lately. A key factor in this success is the introduction and availability of benchmark datasets for evaluation. Surprisingly, such a benchmark setup is still missing in the area of computational painting categorization. In this work, we propose a novel large scale dataset of digital paintings. The dataset consists of paintings from 91 different painters. We further show three applications of our dataset namely: artist categorization, style classification and saliency detection. We investigate how local and global features popular in image classification perform for the tasks of artist and style categorization. For both categorization tasks, our experimental results suggest that combining multiple features significantly improves the final performance. We show that state-of-the-art computer vision methods can correctly classify 50 % of unseen paintings to its painter in a large dataset and correctly attribute its artistic style in over 60 % of the cases. Additionally, we explore the task of saliency detection on paintings and show experimental findings using state-of-the-art saliency estimation algorithms. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | |||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | 0932-8092 | ISBN | Medium | |||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; LAMP; 600.074; 600.079 | Approved | no | |||
Call Number | Admin @ si @ KBW2014 | Serial | 2510 | |||
Permanent link to this record | ||||||
Author | Bojana Gajic; Ariel Amato; Ramon Baldrich; Joost Van de Weijer; Carlo Gatta |
![]() ![]() ![]() ![]() |
||||
Title | Area Under the ROC Curve Maximization for Metric Learning | Type | Conference Article | |||
Year | 2022 | Publication | CVPR 2022 Workshop on Efficien Deep Learning for Computer Vision (ECV 2022, 5th Edition) | Abbreviated Journal | ||
Volume | Issue | Pages | ||||
Keywords | Training; Computer vision; Conferences; Area measurement; Benchmark testing; Pattern recognition | |||||
Abstract | Most popular metric learning losses have no direct relation with the evaluation metrics that are subsequently applied to evaluate their performance. We hypothesize that training a metric learning model by maximizing the area under the ROC curve (which is a typical performance measure of recognition systems) can induce an implicit ranking suitable for retrieval problems. This hypothesis is supported by previous work that proved that a curve dominates in ROC space if and only if it dominates in Precision-Recall space. To test this hypothesis, we design and maximize an approximated, derivable relaxation of the area under the ROC curve. The proposed AUC loss achieves state-of-the-art results on two large scale retrieval benchmark datasets (Stanford Online Products and DeepFashion In-Shop). Moreover, the AUC loss achieves comparable performance to more complex, domain specific, state-of-the-art methods for vehicle re-identification. | |||||
Address | New Orleans, USA; 20 June 2022 | |||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | CVPRW | |||
Notes ![]() |
CIC; LAMP; | Approved | no | |||
Call Number | Admin @ si @ GAB2022 | Serial | 3700 | |||
Permanent link to this record | ||||||
Author | C. Alejandro Parraga; Jordi Roca; Dimosthenis Karatzas; Sophie Wuerger |
![]() ![]() ![]() ![]() ![]() |
||||
Title | Limitations of visual gamma corrections in LCD displays | Type | Journal Article | |||
Year | 2014 | Publication | Displays | Abbreviated Journal | Dis | |
Volume | 35 | Issue | 5 | Pages | 227–239 | |
Keywords | Display calibration; Psychophysics; Perceptual; Visual gamma correction; Luminance matching; Observer-based calibration | |||||
Abstract | A method for estimating the non-linear gamma transfer function of liquid–crystal displays (LCDs) without the need of a photometric measurement device was described by Xiao et al. (2011) [1]. It relies on observer’s judgments of visual luminance by presenting eight half-tone patterns with luminances from 1/9 to 8/9 of the maximum value of each colour channel. These half-tone patterns were distributed over the screen both over the vertical and horizontal viewing axes. We conducted a series of photometric and psychophysical measurements (consisting in the simultaneous presentation of half-tone patterns in each trial) to evaluate whether the angular dependency of the light generated by three different LCD technologies would bias the results of these gamma transfer function estimations. Our results show that there are significant differences between the gamma transfer functions measured and produced by observers at different viewing angles. We suggest appropriate modifications to the Xiao et al. paradigm to counterbalance these artefacts which also have the advantage of shortening the amount of time spent in collecting the psychophysical measurements. | |||||
Address | ||||||
Corporate Author | Thesis | |||||
Publisher | Place of Publication | Editor | ||||
Language | Summary Language | Original Title | ||||
Series Editor | Series Title | Abbreviated Series Title | ||||
Series Volume | Series Issue | Edition | ||||
ISSN | ISBN | Medium | ||||
Area | Expedition | Conference | ||||
Notes ![]() |
CIC; DAG; 600.052; 600.077; 600.074 | Approved | no | |||
Call Number | Admin @ si @ PRK2014 | Serial | 2511 | |||
Permanent link to this record |