|
Records |
Links |
|
Author |
Anjan Dutta; Josep Llados; Horst Bunke; Umapada Pal |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
A Product graph based method for dual subgraph matching applied to symbol spotting |
Type |
Conference Article |
|
Year |
2013 |
Publication |
10th IAPR International Workshop on Graphics Recognition |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
|
|
|
Abstract |
Product graph has been shown to be an efficient way for matching subgraphs. This paper reports the extension of the product graph methodology for subgraph matching applied to symbol spotting in graphical documents. This paper focuses on the two major limitations of the previous version of product graph: (1) Spurious nodes and edges in the graph representation and (2) Inefficient node and edge attributes. To deal with noisy information of vectorized graphical documents, we consider a dual graph representation on the original graph representing the graphical information and the product graph is computed between the dual graphs of the query graphs and the input graph.
The dual graph with redundant edges is helpful for efficient and tolerating encoding of the structural information of the graphical documents. The adjacency matrix of the product graph locates similar path information of two graphs and exponentiating the adjacency matrix finds similar paths of greater lengths. Nodes joining similar paths between two graphs are found by combining different exponentials of adjacency matrices. An experimental investigation reveals that the recall obtained by this approach is quite encouraging. |
|
|
Address |
Bethlehem; PA; USA; August 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
GREC |
|
|
Notes |
DAG |
Approved |
no |
|
|
Call Number |
Admin @ si @ DLB2013b |
Serial |
2359 |
|
Permanent link to this record |
|
|
|
|
Author |
Ivan Huerta; Ariel Amato; Xavier Roca; Jordi Gonzalez |
![download PDF file pdf](img/file_PDF.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Exploiting Multiple Cues in Motion Segmentation Based on Background Subtraction |
Type |
Journal Article |
|
Year |
2013 |
Publication |
Neurocomputing |
Abbreviated Journal |
NEUCOM |
|
|
Volume |
100 |
Issue |
|
Pages |
183–196 |
|
|
Keywords |
Motion segmentation; Shadow suppression; Colour segmentation; Edge segmentation; Ghost detection; Background subtraction |
|
|
Abstract |
This paper presents a novel algorithm for mobile-object segmentation from static background scenes, which is both robust and accurate under most of the common problems found in motionsegmentation. In our first contribution, a case analysis of motionsegmentation errors is presented taking into account the inaccuracies associated with different cues, namely colour, edge and intensity. Our second contribution is an hybrid architecture which copes with the main issues observed in the case analysis by fusing the knowledge from the aforementioned three cues and a temporal difference algorithm. On one hand, we enhance the colour and edge models to solve not only global and local illumination changes (i.e. shadows and highlights) but also the camouflage in intensity. In addition, local information is also exploited to solve the camouflage in chroma. On the other hand, the intensity cue is applied when colour and edge cues are not available because their values are beyond the dynamic range. Additionally, temporal difference scheme is included to segment motion where those three cues cannot be reliably computed, for example in those background regions not visible during the training period. Lastly, our approach is extended for handling ghost detection. The proposed method obtains very accurate and robust motionsegmentation results in multiple indoor and outdoor scenarios, while outperforming the most-referred state-of-art approaches. |
|
|
Address |
|
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
Elsevier |
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
|
|
|
Notes |
ISE |
Approved |
no |
|
|
Call Number |
Admin @ si @ HAR2013 |
Serial |
1808 |
|
Permanent link to this record |
|
|
|
|
Author |
Bhaskar Chakraborty; Andrew Bagdanov; Jordi Gonzalez; Xavier Roca |
![download PDF file pdf](img/file_PDF.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Human Action Recognition Using an Ensemble of Body-Part Detectors |
Type |
Journal Article |
|
Year |
2013 |
Publication |
Expert Systems |
Abbreviated Journal |
EXSY |
|
|
Volume |
30 |
Issue |
2 |
Pages |
101-114 |
|
|
Keywords |
Human action recognition;body-part detection;hidden Markov model |
|
|
Abstract |
This paper describes an approach to human action recognition based on a probabilistic optimization model of body parts using hidden Markov model (HMM). Our method is able to distinguish between similar actions by only considering the body parts having major contribution to the actions, for example, legs for walking, jogging and running; arms for boxing, waving and clapping. We apply HMMs to model the stochastic movement of the body parts for action recognition. The HMM construction uses an ensemble of body-part detectors, followed by grouping of part detections, to perform human identification. Three example-based body-part detectors are trained to detect three components of the human body: the head, legs and arms. These detectors cope with viewpoint changes and self-occlusions through the use of ten sub-classifiers that detect body parts over a specific range of viewpoints. Each sub-classifier is a support vector machine trained on features selected for the discriminative power for each particular part/viewpoint combination. Grouping of these detections is performed using a simple geometric constraint model that yields a viewpoint-invariant human detector. We test our approach on three publicly available action datasets: the KTH dataset, Weizmann dataset and HumanEva dataset. Our results illustrate that with a simple and compact representation we can achieve robust recognition of human actions comparable to the most complex, state-of-the-art methods. |
|
|
Address |
|
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
|
|
|
Notes |
ISE |
Approved |
no |
|
|
Call Number |
Admin @ si @ CBG2013 |
Serial |
1809 |
|
Permanent link to this record |
|
|
|
|
Author |
Kaida Xiao; Chenyang Fu; D.Mylonas; Dimosthenis Karatzas; S. Wuerger |
![goto web page url](img/www.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Unique Hue Data for Colour Appearance Models. Part ii: Chromatic Adaptation Transform |
Type |
Journal Article |
|
Year |
2013 |
Publication |
Color Research & Application |
Abbreviated Journal |
CRA |
|
|
Volume |
38 |
Issue |
1 |
Pages |
22-29 |
|
|
Keywords |
|
|
|
Abstract |
Unique hue settings of 185 observers under three room-lighting conditions were used to evaluate the accuracy of full and mixed chromatic adaptation transform models of CIECAM02 in terms of unique hue reproduction. Perceptual hue shifts in CIECAM02 were evaluated for both models with no clear difference using the current Commission Internationale de l'Éclairage (CIE) recommendation for mixed chromatic adaptation ratio. Using our large dataset of unique hue data as a benchmark, an optimised parameter is proposed for chromatic adaptation under mixed illumination conditions that produces more accurate results in unique hue reproduction. © 2011 Wiley Periodicals, Inc. Col Res Appl, 2013 |
|
|
Address |
|
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
|
|
|
Notes |
DAG |
Approved |
no |
|
|
Call Number |
Admin @ si @ XFM2013 |
Serial |
1822 |
|
Permanent link to this record |
|
|
|
|
Author |
S.Grau; Ana Puig; Sergio Escalera; Maria Salamo |
![download PDF file pdf](img/file_PDF.gif)
![goto web page (via DOI) doi](img/doi.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Intelligent Interactive Volume Classification |
Type |
Conference Article |
|
Year |
2013 |
Publication |
Pacific Graphics |
Abbreviated Journal |
|
|
|
Volume |
32 |
Issue |
7 |
Pages |
23-28 |
|
|
Keywords |
|
|
|
Abstract |
This paper defines an intelligent and interactive framework to classify multiple regions of interest from the original data on demand, without requiring any preprocessing or previous segmentation. The proposed intelligent and interactive approach is divided in three stages: visualize, training and testing. First, users visualize and label some samples directly on slices of the volume. Training and testing are based on a framework of Error Correcting Output Codes and Adaboost classifiers that learn to classify each region the user has painted. Later, at the testing stage, each classifier is directly applied on the rest of samples and combined to perform multi-class labeling, being used in the final rendering. We also parallelized the training stage using a GPU-based implementation for
obtaining a rapid interaction and classification. |
|
|
Address |
|
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
978-3-905674-50-7 |
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
PG |
|
|
Notes |
HuPBA; 600.046;MILAB |
Approved |
no |
|
|
Call Number |
Admin @ si @ GPE2013b |
Serial |
2355 |
|
Permanent link to this record |
|
|
|
|
Author |
Yainuvis Socarras; Sebastian Ramos; David Vazquez; Antonio Lopez; Theo Gevers |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Adapting Pedestrian Detection from Synthetic to Far Infrared Images |
Type |
Conference Article |
|
Year |
2013 |
Publication |
ICCV Workshop on Visual Domain Adaptation and Dataset Bias |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
Domain Adaptation; Far Infrared; Pedestrian Detection |
|
|
Abstract |
We present different techniques to adapt a pedestrian classifier trained with synthetic images and the corresponding automatically generated annotations to operate with far infrared (FIR) images. The information contained in this kind of images allow us to develop a robust pedestrian detector invariant to extreme illumination changes. |
|
|
Address |
Sydney; Australia; December 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
Sydney, Australy |
Editor |
|
|
|
Language |
English |
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
ICCVW-VisDA |
|
|
Notes |
ADAS; 600.054; 600.055; 600.057; 601.217;ISE |
Approved |
no |
|
|
Call Number |
ADAS @ adas @ SRV2013 |
Serial |
2334 |
|
Permanent link to this record |
|
|
|
|
Author |
V.C.Kieu; Alicia Fornes; M. Visani; N.Journet ; Anjan Dutta |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
The ICDAR/GREC 2013 Music Scores Competition on Staff Removal |
Type |
Conference Article |
|
Year |
2013 |
Publication |
10th IAPR International Workshop on Graphics Recognition |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
Competition; Music scores; Staff Removal |
|
|
Abstract |
The first competition on music scores that was organized at ICDAR and GREC in 2011 awoke the interest of researchers, who participated both at staff removal and writer identification tasks. In this second edition, we propose a staff removal competition where we simulate old music scores. Thus, we have created a new set of images, which contain noise and 3D distortions. This paper describes the distortion methods, metrics, the participant’s methods and the obtained results. |
|
|
Address |
Bethlehem; PA; USA; August 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
GREC |
|
|
Notes |
DAG; 600.045; 600.061 |
Approved |
no |
|
|
Call Number |
Admin @ si @ KFV2013 |
Serial |
2337 |
|
Permanent link to this record |
|
|
|
|
Author |
Jorge Bernal; David Vazquez (eds) |
![download PDF file pdf](img/file_PDF.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Computer vision Trends and Challenges |
Type |
Book Whole |
|
Year |
2013 |
Publication |
Computer vision Trends and Challenges |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
CVCRD; Computer Vision |
|
|
Abstract |
This book contains the papers presented at the Eighth CVC Workshop on Computer Vision Trends and Challenges (CVCR&D'2013). The workshop was held at the Computer Vision Center (Universitat Autònoma de Barcelona), the October 25th, 2013. The CVC workshops provide an excellent opportunity for young researchers and project engineers to share new ideas and knowledge about the progress of their work, and also, to discuss about challenges and future perspectives. In addition, the workshop is the welcome event for new people that recently have joined the institute.
The program of CVCR&D is organized in a single-track single-day workshop. It comprises several sessions dedicated to specific topics. For each session, a doctor working on the topic introduces the general research lines. The PhD students expose their specific research. A poster session will be held for open questions. Session topics cover the current research lines and development projects of the CVC: Medical Imaging, Medical Imaging, Color & Texture Analysis, Object Recognition, Image Sequence Evaluation, Advanced Driver Assistance Systems, Machine Vision, Document Analysis, Pattern Recognition and Applications. We want to thank all paper authors and Program Committee members. Their contribution shows that the CVC has a dynamic, active, and promising scientific community.
We hope you all enjoy this Eighth workshop and we are looking forward to meeting you and new people next year in the Ninth CVCR&D. |
|
|
Address |
|
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
Jorge Bernal; David Vazquez |
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
978-84-940902-2-6 |
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
|
|
|
Notes |
|
Approved |
no |
|
|
Call Number |
ADAS @ adas @ BeV2013 |
Serial |
2339 |
|
Permanent link to this record |
|
|
|
|
Author |
Patricia Marquez; Debora Gil; Aura Hernandez-Sabate |
![download PDF file pdf](img/file_PDF.gif)
![goto web page (via DOI) doi](img/doi.gif)
|
|
Title |
Evaluation of the Capabilities of Confidence Measures for Assessing Optical Flow Quality |
Type |
Conference Article |
|
Year |
2013 |
Publication |
ICCV Workshop on Computer Vision in Vehicle Technology: From Earth to Mars |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
624-631 |
|
|
Keywords |
|
|
|
Abstract |
Assessing Optical Flow (OF) quality is essential for its further use in reliable decision support systems. The absence of ground truth in such situations leads to the computation of OF Confidence Measures (CM) obtained from either input or output data. A fair comparison across the capabilities of the different CM for bounding OF error is required in order to choose the best OF-CM pair for discarding points where OF computation is not reliable. This paper presents a statistical probabilistic framework for assessing the quality of a given CM. Our quality measure is given in terms of the percentage of pixels whose OF error bound can not be determined by CM values. We also provide statistical tools for the computation of CM values that ensures a given accuracy of the flow field. |
|
|
Address |
Sydney; Australia; December 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
CVTT:E2M |
|
|
Notes |
IAM; ADAS; 600.044; 600.057; 601.145 |
Approved |
no |
|
|
Call Number |
Admin @ si @ MGH2013b |
Serial |
2351 |
|
Permanent link to this record |
|
|
|
|
Author |
Marçal Rusiñol; V. Poulain d'Andecy; Dimosthenis Karatzas; Josep Llados |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Classification of Administrative Document Images by Logo Identification |
Type |
Conference Article |
|
Year |
2013 |
Publication |
10th IAPR International Workshop on Graphics Recognition |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
|
|
|
Abstract |
This paper is focused on the categorization of administrative document images (such as invoices) based on the recognition of the supplier's graphical logo. Two different methods are proposed, the first one uses a bag-of-visual-words model whereas the second one tries to locate logo images described by the blurred shape model descriptor within documents by a sliding-window technique. Preliminar results are reported with a dataset of real administrative documents. |
|
|
Address |
Bethlehem; PA; USA; August 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
GREC |
|
|
Notes |
DAG; 600.056; 600.045; 605.203 |
Approved |
no |
|
|
Call Number |
Admin @ si @ |
Serial |
2348 |
|
Permanent link to this record |
|
|
|
|
Author |
Marçal Rusiñol; Dimosthenis Karatzas; Josep Llados |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Spotting Graphical Symbols in Camera-Acquired Documents in Real Time |
Type |
Conference Article |
|
Year |
2013 |
Publication |
10th IAPR International Workshop on Graphics Recognition |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
|
|
|
Abstract |
In this paper we present a system devoted to spot graphical symbols in camera-acquired document images. The system is based on the extraction and further matching of ORB compact local features computed over interest key-points. Then, the FLANN indexing framework based on approximate nearest neighbor search allows to efficiently match local descriptors between the captured scene and the graphical models. Finally, the RANSAC algorithm is used in order to compute the homography between the spotted symbol and its appearance in the document image. The proposed approach is efficient and is able to work in real time. |
|
|
Address |
Bethlehem; PA; USA; August 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
GREC |
|
|
Notes |
DAG; 600.045; 600.055; 600.061; 602.101 |
Approved |
no |
|
|
Call Number |
Admin @ si @ RKL2013 |
Serial |
2347 |
|
Permanent link to this record |
|
|
|
|
Author |
Jiaolong Xu; Sebastian Ramos; Xu Hu; David Vazquez; Antonio Lopez |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Multi-task Bilinear Classifiers for Visual Domain Adaptation |
Type |
Conference Article |
|
Year |
2013 |
Publication |
Advances in Neural Information Processing Systems Workshop |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
Domain Adaptation; Pedestrian Detection; ADAS |
|
|
Abstract |
We propose a method that aims to lessen the significant accuracy degradation
that a discriminative classifier can suffer when it is trained in a specific domain (source domain) and applied in a different one (target domain). The principal reason for this degradation is the discrepancies in the distribution of the features that feed the classifier in different domains. Therefore, we propose a domain adaptation method that maps the features from the different domains into a common subspace and learns a discriminative domain-invariant classifier within it. Our algorithm combines bilinear classifiers and multi-task learning for domain adaptation.
The bilinear classifier encodes the feature transformation and classification
parameters by a matrix decomposition. In this way, specific feature transformations for multiple domains and a shared classifier are jointly learned in a multi-task learning framework. Focusing on domain adaptation for visual object detection, we apply this method to the state-of-the-art deformable part-based model for cross domain pedestrian detection. Experimental results show that our method significantly avoids the domain drift and improves the accuracy when compared to several baselines. |
|
|
Address |
Lake Tahoe; Nevada; USA; December 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
NIPSW |
|
|
Notes |
ADAS; 600.054; 600.057; 601.217;ISE |
Approved |
no |
|
|
Call Number |
ADAS @ adas @ XRH2013 |
Serial |
2340 |
|
Permanent link to this record |
|
|
|
|
Author |
Thanh Ha Do; Salvatore Tabbone; Oriol Ramos Terrades |
![goto web page (via DOI) doi](img/doi.gif)
![find record details (via OpenURL) openurl](img/xref.gif)
|
|
Title |
Document noise removal using sparse representations over learned dictionary |
Type |
Conference Article |
|
Year |
2013 |
Publication |
Symposium on Document engineering |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
161-168 |
|
|
Keywords |
|
|
|
Abstract |
best paper award
In this paper, we propose an algorithm for denoising document images using sparse representations. Following a training set, this algorithm is able to learn the main document characteristics and also, the kind of noise included into the documents. In this perspective, we propose to model the noise energy based on the normalized cross-correlation between pairs of noisy and non-noisy documents. Experimental
results on several datasets demonstrate the robustness of our method compared with the state-of-the-art. |
|
|
Address |
Barcelona; October 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
978-1-4503-1789-4 |
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
ACM-DocEng |
|
|
Notes |
DAG; 600.061 |
Approved |
no |
|
|
Call Number |
Admin @ si @ DTR2013a |
Serial |
2330 |
|
Permanent link to this record |
|
|
|
|
Author |
Alvaro Cepero; Albert Clapes; Sergio Escalera |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Quantitative analysis of non-verbal communication for competence analysis |
Type |
Conference Article |
|
Year |
2013 |
Publication |
16th Catalan Conference on Artificial Intelligence |
Abbreviated Journal |
|
|
|
Volume |
256 |
Issue |
|
Pages |
105-114 |
|
|
Keywords |
|
|
|
Abstract |
|
|
|
Address |
Vic; October 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
CCIA |
|
|
Notes |
HUPBA;MILAB |
Approved |
no |
|
|
Call Number |
Admin @ si @ CCE2013 |
Serial |
2324 |
|
Permanent link to this record |
|
|
|
|
Author |
Lluis Pere de las Heras; David Fernandez; Alicia Fornes; Ernest Valveny; Gemma Sanchez;Josep Llados |
![download PDF file pdf](img/file_PDF.gif)
|
|
Title |
Perceptual retrieval of architectural floor plans |
Type |
Conference Article |
|
Year |
2013 |
Publication |
10th IAPR International Workshop on Graphics Recognition |
Abbreviated Journal |
|
|
|
Volume |
|
Issue |
|
Pages |
|
|
|
Keywords |
|
|
|
Abstract |
This paper proposes a runlength histogram signature as a percetual descriptor of architectural plans in a retrieval scenario. The style of an architectural drawing is characterized by the perception of lines, shapes and texture. Such visual stimuli are the basis for defining semantic concepts as space properties, symmetry, density, etc. We propose runlength histograms extracted in vertical, horizontal and diagonal directions as a characterization of line and space properties in floorplans, so it can be roughly associated to a description of walls and room structure. A retrieval application illustrates the performance of the proposed approach, where given a plan as a query,
similar ones are obtained from a database. A ground truth based on human observation has been constructed to validate the hypothesis. Preliminary results show the interest of the proposed approach and opens a challenging research line in graphics recognition. |
|
|
Address |
Bethlehem; PA; USA; August 2013 |
|
|
Corporate Author |
|
Thesis |
|
|
|
Publisher |
|
Place of Publication |
|
Editor |
|
|
|
Language |
|
Summary Language |
|
Original Title |
|
|
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
|
|
Series Volume |
|
Series Issue |
|
Edition |
|
|
|
ISSN ![sorted by ISSN field, descending order (down)](img/sort_desc.gif) |
|
ISBN |
|
Medium |
|
|
|
Area |
|
Expedition |
|
Conference |
GREC |
|
|
Notes |
DAG; 600.045; 600.056; 600.061 |
Approved |
no |
|
|
Call Number |
Admin @ si @ HFF2013a |
Serial |
2320 |
|
Permanent link to this record |