Home | << 1 2 3 4 5 6 7 8 9 10 >> [11–12] |
Records | |||||
---|---|---|---|---|---|
Author | Isabel Guitart; Jordi Conesa; Luis Villarejo; Agata Lapedriza; David Masip; Antoni Perez; Elena Planas | ||||
Title | Opinion Mining on Educational Resources at the Open University of Catalonia | Type | Conference Article | ||
Year | 2013 | Publication | 3rd International Workshop on Adaptive Learning via Interactive, Collaborative and Emotional approaches. In conjunction with CISIS 2013: The 7th International Conference on Complex, Intelligent, and Software Intensive Systems | Abbreviated Journal | |
Volume | Issue | Pages | 385 - 390 | ||
Keywords | |||||
Abstract | In order to make improvements to teaching, it is vital to know what students think of the way they are taught. With that purpose in mind, exhaustively analyzing the forums associated with the subjects taught at the Universitat Oberta de Cataluya (UOC) would be extremely helpful, as the university's students often post comments on their learning experiences in them. Exploiting the content of such forums is not a simple undertaking. The volume of data involved is very large, and performing the task manually would require a great deal of effort from lecturers. As a first step to solve this problem, we propose a tool to automatically analyze the posts in forums of communities of UOC students and teachers, with a view to systematically mining the opinions they contain. This article defines the architecture of such tool and explains how lexical-semantic and language technology resources can be used to that end. For pilot testing purposes, the tool has been used to identify students' opinions on the UOC's Business Intelligence master's degree course during the last two years. The paper discusses the results of such test. The contribution of this paper is twofold. Firstly, it demonstrates the feasibility of using natural language parsing techniques to help teachers to make decisions. Secondly, it introduces a simple tool that can be refined and adapted to a virtual environment for the purpose in question. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-0-7695-4992-7 | Medium | ||
Area | Expedition | Conference | ALICE | ||
Notes | OR;MV | Approved | no | ||
Call Number | GCV2013 | Serial | 2268 | ||
Permanent link to this record | |||||
Author | Jose Manuel Alvarez; Theo Gevers; Ferran Diego; Antonio Lopez | ||||
Title | Road Geometry Classification by Adaptative Shape Models | Type | Journal Article | ||
Year | 2013 | Publication | IEEE Transactions on Intelligent Transportation Systems | Abbreviated Journal | TITS |
Volume | 14 | Issue | 1 | Pages | 459-468 |
Keywords | road detection | ||||
Abstract | Vision-based road detection is important for different applications in transportation, such as autonomous driving, vehicle collision warning, and pedestrian crossing detection. Common approaches to road detection are based on low-level road appearance (e.g., color or texture) and neglect of the scene geometry and context. Hence, using only low-level features makes these algorithms highly depend on structured roads, road homogeneity, and lighting conditions. Therefore, the aim of this paper is to classify road geometries for road detection through the analysis of scene composition and temporal coherence. Road geometry classification is proposed by building corresponding models from training images containing prototypical road geometries. We propose adaptive shape models where spatial pyramids are steered by the inherent spatial structure of road images. To reduce the influence of lighting variations, invariant features are used. Large-scale experiments show that the proposed road geometry classifier yields a high recognition rate of 73.57% ± 13.1, clearly outperforming other state-of-the-art methods. Including road shape information improves road detection results over existing appearance-based methods. Finally, it is shown that invariant features and temporal information provide robustness against disturbing imaging conditions. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1524-9050 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS;ISE | Approved | no | ||
Call Number | Admin @ si @ AGD2013;; ADAS @ adas @ | Serial | 2269 | ||
Permanent link to this record | |||||
Author | Rahat Khan; Joost Van de Weijer; Dimosthenis Karatzas; Damien Muselet | ||||
Title | Towards multispectral data acquisition with hand-held devices | Type | Conference Article | ||
Year | 2013 | Publication | 20th IEEE International Conference on Image Processing | Abbreviated Journal | |
Volume | Issue | Pages | 2053 - 2057 | ||
Keywords | Multispectral; mobile devices; color measurements | ||||
Abstract | We propose a method to acquire multispectral data with handheld devices with front-mounted RGB cameras. We propose to use the display of the device as an illuminant while the camera captures images illuminated by the red, green and
blue primaries of the display. Three illuminants and three response functions of the camera lead to nine response values which are used for reflectance estimation. Results are promising and show that the accuracy of the spectral reconstruction improves in the range from 30-40% over the spectral reconstruction based on a single illuminant. Furthermore, we propose to compute sensor-illuminant aware linear basis by discarding the part of the reflectances that falls in the sensorilluminant null-space. We show experimentally that optimizing reflectance estimation on these new basis functions decreases the RMSE significantly over basis functions that are independent to sensor-illuminant. We conclude that, multispectral data acquisition is potentially possible with consumer hand-held devices such as tablets, mobiles, and laptops, opening up applications which are currently considered to be unrealistic. |
||||
Address | Melbourne; Australia; September 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICIP | ||
Notes | CIC; DAG; 600.048 | Approved | no | ||
Call Number | Admin @ si @ KWK2013b | Serial | 2265 | ||
Permanent link to this record | |||||
Author | Shida Beigpour; Marc Serra; Joost Van de Weijer; Robert Benavente; Maria Vanrell; Olivier Penacchio; Dimitris Samaras | ||||
Title | Intrinsic Image Evaluation On Synthetic Complex Scenes | Type | Conference Article | ||
Year | 2013 | Publication | 20th IEEE International Conference on Image Processing | Abbreviated Journal | |
Volume | Issue | Pages | 285 - 289 | ||
Keywords | |||||
Abstract | Scene decomposition into its illuminant, shading, and reflectance intrinsic images is an essential step for scene understanding. Collecting intrinsic image groundtruth data is a laborious task. The assumptions on which the ground-truth
procedures are based limit their application to simple scenes with a single object taken in the absence of indirect lighting and interreflections. We investigate synthetic data for intrinsic image research since the extraction of ground truth is straightforward, and it allows for scenes in more realistic situations (e.g, multiple illuminants and interreflections). With this dataset we aim to motivate researchers to further explore intrinsic image decomposition in complex scenes. |
||||
Address | Melbourne; Australia; September 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICIP | ||
Notes | CIC; 600.048; 600.052; 600.051 | Approved | no | ||
Call Number | Admin @ si @ BSW2013 | Serial | 2264 | ||
Permanent link to this record | |||||
Author | Rahat Khan; Joost Van de Weijer; Fahad Shahbaz Khan; Damien Muselet; christophe Ducottet; Cecile Barat | ||||
Title | Discriminative Color Descriptors | Type | Conference Article | ||
Year | 2013 | Publication | IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 2866 - 2873 | ||
Keywords | |||||
Abstract | Color description is a challenging task because of large variations in RGB values which occur due to scene accidental events, such as shadows, shading, specularities, illuminant color changes, and changes in viewing geometry. Traditionally, this challenge has been addressed by capturing the variations in physics-based models, and deriving invariants for the undesired variations. The drawback of this approach is that sets of distinguishable colors in the original color space are mapped to the same value in the photometric invariant space. This results in a drop of discriminative power of the color description. In this paper we take an information theoretic approach to color description. We cluster color values together based on their discriminative power in a classification problem. The clustering has the explicit objective to minimize the drop of mutual information of the final representation. We show that such a color description automatically learns a certain degree of photometric invariance. We also show that a universal color representation, which is based on other data sets than the one at hand, can obtain competing performance. Experiments show that the proposed descriptor outperforms existing photometric invariants. Furthermore, we show that combined with shape description these color descriptors obtain excellent results on four challenging datasets, namely, PASCAL VOC 2007, Flowers-102, Stanford dogs-120 and Birds-200. | ||||
Address | Portland; Oregon; June 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1063-6919 | ISBN | Medium | ||
Area | Expedition | Conference | CVPR | ||
Notes | CIC; 600.048 | Approved | no | ||
Call Number | Admin @ si @ KWK2013a | Serial | 2262 | ||
Permanent link to this record | |||||
Author | Christophe Rigaud; Dimosthenis Karatzas; Joost Van de Weijer; Jean-Christophe Burie; Jean-Marc Ogier | ||||
Title | Automatic text localisation in scanned comic books | Type | Conference Article | ||
Year | 2013 | Publication | Proceedings of the International Conference on Computer Vision Theory and Applications | Abbreviated Journal | |
Volume | Issue | Pages | 814-819 | ||
Keywords | Text localization; comics; text/graphic separation; complex background; unstructured document | ||||
Abstract | Comic books constitute an important cultural heritage asset in many countries. Digitization combined with subsequent document understanding enable direct content-based search as opposed to metadata only search (e.g. album title or author name). Few studies have been done in this direction. In this work we detail a novel approach for the automatic text localization in scanned comics book pages, an essential step towards a fully automatic comics book understanding. We focus on speech text as it is semantically important and represents the majority of the text present in comics. The approach is compared with existing methods of text localization found in the literature and results are presented. | ||||
Address | Barcelona; February 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | VISAPP | ||
Notes | DAG; CIC; 600.056 | Approved | no | ||
Call Number | Admin @ si @ RKW2013b | Serial | 2261 | ||
Permanent link to this record | |||||
Author | Christophe Rigaud; Dimosthenis Karatzas; Joost Van de Weijer; Jean-Christophe Burie; Jean-Marc Ogier | ||||
Title | An active contour model for speech balloon detection in comics | Type | Conference Article | ||
Year | 2013 | Publication | 12th International Conference on Document Analysis and Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 1240-1244 | ||
Keywords | |||||
Abstract | Comic books constitute an important cultural heritage asset in many countries. Digitization combined with subsequent comic book understanding would enable a variety of new applications, including content-based retrieval and content retargeting. Document understanding in this domain is challenging as comics are semi-structured documents, combining semantically important graphical and textual parts. Few studies have been done in this direction. In this work we detail a novel approach for closed and non-closed speech balloon localization in scanned comic book pages, an essential step towards a fully automatic comic book understanding. The approach is compared with existing methods for closed balloon localization found in the literature and results are presented. | ||||
Address | washington; USA; August 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1520-5363 | ISBN | Medium | ||
Area | Expedition | Conference | ICDAR | ||
Notes | DAG; CIC; 600.056 | Approved | no | ||
Call Number | Admin @ si @ RKW2013a | Serial | 2260 | ||
Permanent link to this record | |||||
Author | Laura Igual; Xavier Baro | ||||
Title | Experiencia de aprendizaje de programación basada en proyectos. Simposio-Taller Estrategias y herramientas para el aprendizaje y la evaluación | Type | Miscellaneous | ||
Year | 2013 | Publication | Simposio-Taller Estrategias y herramientas para el aprendizaje y la evaluación, de las XIX Jornadas sobre la Enseñanza Universitaria de la Informática | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | JENUI | ||
Notes | OR;HuPBA;MV | Approved | no | ||
Call Number | Admin @ si @ IgB2013 | Serial | 2257 | ||
Permanent link to this record | |||||
Author | S.Grau; Anna Puig; Sergio Escalera; Maria Salamo; Oscar Amoros | ||||
Title | Efficient complementary viewpoint selection in volume rendering | Type | Conference Article | ||
Year | 2013 | Publication | 21st WSCG Conference on Computer Graphics, | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Dual camera; Visualization; Interactive Interfaces; Dynamic Time Warping. | ||||
Abstract | A major goal of visualization is to appropriately express knowledge of scientific data. Generally, gathering visual information contained in the volume data often requires a lot of expertise from the final user to setup the parameters of the visualization. One way of alleviating this problem is to provide the position of inner structures with different viewpoint locations to enhance the perception and construction of the mental image. To this end, traditional illustrations use two or three different views of the regions of interest. Similarly, with the aim of assisting the users to easily place a good viewpoint location, this paper proposes an automatic and interactive method that locates different complementary viewpoints from a reference camera in volume datasets. Specifically, the proposed method combines the quantity of information each camera provides for each structure and the shape similarity of the projections of the remaining viewpoints based on Dynamic Time Warping. The selected complementary viewpoints allow a better understanding of the focused structure in several applications. Thus, the user interactively receives feedback based on several viewpoints that helps him to understand the visual information. A live-user evaluation on different data sets show a good convergence to useful complementary viewpoints. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-808694374-9 | Medium | ||
Area | Expedition | Conference | WSCG | ||
Notes | HuPBA; 600.046;MILAB | Approved | no | ||
Call Number | Admin @ si @ GPE2013a | Serial | 2255 | ||
Permanent link to this record | |||||
Author | Santiago Segui; Laura Igual; Jordi Vitria | ||||
Title | Bagged One Class Classifiers in the Presence of Outliers | Type | Journal Article | ||
Year | 2013 | Publication | International Journal of Pattern Recognition and Artificial Intelligence | Abbreviated Journal | IJPRAI |
Volume | 27 | Issue | 5 | Pages | 1350014-1350035 |
Keywords | One-class Classifier; Ensemble Methods; Bagging and Outliers | ||||
Abstract | The problem of training classifiers only with target data arises in many applications where non-target data are too costly, difficult to obtain, or not available at all. Several one-class classification methods have been presented to solve this problem, but most of the methods are highly sensitive to the presence of outliers in the target class. Ensemble methods have therefore been proposed as a powerful way to improve the classification performance of binary/multi-class learning algorithms by introducing diversity into classifiers.
However, their application to one-class classification has been rather limited. In this paper, we present a new ensemble method based on a non-parametric weighted bagging strategy for one-class classification, to improve accuracy in the presence of outliers. While the standard bagging strategy assumes a uniform data distribution, the method we propose here estimates a probability density based on a forest structure of the data. This assumption allows the estimation of data distribution from the computation of simple univariate and bivariate kernel densities. Experiments using original and noisy versions of 20 different datasets show that bagging ensemble methods applied to different one-class classifiers outperform base one-class classification methods. Moreover, we show that, in noisy versions of the datasets, the non-parametric weighted bagging strategy we propose outperforms the classical bagging strategy in a statistically significant way. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | OR; 600.046;MV | Approved | no | ||
Call Number | Admin @ si @ SIV2013 | Serial | 2256 | ||
Permanent link to this record | |||||
Author | Vitaliy Konovalov; Albert Clapes; Sergio Escalera | ||||
Title | Automatic Hand Detection in RGB-Depth Data Sequences | Type | Conference Article | ||
Year | 2013 | Publication | 16th Catalan Conference on Artificial Intelligence | Abbreviated Journal | |
Volume | Issue | Pages | 91-100 | ||
Keywords | |||||
Abstract | Detecting hands in multi-modal RGB-Depth visual data has become a challenging Computer Vision problem with several applications of interest. This task involves dealing with changes in illumination, viewpoint variations, the articulated nature of the human body, the high flexibility of the wrist articulation, and the deformability of the hand itself. In this work, we propose an accurate and efficient automatic hand detection scheme to be applied in Human-Computer Interaction (HCI) applications in which the user is seated at the desk and, thus, only the upper body is visible. Our main hypothesis is that hand landmarks remain at a nearly constant geodesic distance from an automatically located anatomical reference point.
In a given frame, the human body is segmented first in the depth image. Then, a graph representation of the body is built in which the geodesic paths are computed from the reference point. The dense optical flow vectors on the corresponding RGB image are used to reduce ambiguities of the geodesic paths’ connectivity, allowing to eliminate false edges interconnecting different body parts. Finally, we are able to detect the position of both hands based on invariant geodesic distances and optical flow within the body region, without involving costly learning procedures. |
||||
Address | Vic; October 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CCIA | ||
Notes | HuPBA;MILAB | Approved | no | ||
Call Number | Admin @ si @ KCE2013 | Serial | 2323 | ||
Permanent link to this record | |||||
Author | Andreas Møgelmose; Chris Bahnsen; Thomas B. Moeslund; Albert Clapes; Sergio Escalera | ||||
Title | Tri-modal Person Re-identification with RGB, Depth and Thermal Features | Type | Conference Article | ||
Year | 2013 | Publication | 9th IEEE Workshop on Perception beyond the visible Spectrum, Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 301-307 | ||
Keywords | |||||
Abstract | Person re-identification is about recognizing people who have passed by a sensor earlier. Previous work is mainly based on RGB data, but in this work we for the first time present a system where we combine RGB, depth, and thermal data for re-identification purposes. First, from each of the three modalities, we obtain some particular features: from RGB data, we model color information from different regions of the body, from depth data, we compute different soft body biometrics, and from thermal data, we extract local structural information. Then, the three information types are combined in a joined classifier. The tri-modal system is evaluated on a new RGB-D-T dataset, showing successful results in re-identification scenarios. | ||||
Address | Portland; oregon; June 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-0-7695-4990-3 | Medium | ||
Area | Expedition | Conference | CVPRW | ||
Notes | HUPBA;MILAB | Approved | no | ||
Call Number | Admin @ si @ MBM2013 | Serial | 2253 | ||
Permanent link to this record | |||||
Author | Fernando Barrera; Felipe Lumbreras; Angel Sappa | ||||
Title | Multispectral Piecewise Planar Stereo using Manhattan-World Assumption | Type | Journal Article | ||
Year | 2013 | Publication | Pattern Recognition Letters | Abbreviated Journal | PRL |
Volume | 34 | Issue | 1 | Pages | 52-61 |
Keywords | Multispectral stereo rig; Dense disparity maps from multispectral stereo; Color and infrared images | ||||
Abstract | This paper proposes a new framework for extracting dense disparity maps from a multispectral stereo rig. The system is constructed with an infrared and a color camera. It is intended to explore novel multispectral stereo matching approaches that will allow further extraction of semantic information. The proposed framework consists of three stages. Firstly, an initial sparse disparity map is generated by using a cost function based on feature matching in a multiresolution scheme. Then, by looking at the color image, a set of planar hypotheses is defined to describe the surfaces on the scene. Finally, the previous stages are combined by reformulating the disparity computation as a global minimization problem. The paper has two main contributions. The first contribution combines mutual information with a shape descriptor based on gradient in a multiresolution scheme. The second contribution, which is based on the Manhattan-world assumption, extracts a dense disparity representation using the graph cut algorithm. Experimental results in outdoor scenarios are provided showing the validity of the proposed framework. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS; 600.054; 600.055; 605.203 | Approved | no | ||
Call Number | Admin @ si @ BLS2013 | Serial | 2245 | ||
Permanent link to this record | |||||
Author | Olivier Penacchio; Xavier Otazu; Laura Dempere-Marco | ||||
Title | A Neurodynamical Model of Brightness Induction in V1 | Type | Journal Article | ||
Year | 2013 | Publication | PloS ONE | Abbreviated Journal | Plos |
Volume | 8 | Issue | 5 | Pages | e64086 |
Keywords | |||||
Abstract | Brightness induction is the modulation of the perceived intensity of an area by the luminance of surrounding areas. Recent neurophysiological evidence suggests that brightness information might be explicitly represented in V1, in contrast to the more common assumption that the striate cortex is an area mostly responsive to sensory information. Here we investigate possible neural mechanisms that offer a plausible explanation for such phenomenon. To this end, a neurodynamical model which is based on neurophysiological evidence and focuses on the part of V1 responsible for contextual influences is presented. The proposed computational model successfully accounts for well known psychophysical effects for static contexts and also for brightness induction in dynamic contexts defined by modulating the luminance of surrounding areas. This work suggests that intra-cortical interactions in V1 could, at least partially, explain brightness induction effects and reveals how a common general architecture may account for several different fundamental processes, such as visual saliency and brightness induction, which emerge early in the visual processing pathway. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | CIC | Approved | no | ||
Call Number | Admin @ si @ POD2013 | Serial | 2242 | ||
Permanent link to this record | |||||
Author | Alicia Fornes; Xavier Otazu; Josep Llados | ||||
Title | Show through cancellation and image enhancement by multiresolution contrast processing | Type | Conference Article | ||
Year | 2013 | Publication | 12th International Conference on Document Analysis and Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 200-204 | ||
Keywords | |||||
Abstract | Historical documents suffer from different types of degradation and noise such as background variation, uneven illumination or dark spots. In case of double-sided documents, another common problem is that the back side of the document usually interferes with the front side because of the transparency of the document or ink bleeding. This effect is called the show through phenomenon. Many methods are developed to solve these problems, and in the case of show-through, by scanning and matching both the front and back sides of the document. In contrast, our approach is designed to use only one side of the scanned document. We hypothesize that show-trough are low contrast components, while foreground components are high contrast ones. A Multiresolution Contrast (MC) decomposition is presented in order to estimate the contrast of features at different spatial scales. We cancel the show-through phenomenon by thresholding these low contrast components. This decomposition is also able to enhance the image removing shadowed areas by weighting spatial scales. Results show that the enhanced images improve the readability of the documents, allowing scholars both to recover unreadable words and to solve ambiguities. | ||||
Address | Washington; USA; August 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1520-5363 | ISBN | Medium | ||
Area | Expedition | Conference | ICDAR | ||
Notes | DAG; 602.006; 600.045; 600.061; 600.052;CIC | Approved | no | ||
Call Number | Admin @ si @ FOL2013 | Serial | 2241 | ||
Permanent link to this record |