Home | [111–120] << 121 122 123 124 125 126 127 128 129 130 >> [131–140] |
![]() |
Records | |||||
---|---|---|---|---|---|
Author | Javier Marin; David Vazquez; Antonio Lopez; Jaume Amores; Ludmila I. Kuncheva | ||||
Title | Occlusion handling via random subspace classifiers for human detection | Type | Journal Article | ||
Year | 2014 | Publication | IEEE Transactions on Systems, Man, and Cybernetics (Part B) | Abbreviated Journal | TSMCB |
Volume | 44 | Issue | 3 | Pages | 342-354 |
Keywords | Pedestriand Detection; occlusion handling | ||||
Abstract | This paper describes a general method to address partial occlusions for human detection in still images. The Random Subspace Method (RSM) is chosen for building a classifier ensemble robust against partial occlusions. The component classifiers are chosen on the basis of their individual and combined performance. The main contribution of this work lies in our approach’s capability to improve the detection rate when partial occlusions are present without compromising the detection performance on non occluded data. In contrast to many recent approaches, we propose a method which does not require manual labelling of body parts, defining any semantic spatial components, or using additional data coming from motion or stereo. Moreover, the method can be easily extended to other object classes. The experiments are performed on three large datasets: the INRIA person dataset, the Daimler Multicue dataset, and a new challenging dataset, called PobleSec, in which a considerable number of targets are partially occluded. The different approaches are evaluated at the classification and detection levels for both partially occluded and non-occluded data. The experimental results show that our detector outperforms state-of-the-art approaches in the presence of partial occlusions, while offering performance and reliability similar to those of the holistic approach on non-occluded data. The datasets used in our experiments have been made publicly available for benchmarking purposes | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 2168-2267 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS; 605.203; 600.057; 600.054; 601.042; 601.187; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ MVL2014 | Serial | 2213 | ||
Permanent link to this record | |||||
Author | Jiaolong Xu; David Vazquez; Antonio Lopez; Javier Marin; Daniel Ponsa | ||||
Title | Learning a Multiview Part-based Model in Virtual World for Pedestrian Detection | Type | Conference Article | ||
Year | 2013 | Publication | IEEE Intelligent Vehicles Symposium | Abbreviated Journal | |
Volume | Issue | Pages | 467 - 472 | ||
Keywords | Pedestrian Detection; Virtual World; Part based | ||||
Abstract | State-of-the-art deformable part-based models based on latent SVM have shown excellent results on human detection. In this paper, we propose to train a multiview deformable part-based model with automatically generated part examples from virtual-world data. The method is efficient as: (i) the part detectors are trained with precisely extracted virtual examples, thus no latent learning is needed, (ii) the multiview pedestrian detector enhances the performance of the pedestrian root model, (iii) a top-down approach is used for part detection which reduces the searching space. We evaluate our model on Daimler and Karlsruhe Pedestrian Benchmarks with publicly available Caltech pedestrian detection evaluation framework and the result outperforms the state-of-the-art latent SVM V4.0, on both average miss rate and speed (our detector is ten times faster). | ||||
Address | Gold Coast; Australia; June 2013 | ||||
Corporate Author | Thesis ![]() |
||||
Publisher | IEEE | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1931-0587 | ISBN | 978-1-4673-2754-1 | Medium | |
Area | Expedition | Conference | IV | ||
Notes | ADAS; 600.054; 600.057 | Approved | no | ||
Call Number | XVL2013; ADAS @ adas @ xvl2013a | Serial | 2214 | ||
Permanent link to this record | |||||
Author | Juan Ramon Terven Salinas; Joaquin Salas; Bogdan Raducanu | ||||
Title | Estado del Arte en Sistemas de Vision Artificial para Personas Invidentes | Type | Journal | ||
Year | 2013 | Publication | Komputer Sapiens | Abbreviated Journal | KS |
Volume | 1 | Issue | Pages | 20-25 | |
Keywords | |||||
Abstract | |||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | OR;MV | Approved | no | ||
Call Number | Admin @ si @ TSR2013 | Serial | 2231 | ||
Permanent link to this record | |||||
Author | Patricia Marquez; Debora Gil; Aura Hernandez-Sabate; Daniel Kondermann | ||||
Title | When Is A Confidence Measure Good Enough? | Type | Conference Article | ||
Year | 2013 | Publication | 9th International Conference on Computer Vision Systems | Abbreviated Journal | |
Volume | 7963 | Issue | Pages | 344-353 | |
Keywords | Optical flow, confidence measure, performance evaluation | ||||
Abstract | Confidence estimation has recently become a hot topic in image processing and computer vision.Yet, several definitions exist of the term “confidence” which are sometimes used interchangeably. This is a position paper, in which we aim to give an overview on existing definitions,
thereby clarifying the meaning of the used terms to facilitate further research in this field. Based on these clarifications, we develop a theory to compare confidence measures with respect to their quality. |
||||
Address | St Petersburg; Russia; July 2013 | ||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Link | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-39401-0 | Medium | |
Area | Expedition | Conference | ICVS | ||
Notes | IAM;ADAS; 600.044; 600.057; 600.060; 601.145 | Approved | no | ||
Call Number | IAM @ iam @ MGH2013a | Serial | 2218 | ||
Permanent link to this record | |||||
Author | David Vazquez; Jiaolong Xu; Sebastian Ramos; Antonio Lopez; Daniel Ponsa | ||||
Title | Weakly Supervised Automatic Annotation of Pedestrian Bounding Boxes | Type | Conference Article | ||
Year | 2013 | Publication | CVPR Workshop on Ground Truth – What is a good dataset? | Abbreviated Journal | |
Volume | Issue | Pages | 706 - 711 | ||
Keywords | Pedestrian Detection; Domain Adaptation | ||||
Abstract | Among the components of a pedestrian detector, its trained pedestrian classifier is crucial for achieving the desired performance. The initial task of the training process consists in collecting samples of pedestrians and background, which involves tiresome manual annotation of pedestrian bounding boxes (BBs). Thus, recent works have assessed the use of automatically collected samples from photo-realistic virtual worlds. However, learning from virtual-world samples and testing in real-world images may suffer the dataset shift problem. Accordingly, in this paper we assess an strategy to collect samples from the real world and retrain with them, thus avoiding the dataset shift, but in such a way that no BBs of real-world pedestrians have to be provided. In particular, we train a pedestrian classifier based on virtual-world samples (no human annotation required). Then, using such a classifier we collect pedestrian samples from real-world images by detection. After, a human oracle rejects the false detections efficiently (weak annotation). Finally, a new classifier is trained with the accepted detections. We show that this classifier is competitive with respect to the counterpart trained with samples collected by manually annotating hundreds of pedestrian BBs. | ||||
Address | Portland; Oregon; June 2013 | ||||
Corporate Author | Thesis ![]() |
||||
Publisher | IEEE | Place of Publication | Editor | ||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CVPRW | ||
Notes | ADAS; 600.054; 600.057; 601.217 | Approved | no | ||
Call Number | ADAS @ adas @ VXR2013a | Serial | 2219 | ||
Permanent link to this record | |||||
Author | Jiaolong Xu; David Vazquez; Sebastian Ramos; Antonio Lopez; Daniel Ponsa | ||||
Title | Adapting a Pedestrian Detector by Boosting LDA Exemplar Classifiers | Type | Conference Article | ||
Year | 2013 | Publication | CVPR Workshop on Ground Truth – What is a good dataset? | Abbreviated Journal | |
Volume | Issue | Pages | 688 - 693 | ||
Keywords | Pedestrian Detection; Domain Adaptation | ||||
Abstract | Training vision-based pedestrian detectors using synthetic datasets (virtual world) is a useful technique to collect automatically the training examples with their pixel-wise ground truth. However, as it is often the case, these detectors must operate in real-world images, experiencing a significant drop of their performance. In fact, this effect also occurs among different real-world datasets, i.e. detectors' accuracy drops when the training data (source domain) and the application scenario (target domain) have inherent differences. Therefore, in order to avoid this problem, it is required to adapt the detector trained with synthetic data to operate in the real-world scenario. In this paper, we propose a domain adaptation approach based on boosting LDA exemplar classifiers from both virtual and real worlds. We evaluate our proposal on multiple real-world pedestrian detection datasets. The results show that our method can efficiently adapt the exemplar classifiers from virtual to real world, avoiding drops in average precision over the 15%. | ||||
Address | Portland; oregon; June 2013 | ||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CVPRW | ||
Notes | ADAS; 600.054; 600.057; 601.217 | Approved | yes | ||
Call Number | XVR2013; ADAS @ adas @ xvr2013a | Serial | 2220 | ||
Permanent link to this record | |||||
Author | Thierry Brouard; Jordi Gonzalez; Caifeng Shan; Massimo Piccardi; Larry S. Davis | ||||
Title | Special issue on background modeling for foreground detection in real-world dynamic scenes | Type | Journal Article | ||
Year | 2014 | Publication | Machine Vision and Applications | Abbreviated Journal | MVAP |
Volume | 25 | Issue | 5 | Pages | 1101-1103 |
Keywords | |||||
Abstract | Although background modeling and foreground detection are not mandatory steps for computer vision applications, they may prove useful as they separate the primal objects usually called “foreground” from the remaining part of the scene called “background”, and permits different algorithmic treatment in the video processing field such as video surveillance, optical motion capture, multimedia applications, teleconferencing and human–computer interfaces. Conventional background modeling methods exploit the temporal variation of each pixel to model the background, and the foreground detection is made using change detection. The last decade witnessed very significant publications on background modeling but recently new applications in which background is not static, such as recordings taken from mobile devices or Internet videos, need new developments to detect robustly moving objects in challenging environments. Thus, effective methods for robustness to deal both with dynamic backgrounds, i | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0932-8092 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ISE; 600.078 | Approved | no | ||
Call Number | BGS2014a | Serial | 2411 | ||
Permanent link to this record | |||||
Author | Joan Mas; Gemma Sanchez; Josep Llados | ||||
Title | SSP: Sketching slide Presentations, a Syntactic Approach | Type | Book Chapter | ||
Year | 2010 | Publication | Graphics Recognition. Achievements, Challenges, and Evolution. 8th International Workshop, GREC 2009. Selected Papers | Abbreviated Journal | |
Volume | 6020 | Issue | Pages | 118-129 | |
Keywords | |||||
Abstract | The design of a slide presentation is a creative process. In this process first, humans visualize in their minds what they want to explain. Then, they have to be able to represent this knowledge in an understandable way. There exists a lot of commercial software that allows to create our own slide presentations but the creativity of the user is rather limited. In this article we present an application that allows the user to create and visualize a slide presentation from a sketch. A slide may be seen as a graphical document or a diagram where its elements are placed in a particular spatial arrangement. To describe and recognize slides a syntactic approach is proposed. This approach is based on an Adjacency Grammar and a parsing methodology to cope with this kind of grammars. The experimental evaluation shows the performance of our methodology from a qualitative and a quantitative point of view. Six different slides containing different number of symbols, from 4 to 7, have been given to the users and they have drawn them without restrictions in the order of the elements. The quantitative results give an idea on how suitable is our methodology to describe and recognize the different elements in a slide. | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-13727-3 | Medium | |
Area | Expedition | Conference | GREC | ||
Notes | DAG | Approved | no | ||
Call Number | MSL2010 | Serial | 2405 | ||
Permanent link to this record | |||||
Author | Mathieu Nicolas Delalandre; Jean-Yves Ramel; Ernest Valveny; Muhammad Muzzamil Luqman | ||||
Title | A Performance Characterization Algorithm for Symbol Localization | Type | Book Chapter | ||
Year | 2010 | Publication | Graphics Recognition. Achievements, Challenges, and Evolution. 8th International Workshop, GREC 2009. Selected Papers | Abbreviated Journal | |
Volume | 6020 | Issue | Pages | 260–271 | |
Keywords | |||||
Abstract | In this paper we present an algorithm for performance characterization of symbol localization systems. This algorithm is aimed to be a more “reliable” and “open” solution to characterize the performance. To achieve that, it exploits only single points as the result of localization and offers the possibility to reconsider the localization results provided by a system. We use the information about context in groundtruth, and overall localization results, to detect the ambiguous localization results. A probability score is computed for each matching between a localization point and a groundtruth region, depending on the spatial distribution of the other regions in the groundtruth. Final characterization is given with detection rate/probability score plots, describing the sets of possible interpretations of the localization results, according to a given confidence rate. We present experimentation details along with the results for the symbol localization system of [1], exploiting a synthetic dataset of architectural floorplans and electrical diagrams (composed of 200 images and 3861 symbols). | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-13727-3 | Medium | |
Area | Expedition | Conference | GREC | ||
Notes | DAG | Approved | no | ||
Call Number | Admin @ si @ DRV2010 | Serial | 2406 | ||
Permanent link to this record | |||||
Author | Marçal Rusiñol; K. Bertet; Jean-Marc Ogier; Josep Llados | ||||
Title | Symbol Recognition Using a Concept Lattice of Graphical Patterns | Type | Book Chapter | ||
Year | 2010 | Publication | Graphics Recognition. Achievements, Challenges, and Evolution. 8th International Workshop, GREC 2009. Selected Papers | Abbreviated Journal | |
Volume | 6020 | Issue | Pages | 187-198 | |
Keywords | |||||
Abstract | In this paper we propose a new approach to recognize symbols by the use of a concept lattice. We propose to build a concept lattice in terms of graphical patterns. Each model symbol is decomposed in a set of composing graphical patterns taken as primitives. Each one of these primitives is described by boundary moment invariants. The obtained concept lattice relates which symbolic patterns compose a given graphical symbol. A Hasse diagram is derived from the context and is used to recognize symbols affected by noise. We present some preliminary results over a variation of the dataset of symbols from the GREC 2005 symbol recognition contest. | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-13727-3 | Medium | |
Area | Expedition | Conference | |||
Notes | DAG | Approved | no | ||
Call Number | Admin @ si @ RBO2010 | Serial | 2407 | ||
Permanent link to this record | |||||
Author | Partha Pratim Roy; Umapada Pal; Josep Llados | ||||
Title | Touching Text Character Localization in Graphical Documents using SIFT | Type | Book Chapter | ||
Year | 2010 | Publication | Graphics Recognition. Achievements, Challenges, and Evolution. 8th International Workshop, GREC 2009. Selected Papers | Abbreviated Journal | |
Volume | 6020 | Issue | Pages | 199-211 | |
Keywords | Support Vector Machine; Text Component; Graphical Line; Document Image; Scale Invariant Feature Transform | ||||
Abstract | Interpretation of graphical document images is a challenging task as it requires proper understanding of text/graphics symbols present in such documents. Difficulties arise in graphical document recognition when text and symbol overlapped/touched. Intersection of text and symbols with graphical lines and curves occur frequently in graphical documents and hence separation of such symbols is very difficult.
Several pattern recognition and classification techniques exist to recognize isolated text/symbol. But, the touching/overlapping text and symbol recognition has not yet been dealt successfully. An interesting technique, Scale Invariant Feature Transform (SIFT), originally devised for object recognition can take care of overlapping problems. Even if SIFT features have emerged as a very powerful object descriptors, their employment in graphical documents context has not been investigated much. In this paper we present the adaptation of the SIFT approach in the context of text character localization (spotting) in graphical documents. We evaluate the applicability of this technique in such documents and discuss the scope of improvement by combining some state-of-the-art approaches. |
||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-13727-3 | Medium | |
Area | Expedition | Conference | |||
Notes | DAG | Approved | no | ||
Call Number | Admin @ si @ RPL2010c | Serial | 2408 | ||
Permanent link to this record | |||||
Author | Bhaskar Chakraborty; Jordi Gonzalez; Xavier Roca | ||||
Title | Large scale continuous visual event recognition using max-margin Hough transformation framework | Type | Journal Article | ||
Year | 2013 | Publication | Computer Vision and Image Understanding | Abbreviated Journal | CVIU |
Volume | 117 | Issue | 10 | Pages | 1356–1368 |
Keywords | |||||
Abstract | In this paper we propose a novel method for continuous visual event recognition (CVER) on a large scale video dataset using max-margin Hough transformation framework. Due to high scalability, diverse real environmental state and wide scene variability direct application of action recognition/detection methods such as spatio-temporal interest point (STIP)-local feature based technique, on the whole dataset is practically infeasible. To address this problem, we apply a motion region extraction technique which is based on motion segmentation and region clustering to identify possible candidate “event of interest” as a preprocessing step. On these candidate regions a STIP detector is applied and local motion features are computed. For activity representation we use generalized Hough transform framework where each feature point casts a weighted vote for possible activity class centre. A max-margin frame work is applied to learn the feature codebook weight. For activity detection, peaks in the Hough voting space are taken into account and initial event hypothesis is generated using the spatio-temporal information of the participating STIPs. For event recognition a verification Support Vector Machine is used. An extensive evaluation on benchmark large scale video surveillance dataset (VIRAT) and as well on a small scale benchmark dataset (MSR) shows that the proposed method is applicable on a wide range of continuous visual event recognition applications having extremely challenging conditions. | ||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1077-3142 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ISE | Approved | no | ||
Call Number | Admin @ si @ CGR2013 | Serial | 2413 | ||
Permanent link to this record | |||||
Author | Adria Ruiz; Joost Van de Weijer; Xavier Binefa | ||||
Title | Regularized Multi-Concept MIL for weakly-supervised facial behavior categorization | Type | Conference Article | ||
Year | 2014 | Publication | 25th British Machine Vision Conference | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | We address the problem of estimating high-level semantic labels for videos of recorded people by means of analysing their facial expressions. This problem, to which we refer as facial behavior categorization, is a weakly-supervised learning problem where we do not have access to frame-by-frame facial gesture annotations but only weak-labels at the video level are available. Therefore, the goal is to learn a set of discriminative expressions and how they determine the video weak-labels. Facial behavior categorization can be posed as a Multi-Instance-Learning (MIL) problem and we propose a novel MIL method called Regularized Multi-Concept MIL to solve it. In contrast to previous approaches applied in facial behavior analysis, RMC-MIL follows a Multi-Concept assumption which allows different facial expressions (concepts) to contribute differently to the video-label. Moreover, to handle with the high-dimensional nature of facial-descriptors, RMC-MIL uses a discriminative approach to model the concepts and structured sparsity regularization to discard non-informative features. RMC-MIL is posed as a convex-constrained optimization problem where all the parameters are jointly learned using the Projected-Quasi-Newton method. In our experiments, we use two public data-sets to show the advantages of the Regularized Multi-Concept approach and its improvement compared to existing MIL methods. RMC-MIL outperforms state-of-the-art results in the UNBC data-set for pain detection. | ||||
Address | Nottingham; UK; September 2014 | ||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | BMVC | ||
Notes | LAMP; CIC; 600.074; 600.079 | Approved | no | ||
Call Number | Admin @ si @ RWB2014 | Serial | 2508 | ||
Permanent link to this record | |||||
Author | T.Chauhan; E.Perales; Kaida Xiao; E.Hird ; Dimosthenis Karatzas; Sophie Wuerger | ||||
Title | The achromatic locus: Effect of navigation direction in color space | Type | Journal Article | ||
Year | 2014 | Publication | Journal of Vision | Abbreviated Journal | VSS |
Volume | 14 (1) | Issue | 25 | Pages | 1-11 |
Keywords | achromatic; unique hues; color constancy; luminance; color space | ||||
Abstract | 5Y Impact Factor: 2.99 / 1st (Ophthalmology)
An achromatic stimulus is defined as a patch of light that is devoid of any hue. This is usually achieved by asking observers to adjust the stimulus such that it looks neither red nor green and at the same time neither yellow nor blue. Despite the theoretical and practical importance of the achromatic locus, little is known about the variability in these settings. The main purpose of the current study was to evaluate whether achromatic settings were dependent on the task of the observers, namely the navigation direction in color space. Observers could either adjust the test patch along the two chromatic axes in the CIE u*v* diagram or, alternatively, navigate along the unique-hue lines. Our main result is that the navigation method affects the reliability of these achromatic settings. Observers are able to make more reliable achromatic settings when adjusting the test patch along the directions defined by the four unique hues as opposed to navigating along the main axes in the commonly used CIE u*v* chromaticity plane. This result holds across different ambient viewing conditions (Dark, Daylight, Cool White Fluorescent) and different test luminance levels (5, 20, and 50 cd/m2). The reduced variability in the achromatic settings is consistent with the idea that internal color representations are more aligned with the unique-hue lines than the u* and v* axes. |
||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | DAG; 600.077 | Approved | no | ||
Call Number | Admin @ si @ CPX2014 | Serial | 2418 | ||
Permanent link to this record | |||||
Author | Antonio Clavelli; Dimosthenis Karatzas; Josep Llados; Mario Ferraro; Giuseppe Boccignone | ||||
Title | Modelling task-dependent eye guidance to objects in pictures | Type | Journal Article | ||
Year | 2014 | Publication | Cognitive Computation | Abbreviated Journal | CoCom |
Volume | 6 | Issue | 3 | Pages | 558-584 |
Keywords | Visual attention; Gaze guidance; Value; Payoff; Stochastic fixation prediction | ||||
Abstract | 5Y Impact Factor: 1.14 / 3rd (Computer Science, Artificial Intelligence)
We introduce a model of attentional eye guidance based on the rationale that the deployment of gaze is to be considered in the context of a general action-perception loop relying on two strictly intertwined processes: sensory processing, depending on current gaze position, identifies sources of information that are most valuable under the given task; motor processing links such information with the oculomotor act by sampling the next gaze position and thus performing the gaze shift. In such a framework, the choice of where to look next is task-dependent and oriented to classes of objects embedded within pictures of complex scenes. The dependence on task is taken into account by exploiting the value and the payoff of gazing at certain image patches or proto-objects that provide a sparse representation of the scene objects. The different levels of the action-perception loop are represented in probabilistic form and eventually give rise to a stochastic process that generates the gaze sequence. This way the model also accounts for statistical properties of gaze shifts such as individual scan path variability. Results of the simulations are compared either with experimental data derived from publicly available datasets and from our own experiments. |
||||
Address | |||||
Corporate Author | Thesis ![]() |
||||
Publisher | Springer US | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1866-9956 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | DAG; 600.056; 600.045; 605.203; 601.212; 600.077 | Approved | no | ||
Call Number | Admin @ si @ CKL2014 | Serial | 2419 | ||
Permanent link to this record |