Records |
Author |
Marc Castello; Jordi Gonzalez; Ariel Amato; Pau Baiget; Carles Fernandez; Josep M. Gonfaus; Ramon Mollineda; Marco Pedersoli; Nicolas Perez de la Blanca; Xavier Roca |
Title |
Exploiting Multimodal Interaction Techniques for Video-Surveillance |
Type |
Book Chapter |
Year |
2013 |
Publication |
Multimodal Interaction in Image and Video Applications Intelligent Systems Reference Library |
Abbreviated Journal |
|
Volume |
48 |
Issue |
8 |
Pages |
135-151 |
Keywords |
|
Abstract |
In this paper we present an example of a video surveillance application that exploits Multimodal Interactive (MI) technologies. The main objective of the so-called VID-Hum prototype was to develop a cognitive artificial system for both the detection and description of a particular set of human behaviours arising from real-world events. The main procedure of the prototype described in this chapter entails: (i) adaptation, since the system adapts itself to the most common behaviours (qualitative data) inferred from tracking (quantitative data) thus being able to recognize abnormal behaviors; (ii) feedback, since an advanced interface based on Natural Language understanding allows end-users the communicationwith the prototype by means of conceptual sentences; and (iii) multimodality, since a virtual avatar has been designed to describe what is happening in the scene, based on those textual interpretations generated by the prototype. Thus, the MI methodology has provided an adequate framework for all these cooperating processes. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer Berlin Heidelberg |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1868-4394 |
ISBN |
978-3-642-35931-6 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
ISE; 605.203; 600.049 |
Approved |
no |
Call Number |
CGA2013 |
Serial |
2222 |
Permanent link to this record |
|
|
|
Author |
David Vazquez; Antonio Lopez; Daniel Ponsa; David Geronimo |
Title |
Interactive Training of Human Detectors |
Type |
Book Chapter |
Year |
2013 |
Publication |
Multiodal Interaction in Image and Video Applications |
Abbreviated Journal |
|
Volume |
48 |
Issue |
|
Pages |
169-182 |
Keywords |
Pedestrian Detection; Virtual World; AdaBoost; Domain Adaptation |
Abstract |
Image based human detection remains as a challenging problem. Most promising detectors rely on classifiers trained with labelled samples. However, labelling is a manual labor intensive step. To overcome this problem we propose to collect images of pedestrians from a virtual city, i.e., with automatic labels, and train a pedestrian detector with them, which works fine when such virtual-world data are similar to testing one, i.e., real-world pedestrians in urban areas. When testing data is acquired in different conditions than training one, e.g., human detection in personal photo albums, dataset shift appears. In previous work, we cast this problem as one of domain adaptation and solve it with an active learning procedure. In this work, we focus on the same problem but evaluating a different set of faster to compute features, i.e., Haar, EOH and their combination. In particular, we train a classifier with virtual-world data, using such features and Real AdaBoost as learning machine. This classifier is applied to real-world training images. Then, a human oracle interactively corrects the wrong detections, i.e., few miss detections are manually annotated and some false ones are pointed out too. A low amount of manual annotation is fixed as restriction. Real- and virtual-world difficult samples are combined within what we call cool world and we retrain the classifier with this data. Our experiments show that this adapted classifier is equivalent to the one trained with only real-world data but requiring 90% less manual annotations. |
Address |
Springer Heidelberg New York Dordrecht London |
Corporate Author |
|
Thesis |
|
Publisher |
Springer Berlin Heidelberg |
Place of Publication |
|
Editor |
|
Language |
English |
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1868-4394 |
ISBN |
978-3-642-35931-6 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
ADAS; 600.057; 600.054; 605.203 |
Approved |
no |
Call Number |
VLP2013; ADAS @ adas @ vlp2013 |
Serial |
2193 |
Permanent link to this record |
|
|
|
Author |
Angel Sappa; Jordi Vitria |
Title |
Multimodal Interaction in Image and Video Applications |
Type |
Book Whole |
Year |
2013 |
Publication |
Multimodal Interaction in Image and Video Applications |
Abbreviated Journal |
|
Volume |
48 |
Issue |
|
Pages |
|
Keywords |
|
Abstract |
Book Series Intelligent Systems Reference Library |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer Berlin Heidelberg |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1868-4394 |
ISBN |
978-3-642-35931-6 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
ADAS; OR;MV |
Approved |
no |
Call Number |
Admin @ si @ SaV2013 |
Serial |
2199 |
Permanent link to this record |
|
|
|
Author |
Jiaolong Xu; David Vazquez; Antonio Lopez; Javier Marin; Daniel Ponsa |
Title |
Learning a Multiview Part-based Model in Virtual World for Pedestrian Detection |
Type |
Conference Article |
Year |
2013 |
Publication |
IEEE Intelligent Vehicles Symposium |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
467 - 472 |
Keywords |
Pedestrian Detection; Virtual World; Part based |
Abstract |
State-of-the-art deformable part-based models based on latent SVM have shown excellent results on human detection. In this paper, we propose to train a multiview deformable part-based model with automatically generated part examples from virtual-world data. The method is efficient as: (i) the part detectors are trained with precisely extracted virtual examples, thus no latent learning is needed, (ii) the multiview pedestrian detector enhances the performance of the pedestrian root model, (iii) a top-down approach is used for part detection which reduces the searching space. We evaluate our model on Daimler and Karlsruhe Pedestrian Benchmarks with publicly available Caltech pedestrian detection evaluation framework and the result outperforms the state-of-the-art latent SVM V4.0, on both average miss rate and speed (our detector is ten times faster). |
Address |
Gold Coast; Australia; June 2013 |
Corporate Author |
|
Thesis |
|
Publisher |
IEEE |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1931-0587 |
ISBN |
978-1-4673-2754-1 |
Medium |
|
Area |
|
Expedition |
|
Conference |
IV |
Notes |
ADAS; 600.054; 600.057 |
Approved |
no |
Call Number |
XVL2013; ADAS @ adas @ xvl2013a |
Serial |
2214 |
Permanent link to this record |
|
|
|
Author |
H. Emrah Tasli; Cevahir Çigla; Theo Gevers; A. Aydin Alatan |
Title |
Super pixel extraction via convexity induced boundary adaptation |
Type |
Conference Article |
Year |
2013 |
Publication |
14th IEEE International Conference on Multimedia and Expo |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
1-6 |
Keywords |
|
Abstract |
This study presents an efficient super-pixel extraction algorithm with major contributions to the state-of-the-art in terms of accuracy and computational complexity. Segmentation accuracy is improved through convexity constrained geodesic distance utilization; while computational efficiency is achieved by replacing complete region processing with boundary adaptation idea. Starting from the uniformly distributed rectangular equal-sized super-pixels, region boundaries are adapted to intensity edges iteratively by assigning boundary pixels to the most similar neighboring super-pixels. At each iteration, super-pixel regions are updated and hence progressively converging to compact pixel groups. Experimental results with state-of-the-art comparisons, validate the performance of the proposed technique in terms of both accuracy and speed. |
Address |
San Jose; USA; July 2013 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1945-7871 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
ICME |
Notes |
ALTRES;ISE |
Approved |
no |
Call Number |
Admin @ si @ TÇG2013 |
Serial |
2367 |
Permanent link to this record |
|
|
|
Author |
Zeynep Yucel; Albert Ali Salah; Çetin Meriçli; Tekin Meriçli; Roberto Valenti; Theo Gevers |
Title |
Joint Attention by Gaze Interpolation and Saliency |
Type |
Journal |
Year |
2013 |
Publication |
IEEE Transactions on cybernetics |
Abbreviated Journal |
T-CIBER |
Volume |
43 |
Issue |
3 |
Pages |
829-842 |
Keywords |
|
Abstract |
Joint attention, which is the ability of coordination of a common point of reference with the communicating party, emerges as a key factor in various interaction scenarios. This paper presents an image-based method for establishing joint attention between an experimenter and a robot. The precise analysis of the experimenter's eye region requires stability and high-resolution image acquisition, which is not always available. We investigate regression-based interpolation of the gaze direction from the head pose of the experimenter, which is easier to track. Gaussian process regression and neural networks are contrasted to interpolate the gaze direction. Then, we combine gaze interpolation with image-based saliency to improve the target point estimates and test three different saliency schemes. We demonstrate the proposed method on a human-robot interaction scenario. Cross-subject evaluations, as well as experiments under adverse conditions (such as dimmed or artificial illumination or motion blur), show that our method generalizes well and achieves rapid gaze estimation for establishing joint attention. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
2168-2267 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
ALTRES;ISE |
Approved |
no |
Call Number |
Admin @ si @ YSM2013 |
Serial |
2363 |
Permanent link to this record |
|
|
|
Author |
Carles Fernandez; Jordi Gonzalez; Joao Manuel R. S. Taveres; Xavier Roca |
Title |
Towards Ontological Cognitive System |
Type |
Book Chapter |
Year |
2013 |
Publication |
Topics in Medical Image Processing and Computational Vision |
Abbreviated Journal |
|
Volume |
8 |
Issue |
|
Pages |
87-99 |
Keywords |
|
Abstract |
The increasing ubiquitousness of digital information in our daily lives has positioned video as a favored information vehicle, and given rise to an astonishing generation of social media and surveillance footage. This raises a series of technological demands for automatic video understanding and management, which together with the compromising attentional limitations of human operators, have motivated the research community to guide its steps towards a better attainment of such capabilities. As a result, current trends on cognitive vision promise to recognize complex events and self-adapt to different environments, while managing and integrating several types of knowledge. Future directions suggest to reinforce the multi-modal fusion of information sources and the communication with end-users. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer Netherlands |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
2212-9391 |
ISBN |
978-94-007-0725-2 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
ISE; 605.203; 302.018; 600.049 |
Approved |
no |
Call Number |
Admin @ si @ FGT2013 |
Serial |
2287 |
Permanent link to this record |