|
Sergio Escalera, Oriol Pujol, & Petia Radeva. (2006). Decoding of Ternary Error Correcting Output Codes. In 11th Iberoamerican Congress on Pattern Recognition (CIARP´06), LNCS 4225: 753–763.
|
|
|
Maryam Asadi-Aghbolaghi, Albert Clapes, Marco Bellantonio, Hugo Jair Escalante, Victor Ponce, Xavier Baro, et al. (2017). Deep Learning for Action and Gesture Recognition in Image Sequences: A Survey. In Gesture Recognition (pp. 539–578).
Abstract: Interest in automatic action and gesture recognition has grown considerably in the last few years. This is due in part to the large number of application domains for this type of technology. As in many other computer vision areas, deep learning based methods have quickly become a reference methodology for obtaining state-of-the-art performance in both tasks. This chapter is a survey of current deep learning based methodologies for action and gesture recognition in sequences of images. The survey reviews both fundamental and cutting edge methodologies reported in the last few years. We introduce a taxonomy that summarizes important aspects of deep learning for approaching both tasks. Details of the proposed architectures, fusion strategies, main datasets, and competitions are reviewed. Also, we summarize and discuss the main works proposed so far with particular interest on how they treat the temporal dimension of data, their highlighting features, and opportunities and challenges for future research. To the best of our knowledge this is the first survey in the topic. We foresee this survey will become a reference in this ever dynamic field of research.
Keywords: Action recognition; Gesture recognition; Deep learning architectures; Fusion strategies
|
|
|
Patricia Suarez, Angel Sappa, & Boris X. Vintimilla. (2021). Deep learning-based vegetation index estimation. In A.Solanki, A.Nayyar, & M.Naved (Eds.), Generative Adversarial Networks for Image-to-Image Translation (pp. 205–234). Elsevier.
|
|
|
Thierry Brouard, A. Delaplace, Muhammad Muzzamil Luqman, H. Cardot, & Jean-Yves Ramel. (2010). Design of Evolutionary Methods Applied to the Learning of Bayesian Nerwork Structures. In Ahmed Rebai (Ed.), Bayesian Network (pp. 13–37). Sciyo.
|
|
|
Michael Teutsch, Angel Sappa, & Riad I. Hammoud. (2022). Detection, Classification, and Tracking. In Computer Vision in the Infrared Spectrum. Synthesis Lectures on Computer Vision (pp. 35–58). SLCV. Springer.
Abstract: Automatic image and video exploitation or content analysis is a technique to extract higher-level information from a scene such as objects, behavior, (inter-)actions, environment, or even weather conditions. The relevant information is assumed to be contained in the two-dimensional signal provided in an image (width and height in pixels) or the three-dimensional signal provided in a video (width, height, and time). But also intermediate-level information such as object classes [196], locations [197], or motion [198] can help applications to fulfill certain tasks such as intelligent compression [199], video summarization [200], or video retrieval [201]. Usually, videos with their temporal dimension are a richer source of data compared to single images [202] and thus certain video content can be extracted from videos only such as object motion or object behavior. Often, machine learning or nowadays deep learning techniques are utilized to model prior knowledge about object or scene appearance using labeled training samples [203, 204]. After a learning phase, these models are then applied in real world applications, which is called inference.
|
|
|
V. Valev, & Petia Radeva. (1992). Determining Structural Description by Boolean Formulas. In H. Bunke (Ed.), Advances in Structural and Syntactic Pattern Recognition (Vol. 5, 131–140). Machine Perception and Artificial Intelligence:. World Scientific.
Abstract: Pattern recognition is an active area of research with many applications, some of which have reached commercial maturity. Structural and syntactic methods are very powerful. They are based on symbolic data structures together with matching, parsing, and reasoning procedures that are able to infer interpretations of complex input patterns.
This book gives an overview of the latest developments and achievements in the field.
|
|
|
Santiago Segui, Laura Igual, Fernando Vilariño, Petia Radeva, Carolina Malagelada, Fernando Azpiroz, et al. (2008). Diagnostic System for Intestinal Motility Disfunctions Using Video Capsule Endoscopy. In and J.K. Tsotsos M. V. A. Gasteratos (Ed.), Computer Vision Systems. 6th International (Vol. 5008, 251–260). LNCS. Berlin Heidelberg: Springer-Verlag.
Abstract: Wireless Video Capsule Endoscopy is a clinical technique consisting of the analysis of images from the intestine which are pro- vided by an ingestible device with a camera attached to it. In this paper we propose an automatic system to diagnose severe intestinal motility disfunctions using the video endoscopy data. The system is based on the application of computer vision techniques within a machine learn- ing framework in order to obtain the characterization of diverse motil- ity events from video sequences. We present experimental results that demonstrate the effectiveness of the proposed system and compare them with the ground-truth provided by the gastroenterologists.
|
|
|
Bogdan Raducanu, & Fadi Dornaika. (2008). Dynamic Vs. Static Recognition of Facial Expressions. In Rabuñal (Ed.), Ambient Intelligence. European Conference (Vol. 5355, 13–25). LNCS.
|
|
|
Angel Sappa, & Boris X. Vintimilla. (2008). Edge Point Linking by Means of Global and Local Schemes. In E. Damiani (Ed.), in Signal Processing for Image Enhancement and Multimedia Processing (Vol. 11, 115–125). Springer.
|
|
|
Ivan Huerta, Dani Rowe, Jordi Gonzalez, & Juan J. Villanueva. (2006). Efficient Incorporation of Motionless Foreground Objects for Adaptive Background Segmentation. In IV Conference on Articulated Motion and Deformable Objects (AMDO´06), LNCS 4069: 424–433.
|
|
|
Jaume Garcia, Debora Gil, & Aura Hernandez-Sabate. (2010). Endowing Canonical Geometries to Cardiac Structures. In O. Camara, M. Pop, K. Rhode, M. Sermesant, N. Smith, & A. Young (Eds.), Statistical Atlases And Computational Models Of The Heart (Vol. 6364, pp. 124–133). LNCS. Springer Berlin / Heidelberg.
Abstract: International conference on Cardiac electrophysiological simulation challenge
In this paper, we show that canonical (shape-based) geometries can be endowed to cardiac structures using tubular coordinates defined over their medial axis. We give an analytic formulation of these geometries by means of B-Splines. Since B-Splines present vector space structure PCA can be applied to their control points and statistical models relating boundaries and the interior of the anatomical structures can be derived. We demonstrate the applicability in two cardiac structures, the 3D Left Ventricular volume, and the 2D Left-Right ventricle set in 2D Short Axis view.
|
|
|
Jose M. Armingol, Jorge Alfonso, Nourdine Aliane, Miguel Clavijo, Sergio Campos-Cordobes, Arturo de la Escalera, et al. (2018). Environmental Perception for Intelligent Vehicles. In Intelligent Vehicles. Enabling Technologies and Future Developments (23–101).
Abstract: Environmental perception represents, because of its complexity, a challenge for Intelligent Transport Systems due to the great variety of situations and different elements that can happen in road environments and that must be faced by these systems. In connection with this, so far there are a variety of solutions as regards sensors and methods, so the results of precision, complexity, cost, or computational load obtained by these works are different. In this chapter some systems based on computer vision and laser techniques are presented. Fusion methods are also introduced in order to provide advanced and reliable perception systems.
Keywords: Computer vision; laser techniques; data fusion; advanced driver assistance systems; traffic monitoring systems; intelligent vehicles
|
|
|
Xavier Baro, & Jordi Vitria. (2008). Evolutionary Object Detection by Means of Naive Bayes Models Estimation. In M. Giacobini (Ed.), Applications of Evolutionary Computing. EvoWorkshops (Vol. 4974, 235–244). LNCS.
|
|
|
Agata Lapedriza, & Jordi Vitria. (2005). Experimental Study of the Usefulness of External Face Features for Face Classification. In Artificial Intelligence Research and Development, IOS Press, 99–106.
|
|
|
Marc Castello, Jordi Gonzalez, Ariel Amato, Pau Baiget, Carles Fernandez, Josep M. Gonfaus, et al. (2013). Exploiting Multimodal Interaction Techniques for Video-Surveillance. In Multimodal Interaction in Image and Video Applications Intelligent Systems Reference Library (Vol. 48, pp. 135–151). Springer Berlin Heidelberg.
Abstract: In this paper we present an example of a video surveillance application that exploits Multimodal Interactive (MI) technologies. The main objective of the so-called VID-Hum prototype was to develop a cognitive artificial system for both the detection and description of a particular set of human behaviours arising from real-world events. The main procedure of the prototype described in this chapter entails: (i) adaptation, since the system adapts itself to the most common behaviours (qualitative data) inferred from tracking (quantitative data) thus being able to recognize abnormal behaviors; (ii) feedback, since an advanced interface based on Natural Language understanding allows end-users the communicationwith the prototype by means of conceptual sentences; and (iii) multimodality, since a virtual avatar has been designed to describe what is happening in the scene, based on those textual interpretations generated by the prototype. Thus, the MI methodology has provided an adequate framework for all these cooperating processes.
|
|