Home | << 1 2 3 4 5 6 7 8 9 10 >> [11–11] |
Records | |||||
---|---|---|---|---|---|
Author | Muhammad Anwer Rao; David Vazquez; Antonio Lopez | ||||
Title | Opponent Colors for Human Detection | Type | Conference Article | ||
Year | 2011 | Publication | 5th Iberian Conference on Pattern Recognition and Image Analysis | Abbreviated Journal | |
Volume | 6669 | Issue | Pages | 363-370 | |
Keywords | Pedestrian Detection; Color; Part Based Models | ||||
Abstract | Human detection is a key component in fields such as advanced driving assistance and video surveillance. However, even detecting non-occluded standing humans remains a challenge of intensive research. Finding good features to build human models for further detection is probably one of the most important issues to face. Currently, shape, texture and motion features have deserve extensive attention in the literature. However, color-based features, which are important in other domains (e.g., image categorization), have received much less attention. In fact, the use of RGB color space has become a kind of choice by default. The focus has been put in developing first and second order features on top of RGB space (e.g., HOG and co-occurrence matrices, resp.). In this paper we evaluate the opponent colors (OPP) space as a biologically inspired alternative for human detection. In particular, by feeding OPP space in the baseline framework of Dalal et al. for human detection (based on RGB, HOG and linear SVM), we will obtain better detection performance than by using RGB space. This is a relevant result since, up to the best of our knowledge, OPP space has not been previously used for human detection. This suggests that in the future it could be worth to compute co-occurrence matrices, self-similarity features, etc., also on top of OPP space, i.e., as we have done with HOG in this paper. | ||||
Address | Las Palmas de Gran Canaria. Spain | ||||
Corporate Author | Thesis | ||||
Publisher | Springer | Place of Publication | Berlin Heidelberg | Editor | J. Vitria; J.M. Sanches; M. Hernandez |
Language | English | Summary Language | English | Original Title | Opponent Colors for Human Detection |
Series Editor | Series Title | Lecture Notes on Computer Science | Abbreviated Series Title | LNCS | |
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-21256-7 | Medium | |
Area | Expedition | Conference | IbPRIA | ||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ RVL2011a | Serial | 1666 | ||
Permanent link to this record | |||||
Author | Daniel Ponsa; Joan Serrat; Antonio Lopez | ||||
Title | On-board image-based vehicle detection and tracking | Type | Journal Article | ||
Year | 2011 | Publication | Transactions of the Institute of Measurement and Control | Abbreviated Journal | TIM |
Volume | 33 | Issue | 7 | Pages | 783-805 |
Keywords | vehicle detection | ||||
Abstract | In this paper we present a computer vision system for daytime vehicle detection and localization, an essential step in the development of several types of advanced driver assistance systems. It has a reduced processing time and high accuracy thanks to the combination of vehicle detection with lane-markings estimation and temporal tracking of both vehicles and lane markings. Concerning vehicle detection, our main contribution is a frame scanning process that inspects images according to the geometry of image formation, and with an Adaboost-based detector that is robust to the variability in the different vehicle types (car, van, truck) and lighting conditions. In addition, we propose a new method to estimate the most likely three-dimensional locations of vehicles on the road ahead. With regards to the lane-markings estimation component, we have two main contributions. First, we employ a different image feature to the other commonly used edges: we use ridges, which are better suited to this problem. Second, we adapt RANSAC, a generic robust estimation method, to fit a parametric model of a pair of lane markings to the image features. We qualitatively assess our vehicle detection system in sequences captured on several road types and under very different lighting conditions. The processed videos are available on a web page associated with this paper. A quantitative evaluation of the system has shown quite accurate results (a low number of false positives and negatives) at a reasonable computation time. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ PSL2011 | Serial | 1413 | ||
Permanent link to this record | |||||
Author | Jose Manuel Alvarez; Antonio Lopez | ||||
Title | Road Detection Based on Illuminant Invariance | Type | Journal Article | ||
Year | 2011 | Publication | IEEE Transactions on Intelligent Transportation Systems | Abbreviated Journal | TITS |
Volume | 12 | Issue | 1 | Pages | 184-193 |
Keywords | road detection | ||||
Abstract | By using an onboard camera, it is possible to detect the free road surface ahead of the ego-vehicle. Road detection is of high relevance for autonomous driving, road departure warning, and supporting driver-assistance systems such as vehicle and pedestrian detection. The key for vision-based road detection is the ability to classify image pixels as belonging or not to the road surface. Identifying road pixels is a major challenge due to the intraclass variability caused by lighting conditions. A particularly difficult scenario appears when the road surface has both shadowed and nonshadowed areas. Accordingly, we propose a novel approach to vision-based road detection that is robust to shadows. The novelty of our approach relies on using a shadow-invariant feature space combined with a model-based classifier. The model is built online to improve the adaptability of the algorithm to the current lighting and the presence of other vehicles in the scene. The proposed algorithm works in still images and does not depend on either road shape or temporal restrictions. Quantitative and qualitative experiments on real-world road sequences with heavy traffic and shadows show that the method is robust to shadows and lighting variations. Moreover, the proposed method provides the highest performance when compared with hue-saturation-intensity (HSI)-based algorithms. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ AlL2011 | Serial | 1456 | ||
Permanent link to this record | |||||
Author | Muhammad Anwer Rao; David Vazquez; Antonio Lopez | ||||
Title | Color Contribution to Part-Based Person Detection in Different Types of Scenarios | Type | Conference Article | ||
Year | 2011 | Publication | 14th International Conference on Computer Analysis of Images and Patterns | Abbreviated Journal | |
Volume | 6855 | Issue | II | Pages | 463-470 |
Keywords | Pedestrian Detection; Color | ||||
Abstract | Camera-based person detection is of paramount interest due to its potential applications. The task is diffcult because the great variety of backgrounds (scenarios, illumination) in which persons are present, as well as their intra-class variability (pose, clothe, occlusion). In fact, the class person is one of the included in the popular PASCAL visual object classes (VOC) challenge. A breakthrough for this challenge, regarding person detection, is due to Felzenszwalb et al. These authors proposed a part-based detector that relies on histograms of oriented gradients (HOG) and latent support vector machines (LatSVM) to learn a model of the whole human body and its constitutive parts, as well as their relative position. Since the approach of Felzenszwalb et al. appeared new variants have been proposed, usually giving rise to more complex models. In this paper, we focus on an issue that has not attracted suficient interest up to now. In particular, we refer to the fact that HOG is usually computed from RGB color space, but other possibilities exist and deserve the corresponding investigation. In this paper we challenge RGB space with the opponent color space (OPP), which is inspired in the human vision system.We will compute the HOG on top of OPP, then we train and test the part-based human classifer by Felzenszwalb et al. using PASCAL VOC challenge protocols and person database. Our experiments demonstrate that OPP outperforms RGB. We also investigate possible differences among types of scenarios: indoor, urban and countryside. Interestingly, our experiments suggest that the beneficts of OPP with respect to RGB mainly come for indoor and countryside scenarios, those in which the human visual system was designed by evolution. | ||||
Address | Seville, Spain | ||||
Corporate Author | Thesis | ||||
Publisher | Springer | Place of Publication | Berlin Heidelberg | Editor | P. Real, D. Diaz, H. Molina, A. Berciano, W. Kropatsch |
Language | English | Summary Language | english | Original Title | Color Contribution to Part-Based Person Detection in Different Types of Scenarios |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-23677-8 | Medium | |
Area | Expedition | Conference | CAIP | ||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ RVL2011b | Serial | 1665 | ||
Permanent link to this record | |||||
Author | David Vazquez; Antonio Lopez; Daniel Ponsa; Javier Marin | ||||
Title | Virtual Worlds and Active Learning for Human Detection | Type | Conference Article | ||
Year | 2011 | Publication | 13th International Conference on Multimodal Interaction | Abbreviated Journal | |
Volume | Issue | Pages | 393-400 | ||
Keywords | Pedestrian Detection; Human detection; Virtual; Domain Adaptation; Active Learning | ||||
Abstract | Image based human detection is of paramount interest due to its potential applications in fields such as advanced driving assistance, surveillance and media analysis. However, even detecting non-occluded standing humans remains a challenge of intensive research. The most promising human detectors rely on classifiers developed in the discriminative paradigm, i.e., trained with labelled samples. However, labeling is a manual intensive step, especially in cases like human detection where it is necessary to provide at least bounding boxes framing the humans for training. To overcome such problem, some authors have proposed the use of a virtual world where the labels of the different objects are obtained automatically. This means that the human models (classifiers) are learnt using the appearance of rendered images, i.e., using realistic computer graphics. Later, these models are used for human detection in images of the real world. The results of this technique are surprisingly good. However, these are not always as good as the classical approach of training and testing with data coming from the same camera, or similar ones. Accordingly, in this paper we address the challenge of using a virtual world for gathering (while playing a videogame) a large amount of automatically labelled samples (virtual humans and background) and then training a classifier that performs equal, in real-world images, than the one obtained by equally training from manually labelled real-world samples. For doing that, we cast the problem as one of domain adaptation. In doing so, we assume that a small amount of manually labelled samples from real-world images is required. To collect these labelled samples we propose a non-standard active learning technique. Therefore, ultimately our human model is learnt by the combination of virtual and real world labelled samples (Fig. 1), which has not been done before. We present quantitative results showing that this approach is valid. | ||||
Address | Alicante, Spain | ||||
Corporate Author | Thesis | ||||
Publisher | ACM DL | Place of Publication | New York, NY, USA, USA | Editor | |
Language | English | Summary Language | English | Original Title | Virtual Worlds and Active Learning for Human Detection |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-1-4503-0641-6 | Medium | ||
Area | Expedition | Conference | ICMI | ||
Notes | ADAS | Approved | yes | ||
Call Number | ADAS @ adas @ VLP2011a | Serial | 1683 | ||
Permanent link to this record | |||||
Author | Naveen Onkarappa; Angel Sappa | ||||
Title | Space Variant Representations for Mobile Platform Vision Applications | Type | Conference Article | ||
Year | 2011 | Publication | 14th International Conference on Computer Analysis of Images and Patterns | Abbreviated Journal | |
Volume | 6855 | Issue | II | Pages | 146-154 |
Keywords | |||||
Abstract | The log-polar space variant representation, motivated by biological vision, has been widely studied in the literature. Its data reduction and invariance properties made it useful in many vision applications. However, due to its nature, it fails in preserving features in the periphery. In the current work, as an attempt to overcome this problem, we propose a novel space-variant representation. It is evaluated and proved to be better than the log-polar representation in preserving the peripheral information, crucial for on-board mobile vision applications. The evaluation is performed by comparing log-polar and the proposed representation once they are used for estimating dense optical flow. | ||||
Address | Seville, Spain | ||||
Corporate Author | Thesis | ||||
Publisher | Springer Berlin Heidelberg | Place of Publication | Editor | P. Real, D. Diaz, H. Molina, A. Berciano, W. Kropatsch | |
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0302-9743 | ISBN | 978-3-642-23677-8 | Medium | |
Area | Expedition | Conference | CAIP | ||
Notes | ADAS | Approved | no | ||
Call Number | NaS2011; ADAS @ adas @ | Serial | 1686 | ||
Permanent link to this record | |||||
Author | Carme Julia; Angel Sappa; Felipe Lumbreras; Joan Serrat; Antonio Lopez | ||||
Title | Rank Estimation in Missing Data Matrix Problems | Type | Journal Article | ||
Year | 2011 | Publication | Journal of Mathematical Imaging and Vision | Abbreviated Journal | JMIV |
Volume | 39 | Issue | 2 | Pages | 140-160 |
Keywords | |||||
Abstract | A novel technique for missing data matrix rank estimation is presented. It is focused on matrices of trajectories, where every element of the matrix corresponds to an image coordinate from a feature point of a rigid moving object at a given frame; missing data are represented as empty entries. The objective of the proposed approach is to estimate the rank of a missing data matrix in order to fill in empty entries with some matrix completion method, without using or assuming neither the number of objects contained in the scene nor the kind of their motion. The key point of the proposed technique consists in studying the frequency behaviour of the individual trajectories, which are seen as 1D signals. The main assumption is that due to the rigidity of the moving objects, the frequency content of the trajectories will be similar after filling in their missing entries. The proposed rank estimation approach can be used in different computer vision problems, where the rank of a missing data matrix needs to be estimated. Experimental results with synthetic and real data are provided in order to empirically show the good performance of the proposed approach. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0924-9907 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ JSL2011; | Serial | 1710 | ||
Permanent link to this record | |||||
Author | Carme Julia; Felipe Lumbreras; Angel Sappa | ||||
Title | A Factorization-based Approach to Photometric Stereo | Type | Journal Article | ||
Year | 2011 | Publication | International Journal of Imaging Systems and Technology | Abbreviated Journal | IJIST |
Volume | 21 | Issue | 1 | Pages | 115-119 |
Keywords | |||||
Abstract | This article presents an adaptation of a factorization technique to tackle the photometric stereo problem. That is to recover the surface normals and reflectance of an object from a set of images obtained under different lighting conditions. The main contribution of the proposed approach is to consider pixels in shadow and saturated regions as missing data, in order to reduce their influence to the result. Concretely, an adapted Alternation technique is used to deal with missing data. Experimental results considering both synthetic and real images show the viability of the proposed factorization-based strategy. © 2011 Wiley Periodicals, Inc. Int J Imaging Syst Technol, 21, 115–119, 2011. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ JLS2011; ADAS @ adas @ | Serial | 1711 | ||
Permanent link to this record | |||||
Author | David Vazquez; Antonio Lopez; Daniel Ponsa; Javier Marin | ||||
Title | Cool world: domain adaptation of virtual and real worlds for human detection using active learning | Type | Conference Article | ||
Year | 2011 | Publication | NIPS Domain Adaptation Workshop: Theory and Application | Abbreviated Journal | NIPS-DA |
Volume | Issue | Pages | |||
Keywords | Pedestrian Detection; Virtual; Domain Adaptation; Active Learning | ||||
Abstract | Image based human detection is of paramount interest for different applications. The most promising human detectors rely on discriminatively learnt classifiers, i.e., trained with labelled samples. However, labelling is a manual intensive task, especially in cases like human detection where it is necessary to provide at least bounding boxes framing the humans for training. To overcome such problem, in Marin et al. we have proposed the use of a virtual world where the labels of the different objects are obtained automatically. This means that the human models (classifiers) are learnt using the appearance of realistic computer graphics. Later, these models are used for human detection in images of the real world. The results of this technique are surprisingly good. However, these are not always as good as the classical approach of training and testing with data coming from the same camera and the same type of scenario. Accordingly, in Vazquez et al. we cast the problem as one of supervised domain adaptation. In doing so, we assume that a small amount of manually labelled samples from real-world images is required. To collect these labelled samples we use an active learning technique. Thus, ultimately our human model is learnt by the combination of virtual- and real-world labelled samples which, to the best of our knowledge, was not done before. Here, we term such combined space cool world. In this extended abstract we summarize our proposal, and include quantitative results from Vazquez et al. showing its validity. | ||||
Address | Granada, Spain | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Granada, Spain | Editor | ||
Language | English | Summary Language | English | Original Title | |
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | DA-NIPS | ||
Notes | ADAS | Approved | no | ||
Call Number | ADAS @ adas @ VLP2011b | Serial | 1756 | ||
Permanent link to this record | |||||
Author | Miguel Oliveira; Angel Sappa; V.Santos | ||||
Title | Unsupervised Local Color Correction for Coarsely Registered Images | Type | Conference Article | ||
Year | 2011 | Publication | IEEE conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 201-208 | ||
Keywords | |||||
Abstract | The current paper proposes a new parametric local color correction technique. Initially, several color transfer functions are computed from the output of the mean shift color segmentation algorithm. Secondly, color influence maps are calculated. Finally, the contribution of every color transfer function is merged using the weights from the color influence maps. The proposed approach is compared with both global and local color correction approaches. Results show that our method outperforms the technique ranked first in a recent performance evaluation on this topic. Moreover, the proposed approach is computed in about one tenth of the time. | ||||
Address | Colorado Springs | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1063-6919 | ISBN | 978-1-4577-0394-2 | Medium | |
Area | Expedition | Conference | CVPR | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ OSS2011; ADAS @ adas @ | Serial | 1766 | ||
Permanent link to this record | |||||
Author | Mohammad Rouhani; Angel Sappa | ||||
Title | Implicit B-Spline Fitting Using the 3L Algorithm | Type | Conference Article | ||
Year | 2011 | Publication | 18th IEEE International Conference on Image Processing | Abbreviated Journal | |
Volume | Issue | Pages | 893-896 | ||
Keywords | |||||
Abstract | |||||
Address | Brussels, Belgium | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICIP | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ RoS2011a; ADAS @ adas @ | Serial | 1782 | ||
Permanent link to this record | |||||
Author | Ferran Diego | ||||
Title | Probabilistic Alignment of Video Sequences Recorded by Moving Cameras | Type | Book Whole | ||
Year | 2011 | Publication | PhD Thesis, Universitat Autonoma de Barcelona-CVC | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | Video alignment consists of integrating multiple video sequences recorded independently into a single video sequence. This means to register both in time (synchronize
frames) and space (image registration) so that the two videos sequences can be fused or compared pixel–wise. In spite of being relatively unknown, many applications today may benefit from the availability of robust and efficient video alignment methods. For instance, video surveillance requires to integrate video sequences that are recorded of the same scene at different times in order to detect changes. The problem of aligning videos has been addressed before, but in the relatively simple cases of fixed or rigidly attached cameras and simultaneous acquisition. In addition, most works rely on restrictive assumptions which reduce its difficulty such as linear time correspondence or the knowledge of the complete trajectories of corresponding scene points on the images; to some extent, these assumptions limit the practical applicability of the solutions developed until now. In this thesis, we focus on the challenging problem of aligning sequences recorded at different times from independent moving cameras following similar but not coincident trajectories. More precisely, this thesis covers four studies that advance the state-of-the-art in video alignment. First, we focus on analyzing and developing a probabilistic framework for video alignment, that is, a principled way to integrate multiple observations and prior information. In this way, two different approaches are presented to exploit the combination of several purely visual features (image–intensities, visual words and dense motion field descriptor), and global positioning system (GPS) information. Second, we focus on reformulating the problem into a single alignment framework since previous works on video alignment adopt a divide–and–conquer strategy, i.e., first solve the synchronization, and then register corresponding frames. This also generalizes the ’classic’ case of fixed geometric transform and linear time mapping. Third, we focus on exploiting directly the time domain of the video sequences in order to avoid exhaustive cross–frame search. This provides relevant information used for learning the temporal mapping between pairs of video sequences. Finally, we focus on adapting these methods to the on–line setting for road detection and vehicle geolocation. The qualitative and quantitative results presented in this thesis on a variety of real–world pairs of video sequences show that the proposed method is: robust to varying imaging conditions, different image content (e.g., incoming and outgoing vehicles), variations on camera velocity, and different scenarios (indoor and outdoor) going beyond the state–of–the–art. Moreover, the on–line video alignment has been successfully applied for road detection and vehicle geolocation achieving promising results. |
||||
Address | |||||
Corporate Author | Thesis | Ph.D. thesis | |||
Publisher | Ediciones Graficas Rey | Place of Publication | Editor | Joan Serrat | |
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ Die2011 | Serial | 1787 | ||
Permanent link to this record | |||||
Author | Mohammad Rouhani; Angel Sappa | ||||
Title | Correspondence Free Registration through a Point-to-Model Distance Minimization | Type | Conference Article | ||
Year | 2011 | Publication | 13th IEEE International Conference on Computer Vision | Abbreviated Journal | |
Volume | Issue | Pages | 2150-2157 | ||
Keywords | |||||
Abstract | This paper presents a novel formulation, which derives in a smooth minimization problem, to tackle the rigid registration between a given point set and a model set. Unlike most of the existing works, which are based on minimizing a point-wise correspondence term, we propose to describe the model set by means of an implicit representation. It allows a new definition of the registration error, which works beyond the point level representation. Moreover, it could be used in a gradient-based optimization framework. The proposed approach consists of two stages. Firstly, a novel formulation is proposed that relates the registration parameters with the distance between the model and data set. Secondly, the registration parameters are obtained by means of the Levengberg-Marquardt algorithm. Experimental results and comparisons with state of the art show the validity of the proposed framework. | ||||
Address | Barcelona | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1550-5499 | ISBN | 978-1-4577-1101-5 | Medium | |
Area | Expedition | Conference | ICCV | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ RoS2011b; ADAS @ adas @ | Serial | 1832 | ||
Permanent link to this record | |||||
Author | Fadi Dornaika; Jose Manuel Alvarez; Angel Sappa; Antonio Lopez | ||||
Title | A New Framework for Stereo Sensor Pose through Road Segmentation and Registration | Type | Journal Article | ||
Year | 2011 | Publication | IEEE Transactions on Intelligent Transportation Systems | Abbreviated Journal | TITS |
Volume | 12 | Issue | 4 | Pages | 954-966 |
Keywords | road detection | ||||
Abstract | This paper proposes a new framework for real-time estimation of the onboard stereo head's position and orientation relative to the road surface, which is required for any advanced driver-assistance application. This framework can be used with all road types: highways, urban, etc. Unlike existing works that rely on feature extraction in either the image domain or 3-D space, we propose a framework that directly estimates the unknown parameters from the stream of stereo pairs' brightness. The proposed approach consists of two stages that are invoked for every stereo frame. The first stage segments the road region in one monocular view. The second stage estimates the camera pose using a featureless registration between the segmented monocular road region and the other view in the stereo pair. This paper has two main contributions. The first contribution combines a road segmentation algorithm with a registration technique to estimate the online stereo camera pose. The second contribution solves the registration using a featureless method, which is carried out using two different optimization techniques: 1) the differential evolution algorithm and 2) the Levenberg-Marquardt (LM) algorithm. We provide experiments and evaluations of performance. The results presented show the validity of our proposed framework. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1524-9050 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ DAS2011; ADAS @ adas @ das2011a | Serial | 1833 | ||
Permanent link to this record | |||||
Author | G.D. Evangelidis; Ferran Diego; Joan Serrat; Antonio Lopez | ||||
Title | Slice Matching for Accurate Spatio-Temporal Alignment | Type | Conference Article | ||
Year | 2011 | Publication | In ICCV Workshop on Visual Surveillance | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | video alignment | ||||
Abstract | Video synchronization and alignment is a rather recent topic in computer vision. It usually deals with the problem of aligning sequences recorded simultaneously by static, jointly- or independently-moving cameras. In this paper, we investigate the more difficult problem of matching videos captured at different times from independently-moving cameras, whose trajectories are approximately coincident or parallel. To this end, we propose a novel method that pixel-wise aligns videos and allows thus to automatically highlight their differences. This primarily aims at visual surveillance but the method can be adopted as is by other related video applications, like object transfer (augmented reality) or high dynamic range video. We build upon a slice matching scheme to first synchronize the sequences, while we develop a spatio-temporal alignment scheme to spatially register corresponding frames and refine the temporal mapping. We investigate the performance of the proposed method on videos recorded from vehicles driven along different types of roads and compare with related previous works. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | VS | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ EDS2011; ADAS @ adas @ eds2011a | Serial | 1861 | ||
Permanent link to this record |