Home | << 1 2 3 4 5 6 7 8 9 10 >> |
Records | |||||
---|---|---|---|---|---|
Author | Victor Campmany; Sergio Silva; Juan Carlos Moure; Toni Espinosa; David Vazquez; Antonio Lopez | ||||
Title | GPU-based pedestrian detection for autonomous driving | Type | Conference Article | ||
Year | 2016 | Publication | GPU Technology Conference | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Pedestrian Detection; GPU | ||||
Abstract | Pedestrian detection for autonomous driving is one of the hardest tasks within computer vision, and involves huge computational costs. Obtaining acceptable real-time performance, measured in frames per second (fps), for the most advanced algorithms is nowadays a hard challenge. Taking the work in [1] as our baseline, we propose a CUDA implementation of a pedestrian detection system that includes LBP and HOG as feature descriptors and SVM and Random forest as classifiers. We introduce significant algorithmic adjustments and optimizations to adapt the problem to the NVIDIA GPU architecture. The aim is to deploy a real-time system providing reliable results. | ||||
Address | Silicon Valley; San Francisco; USA; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | GTC | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ CSM2016 | Serial | 2737 | ||
Permanent link to this record | |||||
Author | Daniel Hernandez; Juan Carlos Moure; Toni Espinosa; Alejandro Chacon; David Vazquez; Antonio Lopez | ||||
Title | Real-time 3D Reconstruction for Autonomous Driving via Semi-Global Matching | Type | Conference Article | ||
Year | 2016 | Publication | GPU Technology Conference | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Stereo; Autonomous Driving; GPU; 3d reconstruction | ||||
Abstract | Robust and dense computation of depth information from stereo-camera systems is a computationally demanding requirement for real-time autonomous driving. Semi-Global Matching (SGM) [1] approximates heavy-computation global algorithms results but with lower computational complexity, therefore it is a good candidate for a real-time implementation. SGM minimizes energy along several 1D paths across the image. The aim of this work is to provide a real-time system producing reliable results on energy-efficient hardware. Our design runs on a NVIDIA Titan X GPU at 104.62 FPS and on a NVIDIA Drive PX at 6.7 FPS, promising for real-time platforms | ||||
Address | Silicon Valley; San Francisco; USA; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | GTC | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ HME2016 | Serial | 2738 | ||
Permanent link to this record | |||||
Author | German Ros; Laura Sellart; Joanna Materzynska; David Vazquez; Antonio Lopez | ||||
Title | The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes | Type | Conference Article | ||
Year | 2016 | Publication | 29th IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 3234-3243 | ||
Keywords | Domain Adaptation; Autonomous Driving; Virtual Data; Semantic Segmentation | ||||
Abstract | Vision-based semantic segmentation in urban scenarios is a key functionality for autonomous driving. The irruption of deep convolutional neural networks (DCNNs) allows to foresee obtaining reliable classifiers to perform such a visual task. However, DCNNs require to learn many parameters from raw images; thus, having a sufficient amount of diversified images with this class annotations is needed. These annotations are obtained by a human cumbersome labour specially challenging for semantic segmentation, since pixel-level annotations are required. In this paper, we propose to use a virtual world for automatically generating realistic synthetic images with pixel-level annotations. Then, we address the question of how useful can be such data for the task of semantic segmentation; in particular, when using a DCNN paradigm. In order to answer this question we have generated a synthetic diversified collection of urban images, named SynthCity, with automatically generated class annotations. We use SynthCity in combination with publicly available real-world urban images with manually provided annotations. Then, we conduct experiments on a DCNN setting that show how the inclusion of SynthCity in the training stage significantly improves the performance of the semantic segmentation task | ||||
Address | Las Vegas; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CVPR | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ RSM2016 | Serial | 2739 | ||
Permanent link to this record | |||||
Author | Daniel Hernandez; Alejandro Chacon; Antonio Espinosa; David Vazquez; Juan Carlos Moure; Antonio Lopez | ||||
Title | Embedded real-time stereo estimation via Semi-Global Matching on the GPU | Type | Conference Article | ||
Year | 2016 | Publication | 16th International Conference on Computational Science | Abbreviated Journal | |
Volume | 80 | Issue | Pages | 143-153 | |
Keywords | Autonomous Driving; Stereo; CUDA; 3d reconstruction | ||||
Abstract | Dense, robust and real-time computation of depth information from stereo-camera systems is a computationally demanding requirement for robotics, advanced driver assistance systems (ADAS) and autonomous vehicles. Semi-Global Matching (SGM) is a widely used algorithm that propagates consistency constraints along several paths across the image. This work presents a real-time system producing reliable disparity estimation results on the new embedded energy-efficient GPU devices. Our design runs on a Tegra X1 at 41 frames per second for an image size of 640x480, 128 disparity levels, and using 4 path directions for the SGM method. | ||||
Address | San Diego; CA; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCS | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ HCE2016a | Serial | 2740 | ||
Permanent link to this record | |||||
Author | Victor Campmany; Sergio Silva; Antonio Espinosa; Juan Carlos Moure; David Vazquez; Antonio Lopez | ||||
Title | GPU-based pedestrian detection for autonomous driving | Type | Conference Article | ||
Year | 2016 | Publication | 16th International Conference on Computational Science | Abbreviated Journal | |
Volume | 80 | Issue | Pages | 2377-2381 | |
Keywords | Pedestrian detection; Autonomous Driving; CUDA | ||||
Abstract | We propose a real-time pedestrian detection system for the embedded Nvidia Tegra X1 GPU-CPU hybrid platform. The pipeline is composed by the following state-of-the-art algorithms: Histogram of Local Binary Patterns (LBP) and Histograms of Oriented Gradients (HOG) features extracted from the input image; Pyramidal Sliding Window technique for foreground segmentation; and Support Vector Machine (SVM) for classification. Results show a 8x speedup in the target Tegra X1 platform and a better performance/watt ratio than desktop CUDA platforms in study. | ||||
Address | San Diego; CA; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCS | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ CSE2016 | Serial | 2741 | ||
Permanent link to this record | |||||
Author | Maedeh Aghaei; Mariella Dimiccoli; Petia Radeva | ||||
Title | Multi-face tracking by extended bag-of-tracklets in egocentric photo-streams | Type | Journal Article | ||
Year | 2016 | Publication | Computer Vision and Image Understanding | Abbreviated Journal | CVIU |
Volume | 149 | Issue | Pages | 146-156 | |
Keywords | |||||
Abstract | Wearable cameras offer a hands-free way to record egocentric images of daily experiences, where social events are of special interest. The first step towards detection of social events is to track the appearance of multiple persons involved in them. In this paper, we propose a novel method to find correspondences of multiple faces in low temporal resolution egocentric videos acquired through a wearable camera. This kind of photo-stream imposes additional challenges to the multi-tracking problem with respect to conventional videos. Due to the free motion of the camera and to its low temporal resolution, abrupt changes in the field of view, in illumination condition and in the target location are highly frequent. To overcome such difficulties, we propose a multi-face tracking method that generates a set of tracklets through finding correspondences along the whole sequence for each detected face and takes advantage of the tracklets redundancy to deal with unreliable ones. Similar tracklets are grouped into the so called extended bag-of-tracklets (eBoT), which is aimed to correspond to a specific person. Finally, a prototype tracklet is extracted for each eBoT, where the occurred occlusions are estimated by relying on a new measure of confidence. We validated our approach over an extensive dataset of egocentric photo-streams and compared it to state of the art methods, demonstrating its effectiveness and robustness. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | MILAB; | Approved | no | ||
Call Number | Admin @ si @ ADR2016b | Serial | 2742 | ||
Permanent link to this record | |||||
Author | Onur Ferhat; Fernando Vilariño | ||||
Title | Low Cost Eye Tracking: The Current Panorama | Type | Journal Article | ||
Year | 2016 | Publication | Computational Intelligence and Neuroscience | Abbreviated Journal | CIN |
Volume | Issue | Pages | Article ID 8680541 | ||
Keywords | |||||
Abstract | Despite the availability of accurate, commercial gaze tracker devices working with infrared (IR) technology, visible light gaze tracking constitutes an interesting alternative by allowing scalability and removing hardware requirements. Over the last years, this field has seen examples of research showing performance comparable to the IR alternatives. In this work, we survey the previous work on remote, visible light gaze trackers and analyze the explored techniques from various perspectives such as calibration strategies, head pose invariance, and gaze estimation techniques. We also provide information on related aspects of research such as public datasets to test against, open source projects to build upon, and gaze tracking services to directly use in applications. With all this information, we aim to provide the contemporary and future researchers with a map detailing previously explored ideas and the required tools. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | MV; 605.103; 600.047; 600.097;SIAI | Approved | no | ||
Call Number | Admin @ si @ FeV2016 | Serial | 2744 | ||
Permanent link to this record | |||||
Author | Maria Oliver; G. Haro; Mariella Dimiccoli; B. Mazin; C. Ballester | ||||
Title | A Computational Model for Amodal Completion | Type | Journal Article | ||
Year | 2016 | Publication | Journal of Mathematical Imaging and Vision | Abbreviated Journal | JMIV |
Volume | 56 | Issue | 3 | Pages | 511–534 |
Keywords | Perception; visual completion; disocclusion; Bayesian model;relatability; Euler elastica | ||||
Abstract | This paper presents a computational model to recover the most likely interpretation
of the 3D scene structure from a planar image, where some objects may occlude others. The estimated scene interpretation is obtained by integrating some global and local cues and provides both the complete disoccluded objects that form the scene and their ordering according to depth. Our method first computes several distal scenes which are compatible with the proximal planar image. To compute these different hypothesized scenes, we propose a perceptually inspired object disocclusion method, which works by minimizing the Euler's elastica as well as by incorporating the relatability of partially occluded contours and the convexity of the disoccluded objects. Then, to estimate the preferred scene we rely on a Bayesian model and define probabilities taking into account the global complexity of the objects in the hypothesized scenes as well as the effort of bringing these objects in their relative position in the planar image, which is also measured by an Euler's elastica-based quantity. The model is illustrated with numerical experiments on, both, synthetic and real images showing the ability of our model to reconstruct the occluded objects and the preferred perceptual order among them. We also present results on images of the Berkeley dataset with provided figure-ground ground-truth labeling. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | MILAB; 601.235 | Approved | no | ||
Call Number | Admin @ si @ OHD2016b | Serial | 2745 | ||
Permanent link to this record | |||||
Author | C. Alejandro Parraga; Arash Akbarinia | ||||
Title | NICE: A Computational Solution to Close the Gap from Colour Perception to Colour Categorization | Type | Journal Article | ||
Year | 2016 | Publication | PLoS One | Abbreviated Journal | Plos |
Volume | 11 | Issue | 3 | Pages | e0149538 |
Keywords | |||||
Abstract | The segmentation of visible electromagnetic radiation into chromatic categories by the human visual system has been extensively studied from a perceptual point of view, resulting in several colour appearance models. However, there is currently a void when it comes to relate these results to the physiological mechanisms that are known to shape the pre-cortical and cortical visual pathway. This work intends to begin to fill this void by proposing a new physiologically plausible model of colour categorization based on Neural Isoresponsive Colour Ellipsoids (NICE) in the cone-contrast space defined by the main directions of the visual signals entering the visual cortex. The model was adjusted to fit psychophysical measures that concentrate on the categorical boundaries and are consistent with the ellipsoidal isoresponse surfaces of visual cortical neurons. By revealing the shape of such categorical colour regions, our measures allow for a more precise and parsimonious description, connecting well-known early visual processing mechanisms to the less understood phenomenon of colour categorization. To test the feasibility of our method we applied it to exemplary images and a popular ground-truth chart obtaining labelling results that are better than those of current state-of-the-art algorithms. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | NEUROBIT; 600.068 | Approved | no | ||
Call Number | Admin @ si @ PaA2016a | Serial | 2747 | ||
Permanent link to this record | |||||
Author | Pedro Martins; Paulo Carvalho; Carlo Gatta | ||||
Title | On the completeness of feature-driven maximally stable extremal regions | Type | Journal Article | ||
Year | 2016 | Publication | Pattern Recognition Letters | Abbreviated Journal | PRL |
Volume | 74 | Issue | Pages | 9-16 | |
Keywords | Local features; Completeness; Maximally Stable Extremal Regions | ||||
Abstract | By definition, local image features provide a compact representation of the image in which most of the image information is preserved. This capability offered by local features has been overlooked, despite being relevant in many application scenarios. In this paper, we analyze and discuss the performance of feature-driven Maximally Stable Extremal Regions (MSER) in terms of the coverage of informative image parts (completeness). This type of features results from an MSER extraction on saliency maps in which features related to objects boundaries or even symmetry axes are highlighted. These maps are intended to be suitable domains for MSER detection, allowing this detector to provide a better coverage of informative image parts. Our experimental results, which were based on a large-scale evaluation, show that feature-driven MSER have relatively high completeness values and provide more complete sets than a traditional MSER detection even when sets of similar cardinality are considered. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Elsevier B.V. | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0167-8655 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | LAMP;MILAB; | Approved | no | ||
Call Number | Admin @ si @ MCG2016 | Serial | 2748 | ||
Permanent link to this record | |||||
Author | Eugenio Alcala; Laura Sellart; Vicenc Puig; Joseba Quevedo; Jordi Saludes; David Vazquez; Antonio Lopez | ||||
Title | Comparison of two non-linear model-based control strategies for autonomous vehicles | Type | Conference Article | ||
Year | 2016 | Publication | 24th Mediterranean Conference on Control and Automation | Abbreviated Journal | |
Volume | Issue | Pages | 846-851 | ||
Keywords | Autonomous Driving; Control | ||||
Abstract | This paper presents the comparison of two nonlinear model-based control strategies for autonomous cars. A control oriented model of vehicle based on a bicycle model is used. The two control strategies use a model reference approach. Using this approach, the error dynamics model is developed. Both controllers receive as input the longitudinal, lateral and orientation errors generating as control outputs the steering angle and the velocity of the vehicle. The first control approach is based on a non-linear control law that is designed by means of the Lyapunov direct approach. The second approach is based on a sliding mode-control that defines a set of sliding surfaces over which the error trajectories will converge. The main advantage of the sliding-control technique is the robustness against non-linearities and parametric uncertainties in the model. However, the main drawback of first order sliding mode is the chattering, so it has been implemented a high order sliding mode control. To test and compare the proposed control strategies, different path following scenarios are used in simulation. | ||||
Address | Athens; Greece; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | MED | ||
Notes | ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ ASP2016 | Serial | 2750 | ||
Permanent link to this record | |||||
Author | Joan Mas; Alicia Fornes; Josep Llados | ||||
Title | An Interactive Transcription System of Census Records using Word-Spotting based Information Transfer | Type | Conference Article | ||
Year | 2016 | Publication | 12th IAPR Workshop on Document Analysis Systems | Abbreviated Journal | |
Volume | Issue | Pages | 54-59 | ||
Keywords | |||||
Abstract | This paper presents a system to assist in the transcription of historical handwritten census records in a crowdsourcing platform. Census records have a tabular structured layout. They consist in a sequence of rows with information of homes ordered by street address. For each household snippet in the page, the list of family members is reported. The censuses are recorded in intervals of a few years and the information of individuals in each household is quite stable from a point in time to the next one. This redundancy is used to assist the transcriber, so the redundant information is transferred from the census already transcribed to the next one. Household records are aligned from one year to the next one using the knowledge of the ordering by street address. Given an already transcribed census, a query by string word spotting is applied. Thus, names from the census in time t are used as queries in the corresponding home record in time t+1. Since the search is constrained, the obtained precision-recall values are very high, with an important reduction in the transcription time. The proposed system has been tested in a real citizen-science experience where non expert users transcribe the census data of their home town. | ||||
Address | Santorini; Greece; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | DAS | ||
Notes | DAG; 603.053; 602.006; 600.061; 600.077; 600.097 | Approved | no | ||
Call Number | Admin @ si @ MFL2016 | Serial | 2751 | ||
Permanent link to this record | |||||
Author | Juan Ignacio Toledo; Alicia Fornes; Jordi Cucurull; Josep Llados | ||||
Title | Election Tally Sheets Processing System | Type | Conference Article | ||
Year | 2016 | Publication | 12th IAPR Workshop on Document Analysis Systems | Abbreviated Journal | |
Volume | Issue | Pages | 364-368 | ||
Keywords | |||||
Abstract | In paper based elections, manual tallies at polling station level produce myriads of documents. These documents share a common form-like structure and a reduced vocabulary worldwide. On the other hand, each tally sheet is filled by a different writer and on different countries, different scripts are used. We present a complete document analysis system for electoral tally sheet processing combining state of the art techniques with a new handwriting recognition subprocess based on unsupervised feature discovery with Variational Autoencoders and sequence classification with BLSTM neural networks. The whole system is designed to be script independent and allows a fast and reliable results consolidation process with reduced operational cost. | ||||
Address | Santorini; Greece; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | DAS | ||
Notes | DAG; 602.006; 600.061; 601.225; 600.077; 600.097 | Approved | no | ||
Call Number | TFC2016 | Serial | 2752 | ||
Permanent link to this record | |||||
Author | Anders Hast; Alicia Fornes | ||||
Title | A Segmentation-free Handwritten Word Spotting Approach by Relaxed Feature Matching | Type | Conference Article | ||
Year | 2016 | Publication | 12th IAPR Workshop on Document Analysis Systems | Abbreviated Journal | |
Volume | Issue | Pages | 150-155 | ||
Keywords | |||||
Abstract | The automatic recognition of historical handwritten documents is still considered challenging task. For this reason, word spotting emerges as a good alternative for making the information contained in these documents available to the user. Word spotting is defined as the task of retrieving all instances of the query word in a document collection, becoming a useful tool for information retrieval. In this paper we propose a segmentation-free word spotting approach able to deal with large document collections. Our method is inspired on feature matching algorithms that have been applied to image matching and retrieval. Since handwritten words have different shape, there is no exact transformation to be obtained. However, the sufficient degree of relaxation is achieved by using a Fourier based descriptor and an alternative approach to RANSAC called PUMA. The proposed approach is evaluated on historical marriage records, achieving promising results. | ||||
Address | Santorini; Greece; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | DAS | ||
Notes | DAG; 602.006; 600.061; 600.077; 600.097 | Approved | no | ||
Call Number | HaF2016 | Serial | 2753 | ||
Permanent link to this record | |||||
Author | Alejandro Gonzalez Alzate; Zhijie Fang; Yainuvis Socarras; Joan Serrat; David Vazquez; Jiaolong Xu; Antonio Lopez | ||||
Title | Pedestrian Detection at Day/Night Time with Visible and FIR Cameras: A Comparison | Type | Journal Article | ||
Year | 2016 | Publication | Sensors | Abbreviated Journal | SENS |
Volume | 16 | Issue | 6 | Pages | 820 |
Keywords | Pedestrian Detection; FIR | ||||
Abstract | Despite all the significant advances in pedestrian detection brought by computer vision for driving assistance, it is still a challenging problem. One reason is the extremely varying lighting conditions under which such a detector should operate, namely day and night time. Recent research has shown that the combination of visible and non-visible imaging modalities may increase detection accuracy, where the infrared spectrum plays a critical role. The goal of this paper is to assess the accuracy gain of different pedestrian models (holistic, part-based, patch-based) when training with images in the far infrared spectrum. Specifically, we want to compare detection accuracy on test images recorded at day and nighttime if trained (and tested) using (a) plain color images, (b) just infrared images and (c) both of them. In order to obtain results for the last item we propose an early fusion approach to combine features from both modalities. We base the evaluation on a new dataset we have built for this purpose as well as on the publicly available KAIST multispectral dataset. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1424-8220 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS; 600.085; 600.076; 600.082; 601.281 | Approved | no | ||
Call Number | ADAS @ adas @ GFS2016 | Serial | 2754 | ||
Permanent link to this record |