Home | [181–190] << 191 192 193 194 195 196 197 198 199 200 >> [201–210] |
![]() |
Records | |||||
---|---|---|---|---|---|
Author | Daniel Hernandez; Juan Carlos Moure; Toni Espinosa; Alejandro Chacon; David Vazquez; Antonio Lopez | ||||
Title | Real-time 3D Reconstruction for Autonomous Driving via Semi-Global Matching | Type | Conference Article | ||
Year | 2016 | Publication | GPU Technology Conference | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Stereo; Autonomous Driving; GPU; 3d reconstruction | ||||
Abstract | Robust and dense computation of depth information from stereo-camera systems is a computationally demanding requirement for real-time autonomous driving. Semi-Global Matching (SGM) [1] approximates heavy-computation global algorithms results but with lower computational complexity, therefore it is a good candidate for a real-time implementation. SGM minimizes energy along several 1D paths across the image. The aim of this work is to provide a real-time system producing reliable results on energy-efficient hardware. Our design runs on a NVIDIA Titan X GPU at 104.62 FPS and on a NVIDIA Drive PX at 6.7 FPS, promising for real-time platforms | ||||
Address | Silicon Valley; San Francisco; USA; April 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | GTC | ||
Notes ![]() |
ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ HME2016 | Serial | 2738 | ||
Permanent link to this record | |||||
Author | German Ros; Laura Sellart; Joanna Materzynska; David Vazquez; Antonio Lopez | ||||
Title | The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes | Type | Conference Article | ||
Year | 2016 | Publication | 29th IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 3234-3243 | ||
Keywords | Domain Adaptation; Autonomous Driving; Virtual Data; Semantic Segmentation | ||||
Abstract | Vision-based semantic segmentation in urban scenarios is a key functionality for autonomous driving. The irruption of deep convolutional neural networks (DCNNs) allows to foresee obtaining reliable classifiers to perform such a visual task. However, DCNNs require to learn many parameters from raw images; thus, having a sufficient amount of diversified images with this class annotations is needed. These annotations are obtained by a human cumbersome labour specially challenging for semantic segmentation, since pixel-level annotations are required. In this paper, we propose to use a virtual world for automatically generating realistic synthetic images with pixel-level annotations. Then, we address the question of how useful can be such data for the task of semantic segmentation; in particular, when using a DCNN paradigm. In order to answer this question we have generated a synthetic diversified collection of urban images, named SynthCity, with automatically generated class annotations. We use SynthCity in combination with publicly available real-world urban images with manually provided annotations. Then, we conduct experiments on a DCNN setting that show how the inclusion of SynthCity in the training stage significantly improves the performance of the semantic segmentation task | ||||
Address | Las Vegas; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CVPR | ||
Notes ![]() |
ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ RSM2016 | Serial | 2739 | ||
Permanent link to this record | |||||
Author | Daniel Hernandez; Alejandro Chacon; Antonio Espinosa; David Vazquez; Juan Carlos Moure; Antonio Lopez | ||||
Title | Embedded real-time stereo estimation via Semi-Global Matching on the GPU | Type | Conference Article | ||
Year | 2016 | Publication | 16th International Conference on Computational Science | Abbreviated Journal | |
Volume | 80 | Issue | Pages | 143-153 | |
Keywords | Autonomous Driving; Stereo; CUDA; 3d reconstruction | ||||
Abstract | Dense, robust and real-time computation of depth information from stereo-camera systems is a computationally demanding requirement for robotics, advanced driver assistance systems (ADAS) and autonomous vehicles. Semi-Global Matching (SGM) is a widely used algorithm that propagates consistency constraints along several paths across the image. This work presents a real-time system producing reliable disparity estimation results on the new embedded energy-efficient GPU devices. Our design runs on a Tegra X1 at 41 frames per second for an image size of 640x480, 128 disparity levels, and using 4 path directions for the SGM method. | ||||
Address | San Diego; CA; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCS | ||
Notes ![]() |
ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ HCE2016a | Serial | 2740 | ||
Permanent link to this record | |||||
Author | Victor Campmany; Sergio Silva; Antonio Espinosa; Juan Carlos Moure; David Vazquez; Antonio Lopez | ||||
Title | GPU-based pedestrian detection for autonomous driving | Type | Conference Article | ||
Year | 2016 | Publication | 16th International Conference on Computational Science | Abbreviated Journal | |
Volume | 80 | Issue | Pages | 2377-2381 | |
Keywords | Pedestrian detection; Autonomous Driving; CUDA | ||||
Abstract | We propose a real-time pedestrian detection system for the embedded Nvidia Tegra X1 GPU-CPU hybrid platform. The pipeline is composed by the following state-of-the-art algorithms: Histogram of Local Binary Patterns (LBP) and Histograms of Oriented Gradients (HOG) features extracted from the input image; Pyramidal Sliding Window technique for foreground segmentation; and Support Vector Machine (SVM) for classification. Results show a 8x speedup in the target Tegra X1 platform and a better performance/watt ratio than desktop CUDA platforms in study. | ||||
Address | San Diego; CA; USA; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCS | ||
Notes ![]() |
ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ CSE2016 | Serial | 2741 | ||
Permanent link to this record | |||||
Author | Eugenio Alcala; Laura Sellart; Vicenc Puig; Joseba Quevedo; Jordi Saludes; David Vazquez; Antonio Lopez | ||||
Title | Comparison of two non-linear model-based control strategies for autonomous vehicles | Type | Conference Article | ||
Year | 2016 | Publication | 24th Mediterranean Conference on Control and Automation | Abbreviated Journal | |
Volume | Issue | Pages | 846-851 | ||
Keywords | Autonomous Driving; Control | ||||
Abstract | This paper presents the comparison of two nonlinear model-based control strategies for autonomous cars. A control oriented model of vehicle based on a bicycle model is used. The two control strategies use a model reference approach. Using this approach, the error dynamics model is developed. Both controllers receive as input the longitudinal, lateral and orientation errors generating as control outputs the steering angle and the velocity of the vehicle. The first control approach is based on a non-linear control law that is designed by means of the Lyapunov direct approach. The second approach is based on a sliding mode-control that defines a set of sliding surfaces over which the error trajectories will converge. The main advantage of the sliding-control technique is the robustness against non-linearities and parametric uncertainties in the model. However, the main drawback of first order sliding mode is the chattering, so it has been implemented a high order sliding mode control. To test and compare the proposed control strategies, different path following scenarios are used in simulation. | ||||
Address | Athens; Greece; June 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | MED | ||
Notes ![]() |
ADAS; 600.085; 600.082; 600.076 | Approved | no | ||
Call Number | ADAS @ adas @ ASP2016 | Serial | 2750 | ||
Permanent link to this record | |||||
Author | Ishaan Gulrajani; Kundan Kumar; Faruk Ahmed; Adrien Ali Taiga; Francesco Visin; David Vazquez; Aaron Courville | ||||
Title | PixelVAE: A Latent Variable Model for Natural Images | Type | Conference Article | ||
Year | 2017 | Publication | 5th International Conference on Learning Representations | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Deep Learning; Unsupervised Learning | ||||
Abstract | Natural image modeling is a landmark challenge of unsupervised learning. Variational Autoencoders (VAEs) learn a useful latent representation and generate samples that preserve global structure but tend to suffer from image blurriness. PixelCNNs model sharp contours and details very well, but lack an explicit latent representation and have difficulty modeling large-scale structure in a computationally efficient way. In this paper, we present PixelVAE, a VAE model with an autoregressive decoder based on PixelCNN. The resulting architecture achieves state-of-the-art log-likelihood on binarized MNIST. We extend PixelVAE to a hierarchy of multiple latent variables at different scales; this hierarchical model achieves competitive likelihood on 64x64 ImageNet and generates high-quality samples on LSUN bedrooms. | ||||
Address | Toulon; France; April 2017 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICLR | ||
Notes ![]() |
ADAS; 600.085; 600.076; 601.281; 600.118 | Approved | no | ||
Call Number | ADAS @ adas @ GKA2017 | Serial | 2815 | ||
Permanent link to this record | |||||
Author | Zhijie Fang; David Vazquez; Antonio Lopez | ||||
Title | On-Board Detection of Pedestrian Intentions | Type | Journal Article | ||
Year | 2017 | Publication | Sensors | Abbreviated Journal | SENS |
Volume | 17 | Issue | 10 | Pages | 2193 |
Keywords | pedestrian intention; ADAS; self-driving | ||||
Abstract | Avoiding vehicle-to-pedestrian crashes is a critical requirement for nowadays advanced driver assistant systems (ADAS) and future self-driving vehicles. Accordingly, detecting pedestrians from raw sensor data has a history of more than 15 years of research, with vision playing a central role.
During the last years, deep learning has boosted the accuracy of image-based pedestrian detectors. However, detection is just the first step towards answering the core question, namely is the vehicle going to crash with a pedestrian provided preventive actions are not taken? Therefore, knowing as soon as possible if a detected pedestrian has the intention of crossing the road ahead of the vehicle is essential for performing safe and comfortable maneuvers that prevent a crash. However, compared to pedestrian detection, there is relatively little literature on detecting pedestrian intentions. This paper aims to contribute along this line by presenting a new vision-based approach which analyzes the pose of a pedestrian along several frames to determine if he or she is going to enter the road or not. We present experiments showing 750 ms of anticipation for pedestrians crossing the road, which at a typical urban driving speed of 50 km/h can provide 15 additional meters (compared to a pure pedestrian detector) for vehicle automatic reactions or to warn the driver. Moreover, in contrast with state-of-the-art methods, our approach is monocular, neither requiring stereo nor optical flow information. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes ![]() |
ADAS; 600.085; 600.076; 601.223; 600.116; 600.118 | Approved | no | ||
Call Number | Admin @ si @ FVL2017 | Serial | 2983 | ||
Permanent link to this record | |||||
Author | Alejandro Gonzalez Alzate; Zhijie Fang; Yainuvis Socarras; Joan Serrat; David Vazquez; Jiaolong Xu; Antonio Lopez | ||||
Title | Pedestrian Detection at Day/Night Time with Visible and FIR Cameras: A Comparison | Type | Journal Article | ||
Year | 2016 | Publication | Sensors | Abbreviated Journal | SENS |
Volume | 16 | Issue | 6 | Pages | 820 |
Keywords | Pedestrian Detection; FIR | ||||
Abstract | Despite all the significant advances in pedestrian detection brought by computer vision for driving assistance, it is still a challenging problem. One reason is the extremely varying lighting conditions under which such a detector should operate, namely day and night time. Recent research has shown that the combination of visible and non-visible imaging modalities may increase detection accuracy, where the infrared spectrum plays a critical role. The goal of this paper is to assess the accuracy gain of different pedestrian models (holistic, part-based, patch-based) when training with images in the far infrared spectrum. Specifically, we want to compare detection accuracy on test images recorded at day and nighttime if trained (and tested) using (a) plain color images, (b) just infrared images and (c) both of them. In order to obtain results for the last item we propose an early fusion approach to combine features from both modalities. We base the evaluation on a new dataset we have built for this purpose as well as on the publicly available KAIST multispectral dataset. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1424-8220 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes ![]() |
ADAS; 600.085; 600.076; 600.082; 601.281 | Approved | no | ||
Call Number | ADAS @ adas @ GFS2016 | Serial | 2754 | ||
Permanent link to this record | |||||
Author | Katerine Diaz; Aura Hernandez-Sabate; Antonio Lopez | ||||
Title | A reduced feature set for driver head pose estimation | Type | Journal Article | ||
Year | 2016 | Publication | Applied Soft Computing | Abbreviated Journal | ASOC |
Volume | 45 | Issue | Pages | 98-107 | |
Keywords | Head pose estimation; driving performance evaluation; subspace based methods; linear regression | ||||
Abstract | Evaluation of driving performance is of utmost importance in order to reduce road accident rate. Since driving ability includes visual-spatial and operational attention, among others, head pose estimation of the driver is a crucial indicator of driving performance. This paper proposes a new automatic method for coarse and fine head's yaw angle estimation of the driver. We rely on a set of geometric features computed from just three representative facial keypoints, namely the center of the eyes and the nose tip. With these geometric features, our method combines two manifold embedding methods and a linear regression one. In addition, the method has a confidence mechanism to decide if the classification of a sample is not reliable. The approach has been tested using the CMU-PIE dataset and our own driver dataset. Despite the very few facial keypoints required, the results are comparable to the state-of-the-art techniques. The low computational cost of the method and its robustness makes feasible to integrate it in massive consume devices as a real time application. | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes ![]() |
ADAS; 600.085; 600.076; | Approved | no | ||
Call Number | Admin @ si @ DHL2016 | Serial | 2760 | ||
Permanent link to this record | |||||
Author | Jose Manuel Alvarez; Antonio Lopez; Theo Gevers; Felipe Lumbreras | ||||
Title | Combining Priors, Appearance and Context for Road Detection | Type | Journal Article | ||
Year | 2014 | Publication | IEEE Transactions on Intelligent Transportation Systems | Abbreviated Journal | TITS |
Volume | 15 | Issue | 3 | Pages | 1168-1178 |
Keywords | Illuminant invariance; lane markings; road detection; road prior; road scene understanding; vanishing point; 3-D scene layout | ||||
Abstract | Detecting the free road surface ahead of a moving vehicle is an important research topic in different areas of computer vision, such as autonomous driving or car collision warning.
Current vision-based road detection methods are usually based solely on low-level features. Furthermore, they generally assume structured roads, road homogeneity, and uniform lighting conditions, constraining their applicability in real-world scenarios. In this paper, road priors and contextual information are introduced for road detection. First, we propose an algorithm to estimate road priors online using geographical information, providing relevant initial information about the road location. Then, contextual cues, including horizon lines, vanishing points, lane markings, 3-D scene layout, and road geometry, are used in addition to low-level cues derived from the appearance of roads. Finally, a generative model is used to combine these cues and priors, leading to a road detection method that is, to a large degree, robust to varying imaging conditions, road types, and scenarios. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1524-9050 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes ![]() |
ADAS; 600.076;ISE | Approved | no | ||
Call Number | Admin @ si @ ALG2014 | Serial | 2501 | ||
Permanent link to this record | |||||
Author | Miguel Oliveira; Victor Santos; Angel Sappa; P. Dias | ||||
Title | Scene Representations for Autonomous Driving: an approach based on polygonal primitives | Type | Conference Article | ||
Year | 2015 | Publication | 2nd Iberian Robotics Conference ROBOT2015 | Abbreviated Journal | |
Volume | 417 | Issue | Pages | 503-515 | |
Keywords | Scene reconstruction; Point cloud; Autonomous vehicles | ||||
Abstract | In this paper, we present a novel methodology to compute a 3D scene
representation. The algorithm uses macro scale polygonal primitives to model the scene. This means that the representation of the scene is given as a list of large scale polygons that describe the geometric structure of the environment. Results show that the approach is capable of producing accurate descriptions of the scene. In addition, the algorithm is very efficient when compared to other techniques. |
||||
Address | Lisboa; Portugal; November 2015 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ROBOT | ||
Notes ![]() |
ADAS; 600.076; 600.086 | Approved | no | ||
Call Number | Admin @ si @ OSS2015a | Serial | 2662 | ||
Permanent link to this record | |||||
Author | J.Poujol; Cristhian A. Aguilera-Carrasco; E.Danos; Boris X. Vintimilla; Ricardo Toledo; Angel Sappa | ||||
Title | Visible-Thermal Fusion based Monocular Visual Odometry | Type | Conference Article | ||
Year | 2015 | Publication | 2nd Iberian Robotics Conference ROBOT2015 | Abbreviated Journal | |
Volume | 417 | Issue | Pages | 517-528 | |
Keywords | Monocular Visual Odometry; LWIR-RGB cross-spectral Imaging; Image Fusion. | ||||
Abstract | The manuscript evaluates the performance of a monocular visual odometry approach when images from different spectra are considered, both independently and fused. The objective behind this evaluation is to analyze if classical approaches can be improved when the given images, which are from different spectra, are fused and represented in new domains. The images in these new domains should have some of the following properties: i) more robust to noisy data; ii) less sensitive to changes (e.g., lighting); iii) more rich in descriptive information, among other. In particular in the current work two different image fusion strategies are considered. Firstly, images from the visible and thermal spectrum are fused using a Discrete Wavelet Transform (DWT) approach. Secondly, a monochrome threshold strategy is considered. The obtained
representations are evaluated under a visual odometry framework, highlighting their advantages and disadvantages, using different urban and semi-urban scenarios. Comparisons with both monocular-visible spectrum and monocular-infrared spectrum, are also provided showing the validity of the proposed approach. |
||||
Address | Lisboa; Portugal; November 2015 | ||||
Corporate Author | Thesis | ||||
Publisher | Springer International Publishing | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 2194-5357 | ISBN | 978-3-319-27145-3 | Medium | |
Area | Expedition | Conference | ROBOT | ||
Notes ![]() |
ADAS; 600.076; 600.086 | Approved | no | ||
Call Number | Admin @ si @ PAD2015 | Serial | 2663 | ||
Permanent link to this record | |||||
Author | Cesar de Souza; Adrien Gaidon; Yohann Cabon; Antonio Lopez | ||||
Title | Procedural Generation of Videos to Train Deep Action Recognition Networks | Type | Conference Article | ||
Year | 2017 | Publication | 30th IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 2594-2604 | ||
Keywords | |||||
Abstract | Deep learning for human action recognition in videos is making significant progress, but is slowed down by its dependency on expensive manual labeling of large video collections. In this work, we investigate the generation of synthetic training data for action recognition, as it has recently shown promising results for a variety of other computer vision tasks. We propose an interpretable parametric generative model of human action videos that relies on procedural generation and other computer graphics techniques of modern game engines. We generate a diverse, realistic, and physically plausible dataset of human action videos, called PHAV for ”Procedural Human Action Videos”. It contains a total of 39, 982 videos, with more than 1, 000 examples for each action of 35 categories. Our approach is not limited to existing motion capture sequences, and we procedurally define 14 synthetic actions. We introduce a deep multi-task representation learning architecture to mix synthetic and real videos, even if the action categories differ. Our experiments on the UCF101 and HMDB51 benchmarks suggest that combining our large set of synthetic videos with small real-world datasets can boost recognition performance, significantly
outperforming fine-tuning state-of-the-art unsupervised generative models of videos. |
||||
Address | Honolulu; Hawaii; July 2017 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | CVPR | ||
Notes ![]() |
ADAS; 600.076; 600.085; 600.118 | Approved | no | ||
Call Number | Admin @ si @ SGC2017 | Serial | 3051 | ||
Permanent link to this record | |||||
Author | Cesar de Souza; Adrien Gaidon; Eleonora Vig; Antonio Lopez | ||||
Title | Sympathy for the Details: Dense Trajectories and Hybrid Classification Architectures for Action Recognition | Type | Conference Article | ||
Year | 2016 | Publication | 14th European Conference on Computer Vision | Abbreviated Journal | |
Volume | Issue | Pages | 697-716 | ||
Keywords | |||||
Abstract | Action recognition in videos is a challenging task due to the complexity of the spatio-temporal patterns to model and the difficulty to acquire and learn on large quantities of video data. Deep learning, although a breakthrough for image classification and showing promise for videos, has still not clearly superseded action recognition methods using hand-crafted features, even when training on massive datasets. In this paper, we introduce hybrid video classification architectures based on carefully designed unsupervised representations of hand-crafted spatio-temporal features classified by supervised deep networks. As we show in our experiments on five popular benchmarks for action recognition, our hybrid model combines the best of both worlds: it is data efficient (trained on 150 to 10000 short clips) and yet improves significantly on the state of the art, including recent deep models trained on millions of manually labelled images and videos. | ||||
Address | Amsterdam; The Netherlands; October 2016 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | LNCS | ||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ECCV | ||
Notes ![]() |
ADAS; 600.076; 600.085 | Approved | no | ||
Call Number | Admin @ si @ SGV2016 | Serial | 2824 | ||
Permanent link to this record | |||||
Author | Victor Campmany; Sergio Silva; Juan Carlos Moure; Antoni Espinosa; David Vazquez; Antonio Lopez | ||||
Title | GPU-based pedestrian detection for autonomous driving | Type | Abstract | ||
Year | 2015 | Publication | Programming and Tunning Massive Parallel Systems | Abbreviated Journal | PUMPS |
Volume | Issue | Pages | |||
Keywords | Autonomous Driving; ADAS; CUDA; Pedestrian Detection | ||||
Abstract | Pedestrian detection for autonomous driving has gained a lot of prominence during the last few years. Besides the fact that it is one of the hardest tasks within computer vision, it involves huge computational costs. The real-time constraints in the field are tight, and regular processors are not able to handle the workload obtaining an acceptable ratio of frames per second (fps). Moreover, multiple cameras are required to obtain accurate results, so the need to speed up the process is even higher. Taking the work in [1] as our baseline, we propose a CUDA implementation of a pedestrian detection system. Further, we introduce significant algorithmic adjustments and optimizations to adapt the problem to the GPU architecture. The aim is to provide a system capable of running in real-time obtaining reliable results. | ||||
Address | Barcelona; Spain | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | PUMPS | ||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | PUMPS | ||
Notes ![]() |
ADAS; 600.076; 600.082; 600.085 | Approved | no | ||
Call Number | ADAS @ adas @ CSM2015 | Serial | 2644 | ||
Permanent link to this record |