Records |
Author |
Mohamed Ramzy Ibrahim; Robert Benavente; Daniel Ponsa; Felipe Lumbreras |
Title |
Unveiling the Influence of Image Super-Resolution on Aerial Scene Classification |
Type |
Conference Article |
Year |
2023 |
Publication |
Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications |
Abbreviated Journal |
|
Volume |
14469 |
Issue |
|
Pages |
214–228 |
Keywords |
|
Abstract |
Deep learning has made significant advances in recent years, and as a result, it is now in a stage where it can achieve outstanding results in tasks requiring visual understanding of scenes. However, its performance tends to decline when dealing with low-quality images. The advent of super-resolution (SR) techniques has started to have an impact on the field of remote sensing by enabling the restoration of fine details and enhancing image quality, which could help to increase performance in other vision tasks. However, in previous works, contradictory results for scene visual understanding were achieved when SR techniques were applied. In this paper, we present an experimental study on the impact of SR on enhancing aerial scene classification. Through the analysis of different state-of-the-art SR algorithms, including traditional methods and deep learning-based approaches, we unveil the transformative potential of SR in overcoming the limitations of low-resolution (LR) aerial imagery. By enhancing spatial resolution, more fine details are captured, opening the door for an improvement in scene understanding. We also discuss the effect of different image scales on the quality of SR and its effect on aerial scene classification. Our experimental work demonstrates the significant impact of SR on enhancing aerial scene classification compared to LR images, opening new avenues for improved remote sensing applications. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
LNCS |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
CIARP |
Notes |
MSIAU |
Approved |
no |
Call Number |
Admin @ si @ IBP2023 |
Serial |
4008 |
Permanent link to this record |
|
|
|
Author |
Victor Campmany; Sergio Silva; Juan Carlos Moure; Antoni Espinosa; David Vazquez; Antonio Lopez |
Title |
GPU-based pedestrian detection for autonomous driving |
Type |
Abstract |
Year |
2015 |
Publication |
Programming and Tunning Massive Parallel Systems |
Abbreviated Journal |
PUMPS |
Volume |
|
Issue |
|
Pages |
|
Keywords |
Autonomous Driving; ADAS; CUDA; Pedestrian Detection |
Abstract |
Pedestrian detection for autonomous driving has gained a lot of prominence during the last few years. Besides the fact that it is one of the hardest tasks within computer vision, it involves huge computational costs. The real-time constraints in the field are tight, and regular processors are not able to handle the workload obtaining an acceptable ratio of frames per second (fps). Moreover, multiple cameras are required to obtain accurate results, so the need to speed up the process is even higher. Taking the work in [1] as our baseline, we propose a CUDA implementation of a pedestrian detection system. Further, we introduce significant algorithmic adjustments and optimizations to adapt the problem to the GPU architecture. The aim is to provide a system capable of running in real-time obtaining reliable results. |
Address |
Barcelona; Spain |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
PUMPS |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
PUMPS |
Notes |
ADAS; 600.076; 600.082; 600.085 |
Approved |
no |
Call Number |
ADAS @ adas @ CSM2015 |
Serial |
2644 |
Permanent link to this record |
|
|
|
Author |
Michael Teutsch; Angel Sappa; Riad I. Hammoud |
Title |
Cross-Spectral Image Processing |
Type |
Book Chapter |
Year |
2022 |
Publication |
Computer Vision in the Infrared Spectrum. Synthesis Lectures on Computer Vision |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
23-34 |
Keywords |
|
Abstract |
Although this book is on IR computer vision and its main focus lies on IR image and video processing and analysis, a special attention is dedicated to cross-spectral image processing due to the increasing number of publications and applications in this domain. In these cross-spectral frameworks, IR information is used together with information from other spectral bands to tackle some specific problems by developing more robust solutions. Tasks considered for cross-spectral processing are for instance dehazing, segmentation, vegetation index estimation, or face recognition. This increasing number of applications is motivated by cross- and multi-spectral camera setups available already on the market like for example smartphones, remote sensing multispectral cameras, or multi-spectral cameras for automotive systems or drones. In this chapter, different cross-spectral image processing techniques will be reviewed together with possible applications. Initially, image registration approaches for the cross-spectral case are reviewed: the registration stage is the first image processing task, which is needed to align images acquired by different sensors within the same reference coordinate system. Then, recent cross-spectral image colorization approaches, which are intended to colorize infrared images for different applications are presented. Finally, the cross-spectral image enhancement problem is tackled by including guided super resolution techniques, image dehazing approaches, cross-spectral filtering and edge detection. Figure 3.1 illustrates cross-spectral image processing stages as well as their possible connections. Table 3.1 presents some of the available public cross-spectral datasets generally used as reference data to evaluate cross-spectral image registration, colorization, enhancement, or exploitation results. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SLCV |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
978-3-031-00698-2 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
MSIAU; MACO |
Approved |
no |
Call Number |
Admin @ si @ TSH2022b |
Serial |
3805 |
Permanent link to this record |
|
|
|
Author |
Michael Teutsch; Angel Sappa; Riad I. Hammoud |
Title |
Detection, Classification, and Tracking |
Type |
Book Chapter |
Year |
2022 |
Publication |
Computer Vision in the Infrared Spectrum. Synthesis Lectures on Computer Vision |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
35-58 |
Keywords |
|
Abstract |
Automatic image and video exploitation or content analysis is a technique to extract higher-level information from a scene such as objects, behavior, (inter-)actions, environment, or even weather conditions. The relevant information is assumed to be contained in the two-dimensional signal provided in an image (width and height in pixels) or the three-dimensional signal provided in a video (width, height, and time). But also intermediate-level information such as object classes [196], locations [197], or motion [198] can help applications to fulfill certain tasks such as intelligent compression [199], video summarization [200], or video retrieval [201]. Usually, videos with their temporal dimension are a richer source of data compared to single images [202] and thus certain video content can be extracted from videos only such as object motion or object behavior. Often, machine learning or nowadays deep learning techniques are utilized to model prior knowledge about object or scene appearance using labeled training samples [203, 204]. After a learning phase, these models are then applied in real world applications, which is called inference. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SLCV |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
978-3-031-00698-2 |
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
MSIAU; MACO |
Approved |
no |
Call Number |
Admin @ si @ TSH2022c |
Serial |
3806 |
Permanent link to this record |
|
|
|
Author |
Michael Teutsch; Angel Sappa; Riad I. Hammoud |
Title |
Image and Video Enhancement |
Type |
Book Chapter |
Year |
2022 |
Publication |
Computer Vision in the Infrared Spectrum. Synthesis Lectures on Computer Vision |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
9-21 |
Keywords |
|
Abstract |
Image and video enhancement aims at improving the signal quality relative to imaging artifacts such as noise and blur or atmospheric perturbations such as turbulence and haze. It is usually performed in order to assist humans in analyzing image and video content or simply to present humans visually appealing images and videos. However, image and video enhancement can also be used as a preprocessing technique to ease the task and thus improve the performance of subsequent automatic image content analysis algorithms: preceding dehazing can improve object detection as shown by [23] or explicit turbulence modeling can improve moving object detection as discussed by [24]. But it remains an open question whether image and video enhancement should rather be performed explicitly as a preprocessing step or implicitly for example by feeding affected images directly to a neural network for image content analysis like object detection [25]. Especially for real-time video processing at low latency it can be better to handle image perturbation implicitly in order to minimize the processing time of an algorithm. This can be achieved by making algorithms for image content analysis robust or even invariant to perturbations such as noise or blur. Additionally, mistakes of an individual preprocessing module can obviously affect the quality of the entire processing pipeline. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SLCV |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
MSIAU; MACO |
Approved |
no |
Call Number |
Admin @ si @ TSH2022a |
Serial |
3807 |
Permanent link to this record |
|
|
|
Author |
Jun Wan; Guodong Guo; Sergio Escalera; Hugo Jair Escalante; Stan Z Li |
Title |
Face Presentation Attack Detection (PAD) Challenges |
Type |
Book Chapter |
Year |
2023 |
Publication |
Advances in Face Presentation Attack Detection |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
17–35 |
Keywords |
|
Abstract |
In recent years, the security of face recognition systems has been increasingly threatened. Face Anti-spoofing (FAS) is essential to secure face recognition systems primarily from various attacks. In order to attract researchers and push forward the state of the art in Face Presentation Attack Detection (PAD), we organized three editions of Face Anti-spoofing Workshop and Competition at CVPR 2019, CVPR 2020, and ICCV 2021, which have attracted more than 800 teams from academia and industry, and greatly promoted the algorithms to overcome many challenging problems. In this chapter, we introduce the detailed competition process, including the challenge phases, timeline and evaluation metrics. Along with the workshop, we will introduce the corresponding dataset for each competition including data acquisition details, data processing, statistics, and evaluation protocol. Finally, we provide the available link to download the datasets used in the challenges. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SLCV |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
HUPBA |
Approved |
no |
Call Number |
Admin @ si @ WGE2023b |
Serial |
3956 |
Permanent link to this record |
|
|
|
Author |
Jun Wan; Guodong Guo; Sergio Escalera; Hugo Jair Escalante; Stan Z Li |
Title |
Face Anti-spoofing Progress Driven by Academic Challenges |
Type |
Book Chapter |
Year |
2023 |
Publication |
Advances in Face Presentation Attack Detection |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
1–15 |
Keywords |
|
Abstract |
With the ubiquity of facial authentication systems and the prevalence of security cameras around the world, the impact that facial presentation attack techniques may have is huge. However, research progress in this field has been slowed by a number of factors, including the lack of appropriate and realistic datasets, ethical and privacy issues that prevent the recording and distribution of facial images, the little attention that the community has given to potential ethnic biases among others. This chapter provides an overview of contributions derived from the organization of academic challenges in the context of face anti-spoofing detection. Specifically, we discuss the limitations of benchmarks and summarize our efforts in trying to boost research by the community via the participation in academic challenges |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SLCV |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
HUPBA |
Approved |
no |
Call Number |
Admin @ si @ WGE2023c |
Serial |
3957 |
Permanent link to this record |
|
|
|
Author |
Isabelle Guyon; Lisheng Sun Hosoya; Marc Boulle; Hugo Jair Escalante; Sergio Escalera; Zhengying Liu; Damir Jajetic; Bisakha Ray; Mehreen Saeed; Michele Sebag; Alexander R.Statnikov; Wei-Wei Tu; Evelyne Viegas |
Title |
Analysis of the AutoML Challenge Series 2015-2018. |
Type |
Book Chapter |
Year |
2019 |
Publication |
Automated Machine Learning |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
177-219 |
Keywords |
|
Abstract |
The ChaLearn AutoML Challenge (The authors are in alphabetical order of last name, except the first author who did most of the writing and the second author who produced most of the numerical analyses and plots.) (NIPS 2015 – ICML 2016) consisted of six rounds of a machine learning competition of progressive difficulty, subject to limited computational resources. It was followed bya one-round AutoML challenge (PAKDD 2018). The AutoML setting differs from former model selection/hyper-parameter selection challenges, such as the one we previously organized for NIPS 2006: the participants aim to develop fully automated and computationally efficient systems, capable of being trained and tested without human intervention, with code submission. This chapter analyzes the results of these competitions and provides details about the datasets, which were not revealed to the participants. The solutions of the winners are systematically benchmarked over all datasets of all rounds and compared with canonical machine learning algorithms available in scikit-learn. All materials discussed in this chapter (data and code) have been made publicly available at http://automl.chalearn.org/. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Springer |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title ![sorted by Abbreviated Series Title field, ascending order (up)](img/sort_asc.gif) |
SSCML |
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
HuPBA; no proj |
Approved |
no |
Call Number |
Admin @ si @ GHB2019 |
Serial |
3330 |
Permanent link to this record |