|
Zhong Jin, Zhen Lou, Jing-Yu Yang, & Quan-sen Sun. (2007). Face Detection using Template Matching and Skin-color Information. Neurocomputing, 70(4–6): 794–800.
|
|
|
Zeynep Yucel, Albert Ali Salah, Çetin Meriçli, Tekin Meriçli, Roberto Valenti, & Theo Gevers. (2013). Joint Attention by Gaze Interpolation and Saliency. T-CIBER - IEEE Transactions on cybernetics, 829–842.
Abstract: Joint attention, which is the ability of coordination of a common point of reference with the communicating party, emerges as a key factor in various interaction scenarios. This paper presents an image-based method for establishing joint attention between an experimenter and a robot. The precise analysis of the experimenter's eye region requires stability and high-resolution image acquisition, which is not always available. We investigate regression-based interpolation of the gaze direction from the head pose of the experimenter, which is easier to track. Gaussian process regression and neural networks are contrasted to interpolate the gaze direction. Then, we combine gaze interpolation with image-based saliency to improve the target point estimates and test three different saliency schemes. We demonstrate the proposed method on a human-robot interaction scenario. Cross-subject evaluations, as well as experiments under adverse conditions (such as dimmed or artificial illumination or motion blur), show that our method generalizes well and achieves rapid gaze estimation for establishing joint attention.
|
|
|
Yu Jie, Jaume Amores, N. Sebe, Petia Radeva, & Tian Qi. (2008). Distance Learning for Similarity Estimation. IEEE Trans. on Pattern Analysis and Machine Intelligence, vol.30(3):451–462.
|
|
|
Yong Xu, Jing-Yu Yang, & Zhong Jin. (2003). Theory analysis on FSLDA and ULDA. Pattern Recognition, 36(12): 3031–3033 (IF: 1.611).
|
|
|
Yong Xu, Jing-Yu Yang, & Zhong Jin. (2004). A novel method for Fisher discriminant analysis. Pattern Recognition, 37(2):381–384 (IF: 2.176).
|
|
|
Yasuko Sugito, Trevor Canham, Javier Vazquez, & Marcelo Bertalmio. (2021). A Study of Objective Quality Metrics for HLG-Based HDR/WCG Image Coding. SMPTE - SMPTE Motion Imaging Journal, 53–65.
Abstract: In this work, we study the suitability of high dynamic range, wide color gamut (HDR/WCG) objective quality metrics to assess the perceived deterioration of compressed images encoded using the hybrid log-gamma (HLG) method, which is the standard for HDR television. Several image quality metrics have been developed to deal specifically with HDR content, although in previous work we showed that the best results (i.e., better matches to the opinion of human expert observers) are obtained by an HDR metric that consists simply in applying a given standard dynamic range metric, called visual information fidelity (VIF), directly to HLG-encoded images. However, all these HDR metrics ignore the chroma components for their calculations, that is, they consider only the luminance channel. For this reason, in the current work, we conduct subjective evaluation experiments in a professional setting using compressed HDR/WCG images encoded with HLG and analyze the ability of the best HDR metric to detect perceivable distortions in the chroma components, as well as the suitability of popular color metrics (including ΔITPR , which supports parameters for HLG) to correlate with the opinion scores. Our first contribution is to show that there is a need to consider the chroma components in HDR metrics, as there are color distortions that subjects perceive but that the best HDR metric fails to detect. Our second contribution is the surprising result that VIF, which utilizes only the luminance channel, correlates much better with the subjective evaluation scores than the metrics investigated that do consider the color components.
|
|
|
Xose M. Pardo, Petia Radeva, & D. Cabello. (2003). Discriminant Snakes for 3D Reconstruction of Anatomical Organs. Medical Image Analysis, 7(3): 293–310 (IF: 4.442).
|
|
|
Xim Cerda-Company, Olivier Penacchio, & Xavier Otazu. (2021). Chromatic Induction in Migraine. VISION, 37.
Abstract: The human visual system is not a colorimeter. The perceived colour of a region does not only depend on its colour spectrum, but also on the colour spectra and geometric arrangement of neighbouring regions, a phenomenon called chromatic induction. Chromatic induction is thought to be driven by lateral interactions: the activity of a central neuron is modified by stimuli outside its classical receptive field through excitatory–inhibitory mechanisms. As there is growing evidence of an excitation/inhibition imbalance in migraine, we compared chromatic induction in migraine and control groups. As hypothesised, we found a difference in the strength of induction between the two groups, with stronger induction effects in migraine. On the other hand, given the increased prevalence of visual phenomena in migraine with aura, we also hypothesised that the difference between migraine and control would be more important in migraine with aura than in migraine without aura. Our experiments did not support this hypothesis. Taken together, our results suggest a link between excitation/inhibition imbalance and increased induction effects.
Keywords: migraine; vision; colour; colour perception; chromatic induction; psychophysics
|
|
|
Xiangyang Li, Luis Herranz, & Shuqiang Jiang. (2020). Multifaceted Analysis of Fine-Tuning in Deep Model for Visual Recognition. ACM - ACM Transactions on Data Science.
Abstract: In recent years, convolutional neural networks (CNNs) have achieved impressive performance for various visual recognition scenarios. CNNs trained on large labeled datasets can not only obtain significant performance on most challenging benchmarks but also provide powerful representations, which can be used to a wide range of other tasks. However, the requirement of massive amounts of data to train deep neural networks is a major drawback of these models, as the data available is usually limited or imbalanced. Fine-tuning (FT) is an effective way to transfer knowledge learned in a source dataset to a target task. In this paper, we introduce and systematically investigate several factors that influence the performance of fine-tuning for visual recognition. These factors include parameters for the retraining procedure (e.g., the initial learning rate of fine-tuning), the distribution of the source and target data (e.g., the number of categories in the source dataset, the distance between the source and target datasets) and so on. We quantitatively and qualitatively analyze these factors, evaluate their influence, and present many empirical observations. The results reveal insights into what fine-tuning changes CNN parameters and provide useful and evidence-backed intuitions about how to implement fine-tuning for computer vision tasks.
|
|
|
Xavier Otazu, Maria Vanrell, & C. Alejandro Parraga. (2007). Mutiresolution Wavelet Framework Reproduces Induction Effects. Perception 36:167–167, supp.
|
|
|
Xavier Otazu, Maria Vanrell, & C. Alejandro Parraga. (2008). Multiresolution Wavelet Framework Models Brightness Induction Effects. VR - Vision Research, 733–751.
|
|
|
Xavier Otazu, Maria Vanrell, & C. Alejandro Parraga. (2008). Colour induction effects are modelled by a low-level multiresolution wavelet framework. Perception 37(Suppl.): 107.
|
|
|
Xavier Otazu, & Maria Vanrell. (2005). Perceptual representation of textured images. Journal of Imaging Science and Technology, 49(3):262–271 (IF: 0.522).
|
|
|
Xavier Otazu, & Maria Vanrell. (2006). Several lightness induction effects with a computational multiresolution wavelet framework. 29th European Conference on Visual Perception (ECVP’06), Perception Suppl s, 32: 56–56.
|
|
|
Xavier Otazu, M. Ribo, M. Peracaula, J.M. Paredes, & J. Nuñez. (2002). Detection of superimposed periodic signals using wavelets. Monthly Notices of the Royal Astronomical Society, 333, 2: 365–372 (IF: 4.671).
|
|