|
Laura Igual, & Xavier Baro. (2013). Experiencia de aprendizaje de programación basada en proyectos. Simposio-Taller Estrategias y herramientas para el aprendizaje y la evaluación.
|
|
|
Hugo Jair Escalante, Heysem Kaya, Albert Ali Salah, Sergio Escalera, Yagmur Gucluturk, Umut Guclu, et al. (2018). Explaining First Impressions: Modeling, Recognizing, and Explaining Apparent Personality from Videos.
Abstract: Explainability and interpretability are two critical aspects of decision support systems. Within computer vision, they are critical in certain tasks related to human behavior analysis such as in health care applications. Despite their importance, it is only recently that researchers are starting to explore these aspects. This paper provides an introduction to explainability and interpretability in the context of computer vision with an emphasis on looking at people tasks. Specifically, we review and study those mechanisms in the context of first impressions analysis. To the best of our knowledge, this is the first effort in this direction. Additionally, we describe a challenge we organized on explainability in first impressions analysis from video. We analyze in detail the newly introduced data set, the evaluation protocol, and summarize the results of the challenge. Finally, derived from our study, we outline research opportunities that we foresee will be decisive in the near future for the development of the explainable computer vision field.
|
|
|
Chuanming Tang, Kai Wang, Joost van de Weijer, Jianlin Zhang, & Yongmei Huang. (2023). Exploiting Image-Related Inductive Biases in Single-Branch Visual Tracking.
Abstract: Despite achieving state-of-the-art performance in visual tracking, recent single-branch trackers tend to overlook the weak prior assumptions associated with the Vision Transformer (ViT) encoder and inference pipeline. Moreover, the effectiveness of discriminative trackers remains constrained due to the adoption of the dual-branch pipeline. To tackle the inferior effectiveness of the vanilla ViT, we propose an Adaptive ViT Model Prediction tracker (AViTMP) to bridge the gap between single-branch network and discriminative models. Specifically, in the proposed encoder AViT-Enc, we introduce an adaptor module and joint target state embedding to enrich the dense embedding paradigm based on ViT. Then, we combine AViT-Enc with a dense-fusion decoder and a discriminative target model to predict accurate location. Further, to mitigate the limitations of conventional inference practice, we present a novel inference pipeline called CycleTrack, which bolsters the tracking robustness in the presence of distractors via bidirectional cycle tracking verification. Lastly, we propose a dual-frame update inference strategy that adeptively handles significant challenges in long-term scenarios. In the experiments, we evaluate AViTMP on ten tracking benchmarks for a comprehensive assessment, including LaSOT, LaSOTExtSub, AVisT, etc. The experimental results unequivocally establish that AViTMP attains state-of-the-art performance, especially on long-time tracking and robustness.
|
|
|
Soumick Chatterjee, Fatima Saad, Chompunuch Sarasaen, Suhita Ghosh, Rupali Khatun, Petia Radeva, et al. (2020). Exploration of Interpretability Techniques for Deep COVID-19 Classification using Chest X-ray Images.
Abstract: CoRR abs/2006.02570
The outbreak of COVID-19 has shocked the entire world with its fairly rapid spread and has challenged different sectors. One of the most effective ways to limit its spread is the early and accurate diagnosis of infected patients. Medical imaging such as X-ray and Computed Tomography (CT) combined with the potential of Artificial Intelligence (AI) plays an essential role in supporting the medical staff in the diagnosis process. Thereby, the use of five different deep learning models (ResNet18, ResNet34, InceptionV3, InceptionResNetV2, and DenseNet161) and their Ensemble have been used in this paper, to classify COVID-19, pneumoniæ and healthy subjects using Chest X-Ray. Multi-label classification was performed to predict multiple pathologies for each patient, if present. Foremost, the interpretability of each of the networks was thoroughly studied using techniques like occlusion, saliency, input X gradient, guided backpropagation, integrated gradients, and DeepLIFT. The mean Micro-F1 score of the models for COVID-19 classifications ranges from 0.66 to 0.875, and is 0.89 for the Ensemble of the network models. The qualitative results depicted the ResNets to be the most interpretable model.
|
|
|
Georg Langs, Petia Radeva, David Rotger, & Francesc Carreras. (2004). Explorative Building of 3D Vessel Tree Models.
|
|
|
Francesc Tous, Maria Vanrell, & Ramon Baldrich. (2004). Exploring Colour Constancy Solutions..
|
|
|
Antonio Lopez, Ricardo Toledo, Joan Serrat, & Juan J. Villanueva. (1999). Extraction of vessel centerlines from 2D coronary angiographies.
|
|
|
Francisco Javier Orozco, Pau Baiget, Jordi Gonzalez, & Xavier Roca. (2006). Eyelids and Face Tracking in Real-Time.
|
|
|
Zhong Jin, Zhen Lou, Jing-Yu Yang, & Quan-sen Sun. (2005). Face detection using template matching and skin color information.
|
|
|
Agata Lapedriza, David Masip, & Jordi Vitria. (2006). Face Verification using External Features.
|
|
|
Zhong Jin, Franck Davoine, & Zhen Lou. (2003). Facial expression analysis by using KPCA.
|
|
|
Fadi Dornaika, & Franck Davoine. (2005). Facial expression recognition in continuous videos using dynamic programming.
|
|
|
Fadi Dornaika, & Franck Davoine. (2006). Facial expression recognition using auto-regressive models.
|
|
|
Petia Radeva, & Enric Marti. (1995). Facial Features Segmentation by Model-Based Snakes..
|
|
|
J. Filipe, Juan Andrade, & J.L. Ferrier. (2005). FAF 2005.
|
|