|
Gemma Sanchez, & Josep Llados. (2003). Syntactic models to represent perceptually regular repetitive patterns in graphic documents.
|
|
|
Philippe Dosch, & Josep Llados. (2003). Vectorial Signatures for Symbol Discrimination.
|
|
|
David Rotger, Petia Radeva, Cristina Cañero, Juan J. Villanueva, J. Mauri, E Fernandez-Nofrerias, et al. (2001). Corresponding IVUS and Angiogram Image Data.
|
|
|
Oriol Pujol, David Rotger, Petia Radeva, O. Rodriguez, & J. Mauri. (2003). Near Real Time Plaque Segmentation of IVUS.
|
|
|
Petia Radeva, & Enric Marti. (1995). An improved model of snakes for model-based segmentation. In Proceedings of Computer Analysis of Images and Patterns (pp. 515–520).
Abstract: The main advantage of segmentation by snakes consists in its ability to incorporate smoothness constraints on the detected shapes that can occur. Likewise, we propose to model snakes with other properties that reflect the information provided about the object of interest in a different extent. We consider different kinds of snakes, those searching for contours with a certain direction, those preserving an object’s model, those seeking for symmetry, those expanding open, etc. The availability of such a collection of snakes allows not only the more complete use of the knowledge about the segmented object, but also to solve some problems of the existing snakes. Our experiments on segmentation of facial features justify the usefulness of snakes with different properties.
|
|
|
Debora Gil, Petia Radeva, Jordi Saludes, & Josefina Mauri. (2000). Automatic Segmentation of Artery Wall in Coronary IVUS Images: a Probabilistic Approach. In Proceedings of CIC’2000. Cambridge, Massachussets.
Abstract: Intravascular ultrasound images represent a unique tool to analyze the morphology of arteries and vessels (plaques, restenosis, etc). The poor quality of these images makes unsupervised segmentation based on traditional segmentation algorithms (such as edge or ridge/valley detection) fail to achieve the expected results. In this paper we present a probabilistic flexible template to separate different regions in the image. In particular, we use elliptic templates to model and detect the shape of the vessel inner wall in IVUS images. We present the results of successful segmentation obtained from patients undergoing stent treatment. A physician team has validated these results.
|
|
|
A. Martinez, S. Gonzalez, Jordi Vitria, & J. Lopez. (1997). NAT: a robot that recognizes offices..
|
|
|
Gloria Fernandez Esparrach, Jorge Bernal, Cristina Rodriguez de Miguel, Debora Gil, Fernando Vilariño, Henry Cordova, et al. (2015). Colonic polyps are correctly identified by a computer vision method using wm-dova energy maps. In Proceedings of 23 United European- UEG Week 2015.
|
|
|
Francisco Jose Perales, Juan J. Villanueva, & Y. Luo. (1991). An automatic two-camera human motion perception system based on biomechanical model matching..
|
|
|
Josep Llados, & Gemma Sanchez. (2003). Symbol Recognition Using Graphs.
|
|
|
Petia Radeva, & Jordi Vitria. (2001). Region Based Approach for Discriminant Snakes..
|
|
|
Emanuele Vivoli, Ali Furkan Biten, Andres Mafla, Dimosthenis Karatzas, & Lluis Gomez. (2022). MUST-VQA: MUltilingual Scene-text VQA. In Proceedings European Conference on Computer Vision Workshops (Vol. 13804, 345–358). LNCS.
Abstract: In this paper, we present a framework for Multilingual Scene Text Visual Question Answering that deals with new languages in a zero-shot fashion. Specifically, we consider the task of Scene Text Visual Question Answering (STVQA) in which the question can be asked in different languages and it is not necessarily aligned to the scene text language. Thus, we first introduce a natural step towards a more generalized version of STVQA: MUST-VQA. Accounting for this, we discuss two evaluation scenarios in the constrained setting, namely IID and zero-shot and we demonstrate that the models can perform on a par on a zero-shot setting. We further provide extensive experimentation and show the effectiveness of adapting multilingual language models into STVQA tasks.
Keywords: Visual question answering; Scene text; Translation robustness; Multilingual models; Zero-shot transfer; Power of language models
|
|
|
Sergi Garcia Bordils, Andres Mafla, Ali Furkan Biten, Oren Nuriel, Aviad Aberdam, Shai Mazor, et al. (2022). Out-of-Vocabulary Challenge Report. In Proceedings European Conference on Computer Vision Workshops (Vol. 13804, 359–375). LNCS.
Abstract: This paper presents final results of the Out-Of-Vocabulary 2022 (OOV) challenge. The OOV contest introduces an important aspect that is not commonly studied by Optical Character Recognition (OCR) models, namely, the recognition of unseen scene text instances at training time. The competition compiles a collection of public scene text datasets comprising of 326,385 images with 4,864,405 scene text instances, thus covering a wide range of data distributions. A new and independent validation and test set is formed with scene text instances that are out of vocabulary at training time. The competition was structured in two tasks, end-to-end and cropped scene text recognition respectively. A thorough analysis of results from baselines and different participants is presented. Interestingly, current state-of-the-art models show a significant performance gap under the newly studied setting. We conclude that the OOV dataset proposed in this challenge will be an essential area to be explored in order to develop scene text models that achieve more robust and generalized predictions.
|
|
|
Robert Benavente, & Maria Vanrell. (2007). Parametrizacion del Espacio de Categorias de Color.
|
|
|
David Guillamet, & Jordi Vitria. (2001). Unsupervised Learning of Part-Based Representations.
|
|