|
David Roche. (2015)." A Statistical Framework for Terminating Evolutionary Algorithms at their Steady State" (Debora Gil, & Jesus Giraldo, Eds.). Ph.D. thesis, Ediciones Graficas Rey, .
Abstract: As any iterative technique, it is a necessary condition a stop criterion for terminating Evolutionary Algorithms (EA). In the case of optimization methods, the algorithm should stop at the time it has reached a steady state so it can not improve results anymore. Assessing the reliability of termination conditions for EAs is of prime importance. A wrong or weak stop criterion can negatively aect both the computational eort and the nal result.
In this Thesis, we introduce a statistical framework for assessing whether a termination condition is able to stop EA at its steady state. In one hand a numeric approximation to steady states to detect the point in which EA population has lost its diversity has been presented for EA termination. This approximation has been applied to dierent EA paradigms based on diversity and a selection of functions covering the properties most relevant for EA convergence. Experiments show that our condition works regardless of the search space dimension and function landscape and Dierential Evolution (DE) arises as the best paradigm. On the other hand, we use a regression model in order to determine the requirements ensuring that a measure derived from EA evolving population is related to the distance to the optimum in xspace.
Our theoretical framework is analyzed across several benchmark test functions
and two standard termination criteria based on function improvement in f-space and EA population x-space distribution for the DE paradigm. Results validate our statistical framework as a powerful tool for determining the capability of a measure for terminating EA and select the x-space distribution as the best-suited for accurately stopping DE in real-world applications.
|
|
|
Miquel Angel Piera, Jose Luis Muñoz, Debora Gil, Gonzalo Martin, & Jordi Manzano. (2022). "A Socio-Technical Simulation Model for the Design of the Future Single Pilot Cockpit: An Opportunity to Improve Pilot Performance " . IEEE Access, 10, 22330–22343.
Abstract: The future deployment of single pilot operations must be supported by new cockpit computer services. Such services require an adaptive context-aware integration of technical functionalities with the concurrent tasks that a pilot must deal with. Advanced artificial intelligence supporting services and improved communication capabilities are the key enabling technologies that will render future cockpits more integrated with the present digitalized air traffic management system. However, an issue in the integration of such technologies is the lack of socio-technical analysis in the design of these teaming mechanisms. A key factor in determining how and when a service support should be provided is the dynamic evolution of pilot workload. This paper investigates how the socio-technical model-based systems engineering approach paves the way for the design of a digital assistant framework by formalizing this workload. The model was validated in an Airbus A-320 cockpit simulator, and the results confirmed the degraded pilot behavioral model and the performance impact according to different contextual flight deck information. This study contributes to practical knowledge for designing human-machine task-sharing systems.
Keywords: Human factors ; Performance evaluation ; Simulation; Sociotechnical systems ; System performance
|
|
|
Petia Radeva, Joan Serrat, & Enric Marti. (1995). "A snake for model-based segmentation " In Proc. Conf. Fifth Int Computer Vision (pp. 816–821).
Abstract: Despite the promising results of numerous applications, the hitherto proposed snake techniques share some common problems: snake attraction by spurious edge points, snake degeneration (shrinking and attening), convergence and stability of the deformation process, snake initialization and local determination of the parameters of elasticity. We argue here that these problems can be solved only when all the snake aspects are considered. The snakes proposed here implement a new potential eld and external force in order to provide a deformation convergence, attraction by both near and far edges as well as snake behaviour selective according to the edge orientation. Furthermore, we conclude that in the case of model-based seg mentation, the internal force should include structural information about the expected snake shape. Experiments using this kind of snakes for segmenting bones in complex hand radiographs show a signicant improvement.
Keywords: snakes; elastic matching; model-based segmenta tion
|
|
|
Debora Gil, Jaume Garcia, Ruth Aris, Guillaume Houzeaux, & Manuel Vazquez. (2009). "A Riemmanian approach to cardiac fiber architecture modelling " In R. L. R. V. L. Nithiarasu (Ed.), 1st International Conference on Mathematical & Computational Biomedical Engineering (pp. 59–62). Swansea (UK).
Abstract: There is general consensus that myocardial fiber architecture should be modelled in order to fully understand the electromechanical properties of the Left Ventricle (LV). Diffusion Tensor magnetic resonance Imaging (DTI) is the reference image modality for rapid measurement of fiber orientations by means of the tensor principal eigenvectors. In this work, we present a mathematical framework for across subject comparison of the local geometry of the LV anatomy including the fiber architecture from the statistical analysis of DTI studies. We use concepts of differential geometry for defining a parametric domain suitable for statistical analysis of a low number of samples. We use Riemannian metrics to define a consistent computation of DTI principal eigenvector modes of variation. Our framework has been applied to build an atlas of the LV fiber architecture from 7 DTI normal canine hearts.
Keywords: cardiac fiber architecture; diffusion tensor magnetic resonance imaging; differential (Rie- mannian) geometry.
|
|
|
Debora Gil, & Petia Radeva. (2004). "A Regularized Curvature Flow Designed for a Selective Shape Restoration " . IEEE Transactions on Image Processing, 13, 1444–1458.
Abstract: Among all filtering techniques, those based exclu- sively on image level sets (geometric flows) have proven to be the less sensitive to the nature of noise and the most contrast preserving. A common feature to existent curvature flows is that they penalize high curvature, regardless of the curve regularity. This constitutes a major drawback since curvature extreme values are standard descriptors of the contour geometry. We argue that an operator designed with shape recovery purposes should include a term penalizing irregularity in the curvature rather than its magnitude. To this purpose, we present a novel geometric flow that includes a function that measures the degree of local irregularity present in the curve. A main advantage is that it achieves non-trivial steady states representing a smooth model of level curves in a noisy image. Performance of our approach is compared to classical filtering techniques in terms of quality in the restored image/shape and asymptotic behavior. We empirically prove that our approach is the technique that achieves the best compromise between image quality and evolution stabilization.
Keywords: Geometric flows, nonlinear filtering, shape recovery.
|
|
|
Guillermo Torres, Debora Gil, Antonio Rosell, Sonia Baeza, & Carles Sanchez. (2023)." A radiomic biopsy for virtual histology of pulmonary nodules" In IEEE International Symposium on Biomedical Imaging.
|
|
|
Maurizio Mencuccini, Jordi Martinez-Vilalta, Josep Piñol, Lasse Loepfe, Mireia Burnat, Xavier Alvarez, et al. (2010). "A quantitative and statistically robust method for the determination of xylem conduit spatial distribution " . American Journal of Botany, 97(8), 1247–1259.
Abstract: Premise of the study: Because of their limited length, xylem conduits need to connect to each other to maintain water transport from roots to leaves. Conduit spatial distribution in a cross section plays an important role in aiding this connectivity. While indices of conduit spatial distribution already exist, they are not well defined statistically. * Methods: We used point pattern analysis to derive new spatial indices. One hundred and five cross-sectional images from different species were transformed into binary images. The resulting point patterns, based on the locations of the conduit centers-of-area, were analyzed to determine whether they departed from randomness. Conduit distribution was then modeled using a spatially explicit stochastic model. * Key results: The presence of conduit randomness, uniformity, or aggregation depended on the spatial scale of the analysis. The large majority of the images showed patterns significantly different from randomness at least at one spatial scale. A strong phylogenetic signal was detected in the spatial variables. * Conclusions: Conduit spatial arrangement has been largely conserved during evolution, especially at small spatial scales. Species in which conduits were aggregated in clusters had a lower conduit density compared to those with uniform distribution. Statistically sound spatial indices must be employed as an aid in the characterization of distributional patterns across species and in models of xylem water transport. Point pattern analysis is a very useful tool in identifying spatial patterns.
Keywords: Geyer; hydraulic conductivity; point pattern analysis; Ripley; Spatstat; vessel clusters; xylem anatomy; xylem network
|
|
|
Paula Fritzsche, C.Roig, Ana Ripoll, Emilio Luque, & Aura Hernandez-Sabate. (2006). "A Performance Prediction Methodology for Data-dependent Parallel Applications " In Proceedings of the IEEE International Conference on Cluster Computing (pp. 1–8).
Abstract: The increase in the use of parallel distributed architectures in order to solve large-scale scientific problems has generated the need for performance prediction for both deterministic applications and non-deterministic applications. In particular, the performance prediction of data dependent programs is an extremely challenging problem because for a specific issue the input datasets may cause different execution times. Generally, a parallel application is characterized as a collection of tasks and their interrelations. If the application is time-critical it is not enough to work with only one value per task, and consequently knowledge of the distribution of task execution times is crucial. The development of a new prediction methodology to estimate the performance of data-dependent parallel applications is the primary target of this study. This approach makes it possible to evaluate the parallel performance of an application without the need of implementation. A real data-dependent arterial structure detection application model is used to apply the methodology proposed. The predicted times obtained using the new methodology for genuine datasets are compared with predicted times that arise from using only one execution value per task. Finally, the experimental study shows that the new methodology generates more precise predictions.
|
|
|
Enric Marti, Debora Gil, & Carme Julia. (2005). "A PBL experience in the teaching of Computer Graphics " In EUROGRAPHICS Proceedings (Vol. 5, pp. 95–103).
Abstract: Project-Based Learning (PBL) is an educational strategy to improve student’s learning capability that, in recent years, has had a progressive acceptance in undergraduate studies. This methodology is based on solving a problem or project in a student working group. In this way, PBL focuses on learning the necessary tools to correctly find a solution to given problems. Since the learning initiative is transferred to the student, the PBL method promotes students own abilities. This allows a better assessment of the true workload that carries out the student in the subject. It follows that the methodology conforms to the guidelines of the Bologna document, which quantifies the student workload in a subject by means of the European credit transfer system (ECTS). PBL is currently applied in undergraduate studies needing strong practical training such as medicine, nursing or law sciences. Although this is also the case in engineering studies, amazingly, few experiences have been reported. In this paper we propose to use PBL in the educational organization of the Computer Graphics subjects in the Computer Science degree. Our PBL project focuses in the development of a C++ graphical environment based on the OpenGL libraries for visualization and handling of different graphical objects. The starting point is a basic skeleton that already includes lighting functions, perspective projection with mouse interaction to change the point of view and three predefined objects. Students have to complete this skeleton by adding their own functions to solve the project. A total number of 10 projects have been proposed and successfully solved. The exercises range from human face rendering to articulated objects, such as robot arms or puppets. In the present paper we extensively report the statement and educational objectives for two of the projects: solar system visualization and a chess game. We report our earlier educational experience based on the standard classroom theoretical, problem and practice sessions and the reasons that motivated searching for other learning methods. We have mainly chosen PBL because it improves the student learning initiative. We have applied the PBL educational model since the beginning of the second semester. The student’s feedback increases in his interest for the subject. We present a comparative study of the teachers’ and students’ workload between PBL and the classic teaching approach, which suggests that the workload increase in PBL is not as high as it seems.
Keywords: project-based learning; computer graphics education; Open GL; rendering techniques; computer animation techniques; Graphics packages; Hierarchy and geometric transformations; Animation; Color; shading; shadowing and texture; fractals; hidden line/surface removal; Problem Based Learning
|
|
|
Enric Marti, Carme Julia, & Debora Gil. (2007). "A PBL Experience in the Teaching of Computer Graphics " In XVII Congreso Español de Informàtica Gráfica (Vol. 25, pp. 95–103).
Abstract: Project-Based Learning (PBL) is an educational strategy to improve student’s learning capability that, in recent years, has had a progressive acceptance in undergraduate studies. This methodology is based on solving a problem or project in a student working group. In this way, PBL focuses on learning the necessary tools to correctly find a solution to given problems. Since the learning initiative is transferred to the student, the PBL method promotes students own abilities. This allows a better assessment of the true workload that carries out the student in the subject. It follows that the methodology conforms to the guidelines of the Bologna document, which quantifies the student workload in a subject by means of the European credit transfer system (ECTS). PBL is currently applied in undergraduate studies needing strong practical training such as medicine, nursing or law sciences. Although this is also the case in engineering studies, amazingly, few experiences have been reported. In this paper we propose to use PBL in the educational organization of the Computer Graphics subjects in the Computer Science degree. Our PBL project focuses in the development of a C++ graphical environment based on the OpenGL libraries for visualization and handling of different graphical objects. The starting point is a basic skeleton that already includes lighting functions, perspective projection with mouse interaction to change the point of view and three predefined objects. Students have to complete this skeleton by adding their own functions to solve the project. A total number of 10 projects have been proposed and successfully solved. The exercises range from human face rendering to articulated objects, such as robot arms or puppets. In the present paper we extensively report the statement and educational objectives for two of the projects: solar system visualization and a chess game. We report our earlier educational experience based on the standard classroom theoretical, problem and practice sessions and the reasons that motivated searching for other learning methods. We have mainly chosen PBL because it improves the student learning initiative. We have applied the PBL educational model since the beginning of the second semester. The student’s feedback increases in his interest for the subject. We present a comparative study of the teachers’ and students’ workload between PBL and the classic teaching approach, which suggests that the workload increase in PBL is not as high as it seems.
|
|