Home | << 1 2 3 4 5 6 7 8 9 10 >> [11–12] |
Records | |||||
---|---|---|---|---|---|
Author | Jaume Amores | ||||
Title | Multiple Instance Classification: review, taxonomy and comparative study | Type | Journal Article | ||
Year | 2013 | Publication | Artificial Intelligence | Abbreviated Journal | AI |
Volume | 201 | Issue | Pages | 81-105 | |
Keywords | Multi-instance learning; Codebook; Bag-of-Words | ||||
Abstract | Multiple Instance Learning (MIL) has become an important topic in the pattern recognition community, and many solutions to this problemhave been proposed until now. Despite this fact, there is a lack of comparative studies that shed light into the characteristics and behavior of the different methods. In this work we provide such an analysis focused on the classification task (i.e.,leaving out other learning tasks such as regression). In order to perform our study, we implemented
fourteen methods grouped into three different families. We analyze the performance of the approaches across a variety of well-known databases, and we also study their behavior in synthetic scenarios in order to highlight their characteristics. As a result of this analysis, we conclude that methods that extract global bag-level information show a clearly superior performance in general. In this sense, the analysis permits us to understand why some types of methods are more successful than others, and it permits us to establish guidelines in the design of new MIL methods. |
||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Elsevier Science Publishers Ltd. Essex, UK | Place of Publication | Editor | ||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0004-3702 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ADAS; 601.042; 600.057 | Approved | no | ||
Call Number | Admin @ si @ Amo2013 | Serial | 2273 | ||
Permanent link to this record | |||||
Author | German Ros; J. Guerrero; Angel Sappa; Daniel Ponsa; Antonio Lopez | ||||
Title | Fast and Robust l1-averaging-based Pose Estimation for Driving Scenarios | Type | Conference Article | ||
Year | 2013 | Publication | 24th British Machine Vision Conference | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | SLAM | ||||
Abstract | Robust visual pose estimation is at the core of many computer vision applications, being fundamental for Visual SLAM and Visual Odometry problems. During the last decades, many approaches have been proposed to solve these problems, being RANSAC one of the most accepted and used. However, with the arrival of new challenges, such as large driving scenarios for autonomous vehicles, along with the improvements in the data gathering frameworks, new issues must be considered. One of these issues is the capability of a technique to deal with very large amounts of data while meeting the realtime
constraint. With this purpose in mind, we present a novel technique for the problem of robust camera-pose estimation that is more suitable for dealing with large amount of data, which additionally, helps improving the results. The method is based on a combination of a very fast coarse-evaluation function and a robust ℓ1-averaging procedure. Such scheme leads to high-quality results while taking considerably less time than RANSAC. Experimental results on the challenging KITTI Vision Benchmark Suite are provided, showing the validity of the proposed approach. |
||||
Address | Bristol; UK; September 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | BMVC | ||
Notes | ADAS | Approved | no | ||
Call Number | Admin @ si @ RGS2013b; ADAS @ adas @ | Serial | 2274 | ||
Permanent link to this record | |||||
Author | Enric Marti; Ferran Poveda; Antoni Gurgui; Jaume Rocarias; Debora Gil; Aura Hernandez-Sabate | ||||
Title | Una experiencia de estructura, funcionamiento y evaluación de la asignatura de graficos por computador con metodologia de aprendizaje basado en proyectos | Type | Miscellaneous | ||
Year | 2013 | Publication | IV Congreso Internacional UNIVEST | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | IV Congreso Internacional UNIVEST | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | UNIVEST | ||
Notes | IAM; ADAS | Approved | no | ||
Call Number | Admin @ si @ MPG2013b | Serial | 2384 | ||
Permanent link to this record | |||||
Author | Ariel Amato; Angel Sappa; Alicia Fornes; Felipe Lumbreras; Josep Llados | ||||
Title | Divide and Conquer: Atomizing and Parallelizing A Task in A Mobile Crowdsourcing Platform | Type | Conference Article | ||
Year | 2013 | Publication | 2nd International ACM Workshop on Crowdsourcing for Multimedia | Abbreviated Journal | |
Volume | Issue | Pages | 21-22 | ||
Keywords | |||||
Abstract | In this paper we present some conclusions about the advantages of having an efficient task formulation when a crowdsourcing platform is used. In particular we show how the task atomization and distribution can help to obtain results in an efficient way. Our proposal is based on a recursive splitting of the original task into a set of smaller and simpler tasks. As a result both more accurate and faster solutions are obtained. Our evaluation is performed on a set of ancient documents that need to be digitized. | ||||
Address | Barcelona; October 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-1-4503-2396-3 | Medium | ||
Area | Expedition | Conference | CrowdMM | ||
Notes | ADAS; ISE; DAG; 600.054; 600.055; 600.045; 600.061; 602.006 | Approved | no | ||
Call Number | Admin @ si @ SLA2013 | Serial | 2335 | ||
Permanent link to this record | |||||
Author | Sergio Escalera; Jordi Gonzalez; Xavier Baro; Miguel Reyes; Oscar Lopes; Isabelle Guyon; V. Athitsos; Hugo Jair Escalante | ||||
Title | Multi-modal Gesture Recognition Challenge 2013: Dataset and Results | Type | Conference Article | ||
Year | 2013 | Publication | 15th ACM International Conference on Multimodal Interaction | Abbreviated Journal | |
Volume | Issue | Pages | 445-452 | ||
Keywords | |||||
Abstract | The recognition of continuous natural gestures is a complex and challenging problem due to the multi-modal nature of involved visual cues (e.g. fingers and lips movements, subtle facial expressions, body pose, etc.), as well as technical limitations such as spatial and temporal resolution and unreliable
depth cues. In order to promote the research advance on this field, we organized a challenge on multi-modal gesture recognition. We made available a large video database of 13; 858 gestures from a lexicon of 20 Italian gesture categories recorded with a KinectTM camera, providing the audio, skeletal model, user mask, RGB and depth images. The focus of the challenge was on user independent multiple gesture learning. There are no resting positions and the gestures are performed in continuous sequences lasting 1-2 minutes, containing between 8 and 20 gesture instances in each sequence. As a result, the dataset contains around 1:720:800 frames. In addition to the 20 main gesture categories, ‘distracter’ gestures are included, meaning that additional audio and gestures out of the vocabulary are included. The final evaluation of the challenge was defined in terms of the Levenshtein edit distance, where the goal was to indicate the real order of gestures within the sequence. 54 international teams participated in the challenge, and outstanding results were obtained by the first ranked participants. |
||||
Address | Sidney; Australia; December 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | 978-1-4503-2129-7 | Medium | ||
Area | Expedition | Conference | ICMI | ||
Notes | HUPBA; ISE; 600.063;MV;OR | Approved | no | ||
Call Number | Admin @ si @ EGB2013 | Serial | 2373 | ||
Permanent link to this record | |||||
Author | Onur Ferhat; Fernando Vilariño | ||||
Title | A Cheap Portable Eye-Tracker Solution for Common Setups | Type | Conference Article | ||
Year | 2013 | Publication | 17th European Conference on Eye Movements | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | Low cost; eye-tracker; software; webcam; Raspberry Pi | ||||
Abstract | We analyze the feasibility of a cheap eye-tracker where the hardware consists of a single webcam and a Raspberry Pi device. Our aim is to discover the limits of such a system and to see whether it provides an acceptable performance. We base our work on the open source Opengazer (Zielinski, 2013) and we propose several improvements to create a robust, real-time system. After assessing the accuracy of our eye-tracker in elaborated experiments involving 18 subjects under 4 different system setups, we developed a simple game to see how it performs in practice and we also installed it on a Raspberry Pi to create a portable stand-alone eye-tracker which achieves 1.62° horizontal accuracy with 3 fps refresh rate for a building cost of 70 Euros. | ||||
Address | Lund; Sweden; August 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ECEM | ||
Notes | MV;SIAI | Approved | no | ||
Call Number | Admin @ si @ FeV2013 | Serial | 2374 | ||
Permanent link to this record | |||||
Author | Andrew Nolan; Daniel Serrano; Aura Hernandez-Sabate; Daniel Ponsa; Antonio Lopez | ||||
Title | Obstacle mapping module for quadrotors on outdoor Search and Rescue operations | Type | Conference Article | ||
Year | 2013 | Publication | International Micro Air Vehicle Conference and Flight Competition | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | UAV | ||||
Abstract | Obstacle avoidance remains a challenging task for Micro Aerial Vehicles (MAV), due to their limited payload capacity to carry advanced sensors. Unlike larger vehicles, MAV can only carry light weight sensors, for instance a camera, which is our main assumption in this work. We explore passive monocular depth estimation and propose a novel method Position Aided Depth Estimation
(PADE). We analyse PADE performance and compare it against the extensively used Time To Collision (TTC). We evaluate the accuracy, robustness to noise and speed of three Optical Flow (OF) techniques, combined with both depth estimation methods. Our results show PADE is more accurate than TTC at depths between 0-12 meters and is less sensitive to noise. Our findings highlight the potential application of PADE for MAV to perform safe autonomous navigation in unknown and unstructured environments. |
||||
Address | Toulouse; France; September 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | IMAV | ||
Notes | ADAS; 600.054; 600.057;IAM | Approved | no | ||
Call Number | Admin @ si @ NSH2013 | Serial | 2371 | ||
Permanent link to this record | |||||
Author | Anastasios Doulamis; Nikolaos Doulamis; Marco Bertini; Jordi Gonzalez; Thomas B. Moeslund | ||||
Title | Analysis and Retrieval of Tracked Events and Motion in Imagery Streams | Type | Miscellaneous | ||
Year | 2013 | Publication | ACM/IEEE international workshop on Analysis and retrieval of tracked events and motion in imagery stream | Abbreviated Journal | |
Volume | Issue | Pages | |||
Keywords | |||||
Abstract | |||||
Address | Barcelona; October 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | |||
Notes | ISE | Approved | no | ||
Call Number | Admin @ si @ DDB2013 | Serial | 2372 | ||
Permanent link to this record | |||||
Author | H. Emrah Tasli; Cevahir Çigla; Theo Gevers; A. Aydin Alatan | ||||
Title | Super pixel extraction via convexity induced boundary adaptation | Type | Conference Article | ||
Year | 2013 | Publication | 14th IEEE International Conference on Multimedia and Expo | Abbreviated Journal | |
Volume | Issue | Pages | 1-6 | ||
Keywords | |||||
Abstract | This study presents an efficient super-pixel extraction algorithm with major contributions to the state-of-the-art in terms of accuracy and computational complexity. Segmentation accuracy is improved through convexity constrained geodesic distance utilization; while computational efficiency is achieved by replacing complete region processing with boundary adaptation idea. Starting from the uniformly distributed rectangular equal-sized super-pixels, region boundaries are adapted to intensity edges iteratively by assigning boundary pixels to the most similar neighboring super-pixels. At each iteration, super-pixel regions are updated and hence progressively converging to compact pixel groups. Experimental results with state-of-the-art comparisons, validate the performance of the proposed technique in terms of both accuracy and speed. | ||||
Address | San Jose; USA; July 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1945-7871 | ISBN | Medium | ||
Area | Expedition | Conference | ICME | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ TÇG2013 | Serial | 2367 | ||
Permanent link to this record | |||||
Author | H. Emrah Tasli; Jan van Gemert; Theo Gevers | ||||
Title | Spot the differences: from a photograph burst to the single best picture | Type | Conference Article | ||
Year | 2013 | Publication | 21ST ACM International Conference on Multimedia | Abbreviated Journal | |
Volume | Issue | Pages | 729-732 | ||
Keywords | |||||
Abstract | With the rise of the digital camera, people nowadays typically take several near-identical photos of the same scene to maximize the chances of a good shot. This paper proposes a user-friendly tool for exploring a personal photo gallery for selecting or even creating the best shot of a scene between its multiple alternatives. This functionality is realized through a graphical user interface where the best viewpoint can be selected from a generated panorama of the scene. Once the viewpoint is selected, the user is able to go explore possible alternatives coming from the other images. Using this tool, one can explore a photo gallery efficiently. Moreover, additional compositions from other images are also possible. With such additional compositions, one can go from a burst of photographs to the single best one. Even funny compositions of images, where you can duplicate a person in the same image, are possible with our proposed tool. | ||||
Address | Barcelona | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ACM-MM | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | TGG2013 | Serial | 2368 | ||
Permanent link to this record | |||||
Author | Sezer Karaoglu; Jan van Gemert; Theo Gevers | ||||
Title | Con-text: text detection using background connectivity for fine-grained object classification | Type | Conference Article | ||
Year | 2013 | Publication | 21ST ACM International Conference on Multimedia | Abbreviated Journal | |
Volume | Issue | Pages | 757-760 | ||
Keywords | |||||
Abstract | |||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ACM-MM | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ KGG2013 | Serial | 2369 | ||
Permanent link to this record | |||||
Author | Ivo Everts; Jan van Gemert; Theo Gevers | ||||
Title | Evaluation of Color STIPs for Human Action Recognition | Type | Conference Article | ||
Year | 2013 | Publication | IEEE Conference on Computer Vision and Pattern Recognition | Abbreviated Journal | |
Volume | Issue | Pages | 2850-2857 | ||
Keywords | |||||
Abstract | This paper is concerned with recognizing realistic human actions in videos based on spatio-temporal interest points (STIPs). Existing STIP-based action recognition approaches operate on intensity representations of the image data. Because of this, these approaches are sensitive to disturbing photometric phenomena such as highlights and shadows. Moreover, valuable information is neglected by discarding chromaticity from the photometric representation. These issues are addressed by Color STIPs. Color STIPs are multi-channel reformulations of existing intensity-based STIP detectors and descriptors, for which we consider a number of chromatic representations derived from the opponent color space. This enhanced modeling of appearance improves the quality of subsequent STIP detection and description. Color STIPs are shown to substantially outperform their intensity-based counterparts on the challenging UCF~sports, UCF11 and UCF50 action recognition benchmarks. Moreover, the results show that color STIPs are currently the single best low-level feature choice for STIP-based approaches to human action recognition. | ||||
Address | Portland; oregon; June 2013 | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 1063-6919 | ISBN | Medium | ||
Area | Expedition | Conference | CVPR | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ EGG2013 | Serial | 2364 | ||
Permanent link to this record | |||||
Author | Fares Alnajar; Theo Gevers; Roberto Valenti; Sennay Ghebreab | ||||
Title | Calibration-free Gaze Estimation using Human Gaze Patterns | Type | Conference Article | ||
Year | 2013 | Publication | 15th IEEE International Conference on Computer Vision | Abbreviated Journal | |
Volume | Issue | Pages | 137-144 | ||
Keywords | |||||
Abstract | We present a novel method to auto-calibrate gaze estimators based on gaze patterns obtained from other viewers. Our method is based on the observation that the gaze patterns of humans are indicative of where a new viewer will look at [12]. When a new viewer is looking at a stimulus, we first estimate a topology of gaze points (initial gaze points). Next, these points are transformed so that they match the gaze patterns of other humans to find the correct gaze points. In a flexible uncalibrated setup with a web camera and no chin rest, the proposed method was tested on ten subjects and ten images. The method estimates the gaze points after looking at a stimulus for a few seconds with an average accuracy of 4.3 im. Although the reported performance is lower than what could be achieved with dedicated hardware or calibrated setup, the proposed method still provides a sufficient accuracy to trace the viewer attention. This is promising considering the fact that auto-calibration is done in a flexible setup , without the use of a chin rest, and based only on a few seconds of gaze initialization data. To the best of our knowledge, this is the first work to use human gaze patterns in order to auto-calibrate gaze estimators. | ||||
Address | Sydney | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCV | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ AGV2013 | Serial | 2365 | ||
Permanent link to this record | |||||
Author | Hamdi Dibeklioglu; Albert Ali Salah; Theo Gevers | ||||
Title | Like Father, Like Son: Facial Expression Dynamics for Kinship Verification | Type | Conference Article | ||
Year | 2013 | Publication | 15th IEEE International Conference on Computer Vision | Abbreviated Journal | |
Volume | Issue | Pages | 1497-1504 | ||
Keywords | |||||
Abstract | Kinship verification from facial appearance is a difficult problem. This paper explores the possibility of employing facial expression dynamics in this problem. By using features that describe facial dynamics and spatio-temporal appearance over smile expressions, we show that it is possible to improve the state of the art in this problem, and verify that it is indeed possible to recognize kinship by resemblance of facial expressions. The proposed method is tested on different kin relationships. On the average, 72.89% verification accuracy is achieved on spontaneous smiles. | ||||
Address | Sydney | ||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | ISBN | Medium | |||
Area | Expedition | Conference | ICCV | ||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ DSG2013 | Serial | 2366 | ||
Permanent link to this record | |||||
Author | Jasper Uilings; Koen E.A. van de Sande; Theo Gevers; Arnold Smeulders | ||||
Title | Selective Search for Object Recognition | Type | Journal Article | ||
Year | 2013 | Publication | International Journal of Computer Vision | Abbreviated Journal | IJCV |
Volume | 104 | Issue | 2 | Pages | 154-171 |
Keywords | |||||
Abstract | This paper addresses the problem of generating possible object locations for use in object recognition. We introduce selective search which combines the strength of both an exhaustive search and segmentation. Like segmentation, we use the image structure to guide our sampling process. Like exhaustive search, we aim to capture all possible object locations. Instead of a single technique to generate possible object locations, we diversify our search and use a variety of complementary image partitionings to deal with as many image conditions as possible. Our selective search results in a small set of data-driven, class-independent, high quality locations, yielding 99 % recall and a Mean Average Best Overlap of 0.879 at 10,097 locations. The reduced number of locations compared to an exhaustive search enables the use of stronger machine learning techniques and stronger appearance models for object recognition. In this paper we show that our selective search enables the use of the powerful Bag-of-Words model for recognition. The selective search software is made publicly available (Software: http://disi.unitn.it/~uijlings/SelectiveSearch.html). | ||||
Address | |||||
Corporate Author | Thesis | ||||
Publisher | Place of Publication | Editor | |||
Language | Summary Language | Original Title | |||
Series Editor | Series Title | Abbreviated Series Title | |||
Series Volume | Series Issue | Edition | |||
ISSN | 0920-5691 | ISBN | Medium | ||
Area | Expedition | Conference | |||
Notes | ALTRES;ISE | Approved | no | ||
Call Number | Admin @ si @ USG2013 | Serial | 2362 | ||
Permanent link to this record |