Records |
Author |
Naila Murray; Maria Vanrell; Xavier Otazu; C. Alejandro Parraga |
Title |
Low-level SpatioChromatic Grouping for Saliency Estimation |
Type |
Journal Article |
Year |
2013 |
Publication |
IEEE Transactions on Pattern Analysis and Machine Intelligence |
Abbreviated Journal |
TPAMI |
Volume |
35 |
Issue |
11 |
Pages |
2810-2816 |
Keywords |
|
Abstract |
We propose a saliency model termed SIM (saliency by induction mechanisms), which is based on a low-level spatiochromatic model that has successfully predicted chromatic induction phenomena. In so doing, we hypothesize that the low-level visual mechanisms that enhance or suppress image detail are also responsible for making some image regions more salient. Moreover, SIM adds geometrical grouplets to enhance complex low-level features such as corners, and suppress relatively simpler features such as edges. Since our model has been fitted on psychophysical chromatic induction data, it is largely nonparametric. SIM outperforms state-of-the-art methods in predicting eye fixations on two datasets and using two metrics. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
0162-8828 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
CIC; 600.051; 600.052; 605.203 |
Approved |
no |
Call Number |
Admin @ si @ MVO2013 |
Serial |
2289 |
Permanent link to this record |
|
|
|
Author |
Jon Almazan; Albert Gordo; Alicia Fornes; Ernest Valveny |
Title |
Handwritten Word Spotting with Corrected Attributes |
Type |
Conference Article |
Year |
2013 |
Publication |
15th IEEE International Conference on Computer Vision |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
1017-1024 |
Keywords |
|
Abstract |
We propose an approach to multi-writer word spotting, where the goal is to find a query word in a dataset comprised of document images. We propose an attributes-based approach that leads to a low-dimensional, fixed-length representation of the word images that is fast to compute and, especially, fast to compare. This approach naturally leads to an unified representation of word images and strings, which seamlessly allows one to indistinctly perform query-by-example, where the query is an image, and query-by-string, where the query is a string. We also propose a calibration scheme to correct the attributes scores based on Canonical Correlation Analysis that greatly improves the results on a challenging dataset. We test our approach on two public datasets showing state-of-the-art results. |
Address |
Sydney; Australia; December 2013 |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
1550-5499 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
ICCV |
Notes |
DAG |
Approved |
no |
Call Number |
Admin @ si @ AGF2013 |
Serial |
2327 |
Permanent link to this record |
|
|
|
Author |
Albert Clapes; Miguel Reyes; Sergio Escalera |
Title |
Multi-modal User Identification and Object Recognition Surveillance System |
Type |
Journal Article |
Year |
2013 |
Publication |
Pattern Recognition Letters |
Abbreviated Journal |
PRL |
Volume |
34 |
Issue |
7 |
Pages |
799-808 |
Keywords |
Multi-modal RGB-Depth data analysis; User identification; Object recognition; Intelligent surveillance; Visual features; Statistical learning |
Abstract |
We propose an automatic surveillance system for user identification and object recognition based on multi-modal RGB-Depth data analysis. We model a RGBD environment learning a pixel-based background Gaussian distribution. Then, user and object candidate regions are detected and recognized using robust statistical approaches. The system robustly recognizes users and updates the system in an online way, identifying and detecting new actors in the scene. Moreover, segmented objects are described, matched, recognized, and updated online using view-point 3D descriptions, being robust to partial occlusions and local 3D viewpoint rotations. Finally, the system saves the historic of user–object assignments, being specially useful for surveillance scenarios. The system has been evaluated on a novel data set containing different indoor/outdoor scenarios, objects, and users, showing accurate recognition and better performance than standard state-of-the-art approaches. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Elsevier |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
HUPBA; 600.046; 605.203;MILAB |
Approved |
no |
Call Number |
Admin @ si @ CRE2013 |
Serial |
2248 |
Permanent link to this record |
|
|
|
Author |
H. Emrah Tasli; Jan van Gemert; Theo Gevers |
Title |
Spot the differences: from a photograph burst to the single best picture |
Type |
Conference Article |
Year |
2013 |
Publication |
21ST ACM International Conference on Multimedia |
Abbreviated Journal |
|
Volume |
|
Issue |
|
Pages |
729-732 |
Keywords |
|
Abstract |
With the rise of the digital camera, people nowadays typically take several near-identical photos of the same scene to maximize the chances of a good shot. This paper proposes a user-friendly tool for exploring a personal photo gallery for selecting or even creating the best shot of a scene between its multiple alternatives. This functionality is realized through a graphical user interface where the best viewpoint can be selected from a generated panorama of the scene. Once the viewpoint is selected, the user is able to go explore possible alternatives coming from the other images. Using this tool, one can explore a photo gallery efficiently. Moreover, additional compositions from other images are also possible. With such additional compositions, one can go from a burst of photographs to the single best one. Even funny compositions of images, where you can duplicate a person in the same image, are possible with our proposed tool. |
Address |
Barcelona |
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
ACM-MM |
Notes |
ALTRES;ISE |
Approved |
no |
Call Number |
TGG2013 |
Serial |
2368 |
Permanent link to this record |
|
|
|
Author |
Fadi Dornaika; Alireza Bosaghzadeh; Bogdan Raducanu |
Title |
Efficient Graph Construction for Label Propagation based Multi-observation Face Recognition |
Type |
Conference Article |
Year |
2013 |
Publication |
Human Behavior Understanding 4th International Workshop |
Abbreviated Journal |
|
Volume |
8212 |
Issue |
|
Pages |
124-135 |
Keywords |
|
Abstract |
Workshop on Human Behavior Understanding
Human-machine interaction is a hot topic nowadays in the communities of multimedia and computer vision. In this context, face recognition algorithms (used as primary cue for a person’s identity assessment) work well under controlled conditions but degrade significantly when tested in real-world environments. Recently, graph-based label propagation for multi-observation face recognition was proposed. However, the associated graphs were constructed in an ad-hoc manner (e.g., using the KNN graph) that cannot adapt optimally to the data. In this paper, we propose a novel approach for efficient and adaptive graph construction that can be used for multi-observation face recognition as well as for other recognition problems. Experimental results performed on Honda video face database, show a distinct advantage of the proposed method over the standard graph construction methods. |
Address |
Barcelona |
Corporate Author |
|
Thesis |
|
Publisher |
Springer International Publishing |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
0302-9743 |
ISBN |
978-3-319-02713-5 |
Medium |
|
Area |
|
Expedition |
|
Conference |
HBU |
Notes |
OR;MV |
Approved |
no |
Call Number |
Admin @ si @ DBR2013 |
Serial |
2315 |
Permanent link to this record |
|
|
|
Author |
Miguel Reyes; Albert Clapes; Jose Ramirez; Juan R Revilla; Sergio Escalera |
Title |
Automatic Digital Biometry Analysis based on Depth Maps |
Type |
Journal Article |
Year |
2013 |
Publication |
Computers in Industry |
Abbreviated Journal |
COMPUTIND |
Volume |
64 |
Issue |
9 |
Pages |
1316-1325 |
Keywords |
Multi-modal data fusion; Depth maps; Posture analysis; Anthropometric data; Musculo-skeletal disorders; Gesture analysis |
Abstract |
World Health Organization estimates that 80% of the world population is affected by back-related disorders during his life. Current practices to analyze musculo-skeletal disorders (MSDs) are expensive, subjective, and invasive. In this work, we propose a tool for static body posture analysis and dynamic range of movement estimation of the skeleton joints based on 3D anthropometric information from multi-modal data. Given a set of keypoints, RGB and depth data are aligned, depth surface is reconstructed, keypoints are matched, and accurate measurements about posture and spinal curvature are computed. Given a set of joints, range of movement measurements is also obtained. Moreover, gesture recognition based on joint movements is performed to look for the correctness in the development of physical exercises. The system shows high precision and reliable measurements, being useful for posture reeducation purposes to prevent MSDs, as well as tracking the posture evolution of patients in rehabilitation treatments. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
Elsevier |
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
|
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
HuPBA;MILAB |
Approved |
no |
Call Number |
Admin @ si @ RCR2013 |
Serial |
2252 |
Permanent link to this record |
|
|
|
Author |
Albert Gordo; Alicia Fornes; Ernest Valveny |
Title |
Writer identification in handwritten musical scores with bags of notes |
Type |
Journal Article |
Year |
2013 |
Publication |
Pattern Recognition |
Abbreviated Journal |
PR |
Volume |
46 |
Issue |
5 |
Pages |
1337-1345 |
Keywords |
|
Abstract |
Writer Identification is an important task for the automatic processing of documents. However, the identification of the writer in graphical documents is still challenging. In this work, we adapt the Bag of Visual Words framework to the task of writer identification in handwritten musical scores. A vanilla implementation of this method already performs comparably to the state-of-the-art. Furthermore, we analyze the effect of two improvements of the representation: a Bhattacharyya embedding, which improves the results at virtually no extra cost, and a Fisher Vector representation that very significantly improves the results at the cost of a more complex and costly representation. Experimental evaluation shows results more than 20 points above the state-of-the-art in a new, challenging dataset. |
Address |
|
Corporate Author |
|
Thesis |
|
Publisher |
|
Place of Publication |
|
Editor |
|
Language |
|
Summary Language |
|
Original Title |
|
Series Editor |
|
Series Title |
|
Abbreviated Series Title |
|
Series Volume |
|
Series Issue |
|
Edition |
|
ISSN |
0031-3203 |
ISBN |
|
Medium |
|
Area |
|
Expedition |
|
Conference |
|
Notes |
DAG |
Approved |
no |
Call Number |
Admin @ si @ GFV2013 |
Serial |
2307 |
Permanent link to this record |