SID4VAM: a benchmark dataset with synthetic images for visual attention modeling

Loading...
Thumbnail Image
Identifiers
ISSN: 2380-7504
ISBN: 978-1-7281-4803-8

Publication date

Advisors

Tutors

Editors

Journal Title

Journal ISSN

Volume Title

Publisher

IEEE
Metrics
Google Scholar
lacobus
Export

Research Projects

Organizational Units

Journal Issue

Abstract

A benchmark of saliency models performance with a synthetic image dataset is provided. Model performance is evaluated through saliency metrics as well as the influence of model inspiration and consistency with human psychophysics. SID4VAM is composed of 230 synthetic images, with known salient regions. Images were generated with 15 distinct types of low-level features (e.g. orientation, brightness, color, size...) with a target-distractor popout type of synthetic patterns. We have used Free-Viewing and Visual Search task instructions and 7 feature contrasts for each feature category. Our study reveals that state-ofthe- art Deep Learning saliency models do not perform well with synthetic pattern images, instead, models with Spectral/ Fourier inspiration outperform others in saliency metrics and are more consistent with human psychophysical experimentation. This study proposes a new way to evaluate saliency models in the forthcoming literature, accounting for synthetic images with uniquely low-level feature contexts, distinct from previous eye tracking image datasets.

Description

This is the author’s version of the work. The definitive version was published in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), available online at 10.1109/ICCV.2019.00888

Bibliographic citation

D. Berga, X. R. F. Vidal, X. Otazu and X. M. Pardo, "SID4VAM: A Benchmark Dataset With Synthetic Images for Visual Attention Modeling," 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea (South), 2019, pp. 8788-8797, doi: 10.1109/ICCV.2019.00888

Relation

Has part

Has version

Is based on

Is part of

Is referenced by

Is version of

Requires

Sponsors

This work was funded by the MINECO (DPI2017- 89867-C2-1-R, TIN2015-71130-REDT), AGAUR (2017- SGR-649), CERCA Programme / Generalitat de Catalunya, in part by Xunta de Galicia under Project ED431C2017/69, in part by the Conseller´ıa de Cultura, Educación e Ordenación Universitaria (accreditation 20162019, ED431G/08) and the European Regional Development Fund, and in part by Xunta de Galicia and the European Union (European Social Fund). We also acknowledge the generous GPU support from NVIDIA

Rights

© 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
Attribution 4.0 International