Tracking more than 100 arbitrary objects at 25 FPS through deep learning

dc.contributor.affiliationUniversidade de Santiago de Compostela. Centro de Investigación en Tecnoloxías da Informacióngl
dc.contributor.affiliationUniversidade de Santiago de Compostela. Departamento de Electrónica e Computacióngl
dc.contributor.areaÁrea de Enxeñaría e Arquitectura
dc.contributor.authorVaquero Otal, Lorenzo
dc.contributor.authorBrea Sánchez, Víctor Manuel
dc.contributor.authorMucientes Molina, Manuel
dc.date.accessioned2021-08-12T12:21:15Z
dc.date.available2021-08-12T12:21:15Z
dc.date.issued2022
dc.description.abstractMost video analytics applications rely on object detectors to localize objects in frames. However, when real-time is a requirement, running the detector at all the frames is usually not possible. This is somewhat circumvented by instantiating visual object trackers between detector calls, but this does not scale with the number of objects. To tackle this problem, we present SiamMT, a new deep learning multiple visual object tracking solution that applies single-object tracking principles to multiple arbitrary objects in real-time. To achieve this, SiamMT reuses feature computations, implements a novel crop-and-resize operator, and defines a new and efficient pairwise similarity operator. SiamMT naturally scales up to several dozens of targets, reaching 25 fps with 122 simultaneous objects for VGA videos, or up to 100 simultaneous objects in HD720 video. SiamMT has been validated on five large real-time benchmarks, achieving leading performance against current state-of-the-art trackersgl
dc.description.peerreviewedSIgl
dc.description.sponsorshipThis research was partially funded by the Spanish Ministerio de Ciencia e Innovación [grant numbers PID2020-112623GB-I00, RTI2018-097088-B-C32], and the Galician Consellería de Cultura, Educación e Universidade [grant numbers ED431C 2018/29, ED431C 2017/69, accreditation 2016–2019, ED431G/08]. These grants are co-funded by the European Regional Development Fund (ERDF). Lorenzo Vaquero is supported by the Spanish Ministerio de Universidades under the FPU national plan (FPU18/03174)gl
dc.identifier.citationPattern Recognition 2022, 121: 108205. https://doi.org/10.1016/j.patcog.2021.108205gl
dc.identifier.doi10.1016/j.patcog.2021.108205
dc.identifier.issn0031-3203
dc.identifier.urihttp://hdl.handle.net/10347/26777
dc.language.isoenggl
dc.publisherElseviergl
dc.relation.publisherversionhttps://doi.org/10.1016/j.patcog.2021.108205gl
dc.rights© 2021 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)gl
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internacional
dc.rights.accessRightsopen accessgl
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/
dc.subjectMultiple visual object trackinggl
dc.subjectMotion estimationgl
dc.subjectDeep learninggl
dc.subjectSiamese networksgl
dc.subject.classificationinfo:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/PID2020-112623GB-I00/ESgl
dc.subject.classificationinfo:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/RTI2018-097088-B-C32/ES/SENSORES CMOS DE VISION, GESTION DE ENERGIA Y SEGUIMIENTO DE OBJETOS SOBRE GPUS EMPOTRADASgl
dc.titleTracking more than 100 arbitrary objects at 25 FPS through deep learninggl
dc.typejournal articlegl
dc.type.hasVersionVoRgl
dspace.entity.typePublication
relation.isAuthorOfPublication22d4aeb8-73ba-4743-a84e-9118799ab1f2
relation.isAuthorOfPublication21112b72-72a3-4a96-bda4-065e7e2bb262
relation.isAuthorOfPublication.latestForDiscovery22d4aeb8-73ba-4743-a84e-9118799ab1f2

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2022_patcog_vaquero_tracking.pdf
Size:
3.26 MB
Format:
Adobe Portable Document Format
Description: