Robust visual tracking with channel attention and focal loss
| dc.contributor.author | Li, Dongdong | |
| dc.contributor.author | Wen, Gongjian | |
| dc.contributor.author | Kuai, Yangliu | |
| dc.contributor.author | Zhu, Lingxiao | |
| dc.contributor.author | Porikli, Fatih | |
| dc.date.accessioned | 2024-05-10T00:14:35Z | |
| dc.date.issued | 2020 | |
| dc.date.updated | 2023-01-15T07:16:30Z | |
| dc.description.abstract | Recently, the tracking community leads a fashion of end-to-end feature representation learning for visual tracking. Previous works treat all feature channels and training samples equally during training. This ignores channel interdependencies and foreground–background data imbalance, thus limiting the tracking performance. To tackle these problems, we introduce channel attention and focal loss into the network design to enhance feature representation learning. Specifically, a Squeeze-and-Excitation (SE) block is coupled to each convolutional layer to generate channel attention. Channel attention reflects the channelwise importance of each feature channel and is used for feature weighting in online tracking. To alleviate the foreground–background data imbalance, we propose a focal logistic loss by adding a modulating factor to the logistic loss, with two tunable focusing parameters. The focal logistic loss down-weights the loss assigned to easy examples in the background area. Both the SE block and focal logistic loss are computationally lightweight and impose only a slight increase in model complexity. Extensive experiments are performed on three challenging tracking datasets including OTB100, UAV123 and TC128. Experimental results demonstrate that the enhanced tracker achieves significant performance improvement while running at a real-time frame-rate of 66 fps. | en_AU |
| dc.description.sponsorship | This work is supported by the National Natural Science Foundation of China (NSFC) (project no 61902420). | en_AU |
| dc.format.mimetype | application/pdf | en_AU |
| dc.identifier.issn | 0925-2312 | en_AU |
| dc.identifier.uri | http://hdl.handle.net/1885/317411 | |
| dc.language.iso | en_AU | en_AU |
| dc.publisher | Elsevier | en_AU |
| dc.rights | © 2019 Published by Elsevier B.V. | en_AU |
| dc.source | Neurocomputing | en_AU |
| dc.subject | Visual tracking | en_AU |
| dc.subject | Channel attention | en_AU |
| dc.subject | Focal logistic loss | en_AU |
| dc.title | Robust visual tracking with channel attention and focal loss | en_AU |
| dc.type | Journal article | en_AU |
| local.bibliographicCitation.lastpage | 307 | en_AU |
| local.bibliographicCitation.startpage | 295 | en_AU |
| local.contributor.affiliation | Li, Dongdong, National University of Defense Technology | en_AU |
| local.contributor.affiliation | Wen, Gongjian, National University of Defense Technology | en_AU |
| local.contributor.affiliation | Kuai, Yangliu, National University of Defense Technology | en_AU |
| local.contributor.affiliation | Zhu, Lingxiao, National University of Defense Technology | en_AU |
| local.contributor.affiliation | Porikli, Fatih, College of Engineering, Computing and Cybernetics, ANU | en_AU |
| local.contributor.authoruid | Porikli, Fatih, u5405232 | en_AU |
| local.description.embargo | 2099-12-31 | |
| local.description.notes | Imported from ARIES | en_AU |
| local.identifier.absfor | 400900 - Electronics, sensors and digital hardware | en_AU |
| local.identifier.ariespublication | a383154xPUB11365 | en_AU |
| local.identifier.citationvolume | 401 | en_AU |
| local.identifier.doi | 10.1016/j.neucom.2019.10.041 | en_AU |
| local.identifier.scopusID | 2-s2.0-85083118628 | |
| local.identifier.thomsonID | WOS:000544725700027 | |
| local.publisher.url | https://www.elsevier.com/en-au | en_AU |
| local.type.status | Published Version | en_AU |
Downloads
Original bundle
1 - 1 of 1
Loading...
- Name:
- 1-s2.0-S0925231219314201-main.pdf
- Size:
- 3.94 MB
- Format:
- Adobe Portable Document Format
- Description: