Dr Yang Long yang.long@durham.ac.uk
Associate Professor
Zero-Shot Learning Using Synthesised Unseen Visual Data with Diffusion Regularisation
Long, Yang; Liu, Li; Shen, Fumin; Shao, Ling; Li, Xuelong
Authors
Li Liu
Fumin Shen
Ling Shao
Xuelong Li
Abstract
Sufficient training examples are the fundamental requirement for most of the learning tasks. However, collecting well-labelled training examples is costly. Inspired by Zero-shot Learning (ZSL) that can make use of visual attributes or natural language semantics as an intermediate level clue to associate low-level features with high-level classes, in a novel extension of this idea, we aim to synthesise training data for novel classes using only semantic attributes. Despite the simplicity of this idea, there are several challenges. First, how to prevent the synthesised data from over-fitting to training classes? Second, how to guarantee the synthesised data is discriminative for ZSL tasks? Third, we observe that only a few dimensions of the learnt features gain high variances whereas most of the remaining dimensions are not informative. Thus, the question is how to make the concentrated information diffuse to most of the dimensions of synthesised data. To address the above issues, we propose a novel embedding algorithm named Unseen Visual Data Synthesis (UVDS) that projects semantic features to the high-dimensional visual feature space. Two main techniques are introduced in our proposed algorithm. (1) We introduce a latent embedding space which aims to reconcile the structural difference between the visual and semantic spaces, meanwhile preserve the local structure. (2) We propose a novel Diffusion Regularisation (DR) that explicitly forces the variances to diffuse over most dimensions of the synthesised data. By an orthogonal rotation (more precisely, an orthogonal transformation), DR can remove the redundant correlated attributes and further alleviate the over-fitting problem. On four benchmark datasets, we demonstrate the benefit of using synthesised unseen data for zero-shot learning. Extensive experimental results suggest that our proposed approach significantly outperforms the state-of-the-art methods.
Citation
Long, Y., Liu, L., Shen, F., Shao, L., & Li, X. (2018). Zero-Shot Learning Using Synthesised Unseen Visual Data with Diffusion Regularisation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(10), 2498-2512. https://doi.org/10.1109/tpami.2017.2762295
Journal Article Type | Article |
---|---|
Acceptance Date | Oct 4, 2017 |
Online Publication Date | Oct 12, 2017 |
Publication Date | 2018-10 |
Deposit Date | Aug 31, 2019 |
Journal | IEEE Transactions on Pattern Analysis and Machine Intelligence |
Print ISSN | 0162-8828 |
Electronic ISSN | 1939-3539 |
Publisher | Institute of Electrical and Electronics Engineers |
Peer Reviewed | Peer Reviewed |
Volume | 40 |
Issue | 10 |
Pages | 2498-2512 |
DOI | https://doi.org/10.1109/tpami.2017.2762295 |
Public URL | https://durham-repository.worktribe.com/output/1294898 |
Related Public URLs | https://ueaeprints.uea.ac.uk/65217/ |
You might also like
EfficientTDNN: Efficient Architecture Search for Speaker Recognition
(2022)
Journal Article
Kernelized distance learning for zero-shot recognition
(2021)
Journal Article
A plug-in attribute correction module for generalized zero-shot learning
(2020)
Journal Article
Semantic combined network for zero-shot scene parsing
(2019)
Journal Article
Downloadable Citations
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search