Skip to main content

Research Repository

Advanced Search

Zero-Shot Learning Using Synthesised Unseen Visual Data with Diffusion Regularisation

Long, Yang; Liu, Li; Shen, Fumin; Shao, Ling; Li, Xuelong

Authors

Li Liu

Fumin Shen

Ling Shao

Xuelong Li



Abstract

Sufficient training examples are the fundamental requirement for most of the learning tasks. However, collecting well-labelled training examples is costly. Inspired by Zero-shot Learning (ZSL) that can make use of visual attributes or natural language semantics as an intermediate level clue to associate low-level features with high-level classes, in a novel extension of this idea, we aim to synthesise training data for novel classes using only semantic attributes. Despite the simplicity of this idea, there are several challenges. First, how to prevent the synthesised data from over-fitting to training classes? Second, how to guarantee the synthesised data is discriminative for ZSL tasks? Third, we observe that only a few dimensions of the learnt features gain high variances whereas most of the remaining dimensions are not informative. Thus, the question is how to make the concentrated information diffuse to most of the dimensions of synthesised data. To address the above issues, we propose a novel embedding algorithm named Unseen Visual Data Synthesis (UVDS) that projects semantic features to the high-dimensional visual feature space. Two main techniques are introduced in our proposed algorithm. (1) We introduce a latent embedding space which aims to reconcile the structural difference between the visual and semantic spaces, meanwhile preserve the local structure. (2) We propose a novel Diffusion Regularisation (DR) that explicitly forces the variances to diffuse over most dimensions of the synthesised data. By an orthogonal rotation (more precisely, an orthogonal transformation), DR can remove the redundant correlated attributes and further alleviate the over-fitting problem. On four benchmark datasets, we demonstrate the benefit of using synthesised unseen data for zero-shot learning. Extensive experimental results suggest that our proposed approach significantly outperforms the state-of-the-art methods.

Citation

Long, Y., Liu, L., Shen, F., Shao, L., & Li, X. (2018). Zero-Shot Learning Using Synthesised Unseen Visual Data with Diffusion Regularisation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(10), 2498-2512. https://doi.org/10.1109/tpami.2017.2762295

Journal Article Type Article
Acceptance Date Oct 4, 2017
Online Publication Date Oct 12, 2017
Publication Date 2018-10
Deposit Date Aug 31, 2019
Journal IEEE Transactions on Pattern Analysis and Machine Intelligence
Print ISSN 0162-8828
Electronic ISSN 1939-3539
Publisher Institute of Electrical and Electronics Engineers
Peer Reviewed Peer Reviewed
Volume 40
Issue 10
Pages 2498-2512
DOI https://doi.org/10.1109/tpami.2017.2762295
Public URL https://durham-repository.worktribe.com/output/1294898
Related Public URLs https://ueaeprints.uea.ac.uk/65217/