A general transductive regularizer for zero-shot learning

Huaqi Mao, Haofeng Zhang, Yang Long, Shidong Wang, Longzhi Yang

Research output: Contribution to conferencePaperpeer-review

1 Citation (Scopus)


Zero Shot Learning (ZSL) has attracted much attention due to its ability to recognize objects of unseen classes, which is realized by transferring knowledge from seen classes through semantic embeddings. Since the seen classes and unseen classes usually have different distributions, conventional inductive ZSL often suffers from the domain shift problem. Transductive ZSL is a type of method for solving such a problem. However, the regularizers of conventional transductive methods are different from each other, and cannot be applied to other methods. In this paper, we propose a General Transductive Regularizer (GTR), which assigns each unlabeled sample to a fixed attribute by defining a Kullback-Leibler Divergence (KLD) objective. To this end, GTR can be easily applied to many compatible linear and deep inductive ZSL models. Extensive experiments on both linear and deep methods are conducted on four popular datasets, and the results show that GTR can significantly improve the performance comparing to its original inductive method, and also outperform some state-of-the-art methods, especially the extension on deep model.

Original languageEnglish
Number of pages12
Publication statusPublished - 1 Jan 2020
Event30th British Machine Vision Conference, BMVC 2019 - Cardiff, United Kingdom
Duration: 9 Sept 201912 Sept 2019


Conference30th British Machine Vision Conference, BMVC 2019
Country/TerritoryUnited Kingdom


Dive into the research topics of 'A general transductive regularizer for zero-shot learning'. Together they form a unique fingerprint.

Cite this