@inproceedings{liu2018fast, author = {Liu, X. and Nie, X. and Zeng, Wenjun and Cui, C. and Zhu, L. and Yin, Y.}, title = {Fast Discrete Cross-modal Hashing With Regressing From Semantic Labels}, booktitle = {ACM Multimedia (oral)}, year = {2018}, month = {October}, abstract = {Hashing has recently received great attention in cross-modal retrieval. Cross-modal retrieval aims at retrieving information across heterogeneous modalities (e.g., texts vs. images). Cross-modal hashing compresses heterogeneous high-dimensional data into compact binary codes with similarity preserving, which provides efficiency and facility in both retrieval and storage. In this study, we propose a novel fast discrete cross-modal hashing (FDCH) method with regressing from semantic labels to take advantage of supervised labels to improve retrieval performance. In contrast to existing methods that learn the projection from hash codes to semantic labels, the proposed FDCH regresses the semantic labels of training examples to the corresponding hash codes with a drift. It not only accelerates the hash learning process, but also helps generate stable hash codes. Furthermore, the drift can adjust the regression and enhance the discriminative capability of hash codes. Especially in the case of training efficiency, FDCH is much faster than existing methods. Comparisons with several state-of-the-art techniques with three benchmark datasets have demonstrated the superiority of FDCH under various cross-modal retrieval scenarios.}, publisher = {ACM}, url = {http://approjects.co.za/?big=en-us/research/publication/fast-discrete-cross-modal-hashing-with-regressing-from-semantic-labels/}, }