Searched for: +
(1 - 4 of 4)
document
Wang, X. (author), Qiao, T. (author), Zhu, Jihua (author), Hanjalic, A. (author), Scharenborg, O.E. (author)
An estimated half of the world’s languages do not have a written form, making it impossible for these languages to benefit from any existing text-based technologies. In this paper, a speech-to-image generation (S2IG) framework is proposed which translates speech descriptions to photo-realistic images without using any text information, thus...
conference paper 2020
document
Zhan, X. (author), Hanjalic, A. (author), Wang, H. (author)
In this paper, we explore how to effectively suppress the diffusion of (mis)information via blocking/removing the temporal contacts between selected node pairs. Information diffusion can be modelled as, e.g., an SI (Susceptible-Infected) spreading process, on a temporal social network: an infected (information possessing) node spreads the...
conference paper 2020
document
Wang, Tan (author), Hanjalic, A. (author), Xu, Xing (author), Shen, Heng Tao (author), Yang, Yang (author), Song, Jingkuan (author)
A major challenge in matching images and text is that they have intrinsically different data distributions and feature representations. Most existing approaches are based either on embedding or classification, the first one mapping image and text instances into a common embedding space for distance measuring, and the second one regarding...
conference paper 2019
document
Wang, Bokun (author), Yang, Yang (author), Xing, Xu (author), Hanjalic, A. (author), Shen, Heng Tao (author)
Cross-modal retrieval aims to enable flexible retrieval experience across different modalities (e.g., texts vs. images). The core of crossmodal retrieval research is to learn a common subspace where the items of different modalities can be directly compared to each other. In this paper, we present a novel Adversarial Cross-Modal Retrieval ...
conference paper 2017
Searched for: +
(1 - 4 of 4)