Searched for: contributor%3A%22Yamakata%2C+Yoko+%28editor%29%22
(1 - 3 of 3)
document
Zhu, P. (author), Wang, Zhen (author), Okumura, Manabu (author), Yang, J. (author)
Textbook question answering is challenging as it aims to automatically answer various questions on textbook lessons with long text and complex diagrams, requiring reasoning across modalities. In this work, we propose MRHF, a novel framework that incorporates dense passage re-ranking and the mixture-of-experts architecture for TQA. MRHF...
conference paper 2024
document
Yu, Fuyang (author), Wang, Zhen (author), Li, Dongyuan (author), Zhu, P. (author), Liang, Xiaohui (author), Wang, Xiaochuan (author), Okumura, Manabu (author)
Cross-modal retrieval, as an important emerging foundational information retrieval task, benefits from recent advances in multimodal technologies. However, current cross-modal retrieval methods mainly focus on the interaction between textual information and 2D images, lacking research on 3D data, especially point clouds at scene level,...
conference paper 2024
document
Vieira, Romulo (author), Muchaluat-Saade, Debora (author), Cesar, Pablo (author)
The Internet of Multisensory, Multimedia and Musical Things (Io3MT) is a new concept that arises from the confluence of several areas of computer science, arts, and humanities, with the objective of grouping in a single place devices and data that explore the five human senses, besides multimedia aspects and music content. In the context of...
conference paper 2024