Search Results for author: Yingbin Mo

Found 2 papers, 0 papers with code

Game-MUG: Multimodal Oriented Game Situation Understanding and Commentary Generation Dataset

no code implementations30 Apr 2024 Zhihao Zhang, Feiqi Cao, Yingbin Mo, Yiran Zhang, Josiah Poon, Caren Han

In addition, we also propose a new audience conversation augmented commentary dataset by covering the game situation and audience conversation understanding, and introducing a robust joint multimodal dual learning model as a baseline.

Time Series

MCAD: Multi-teacher Cross-modal Alignment Distillation for efficient image-text retrieval

no code implementations30 Oct 2023 Youbo Lei, Feifei He, Chen Chen, Yingbin Mo, Si Jia Li, Defeng Xie, Haonan Lu

Due to the success of large-scale visual-language pretraining (VLP) models and the widespread use of image-text retrieval in industry areas, it is now critically necessary to reduce the model size and streamline their mobile-device deployment.

Retrieval Text Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.