KAIST EE Human Multimodal Research Team (advisor Ro, Yong Man) interview is introduced in nominent IEEE journal news letter.
(May issue of IEEE Consumer Technology Society News Featured People.)
The Human Modal research team has been announcing the following papers in AI top tier conferences and IEEE journals such as NeurIPS, AAAI, CVPR, and ICCV for the past year.
Please refer to the attachment for details of the IEEE CTSoc News on Consumer Technology (NCT).
Interview Link : https://ctsoc.ieee.org/images/CTSOC-NCT-2022-05-FP.pdf
– Journal list
- “Distinguishing Homophenes using Multi-head Visual-audio Memory for Lip Reading.” Minsu Kim, Jeong Hun Yeo, and Yong Man Ro. AAAI. 2022.
- “SyncTalkFace: Talking Face Generation with Precise Lip-syncing via Audio-Lip Memory.” Se Jin Park, Minsu Kim, Joanna Hong, Jeongsoo Choi, and Yong Man Ro. AAAI. 2022.
- “Lip to Speech Synthesis with Visual Context Attentional GAN.” Minsu Kim, Joanna Hong, and Yong Man Ro. NeurIPS (2021).
- “Multi-modality associative bridging through memory: Speech sound recollected from face video.” Minsu Kim*, Joanna Hong*, Se Jin Park, and Yong Man Ro. ICCV. 2021.
- Video prediction recalling long-term motion context via memory alignment learning, S Lee, HG Kim, DH Choi, HI Kim, YM Ro, CVPR 2021.
- “Speech Reconstruction with Reminiscent Sound Via Visual Voice Memory.” Joanna Hong, Minsu Kim, Se Jin Park, Yong Man Ro. IEEE Transactions on Audio, Speech, and Language Processing 29 (2021)
- “Cromm-vsr: Cross-modal memory augmented visual speech recognition.” Minsu Kim, Joanna Hong, Sejin Park, Yong Man Ro. IEEE Transactions on Multimedia (2021).