Distinguishing homophenes using multi-head visual-audio memory for lip reading M Kim, JH Yeo, YM Ro Proceedings of the AAAI conference on artificial intelligence 36 (1), 1174-1182, 2022 | 45 | 2022 |
Multi-temporal lip-audio memory for visual speech recognition JH Yeo, M Kim, YM Ro ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023 | 8 | 2023 |
Akvsr: Audio knowledge empowered visual speech recognition by compressing audio knowledge of a pretrained model JH Yeo, M Kim, J Choi, DH Kim, YM Ro IEEE Transactions on Multimedia, 2024 | 6 | 2024 |
Lip reading for low-resource languages by learning and combining general speech knowledge and language-specific knowledge M Kim, JH Yeo, J Choi, YM Ro Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2023 | 6 | 2023 |
Visual Speech Recognition for Languages with Limited Labeled Data Using Automatic Labels from Whisper JH Yeo, M Kim, S Watanabe, YM Ro ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and …, 2024 | 3* | 2024 |
Towards practical and efficient image-to-speech captioning with vision-language pre-training and multi-modal tokens M Kim, J Choi, S Maiti, JH Yeo, S Watanabe, YM Ro ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and …, 2024 | 3 | 2024 |
Multilingual visual speech recognition with a single model by learning with discrete visual speech units M Kim, JH Yeo, J Choi, SJ Park, YM Ro arXiv preprint arXiv:2401.09802, 2024 | 1 | 2024 |
Where Visual Speech Meets Language: VSP-LLM Framework for Efficient and Context-Aware Visual Speech Processing JH Yeo, S Han, M Kim, YM Ro arXiv preprint arXiv:2402.15151, 2024 | | 2024 |