Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis W Yu, H Xu, Z Yuan, J Wu Proceedings of the AAAI conference on artificial intelligence 35 (12), 10790 …, 2021 | 374 | 2021 |
Ch-sims: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality W Yu, H Xu, F Meng, Y Zhu, Y Ma, J Wu, J Zou, K Yang Proceedings of the 58th annual meeting of the association for computational …, 2020 | 246 | 2020 |
Cogvlm: Visual expert for pretrained language models W Wang, Q Lv, W Yu, W Hong, J Qi, Y Wang, J Ji, Z Yang, L Zhao, X Song, ... arXiv preprint arXiv:2311.03079, 2023 | 218 | 2023 |
Cogagent: A visual language model for gui agents W Hong, W Wang, Q Lv, J Xu, W Yu, J Ji, Y Wang, Z Wang, Y Dong, ... Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2024 | 80 | 2024 |
Transformer-based feature reconstruction network for robust multimodal sentiment analysis Z Yuan, W Li, H Xu, W Yu Proceedings of the 29th ACM International Conference on Multimedia, 4400-4407, 2021 | 72 | 2021 |
Co-attentive multi-task convolutional neural network for facial expression recognition W Yu, H Xu Pattern Recognition 123, 108401, 2022 | 61 | 2022 |
M-SENA: An integrated platform for multimodal sentiment analysis H Mao, Z Yuan, H Xu, W Yu, Y Liu, K Gao arXiv preprint arXiv:2203.12441, 2022 | 38 | 2022 |
CogVLM: Visual expert for large language models W Wang, Q Lv, W Yu, W Hong, J Qi, Y Wang, J Ji, Z Yang, L Zhao, ... | 4 | 2023 |
AlignMMBench: Evaluating Chinese Multimodal Alignment in Large Vision-Language Models Y Wu, W Yu, Y Cheng, Y Wang, X Zhang, J Xu, M Ding, Y Dong arXiv preprint arXiv:2406.09295, 2024 | | 2024 |