Dreamtuner: Single image is enough for subject-driven generation M Hua, J Liu, F Ding, W Liu, J Wu, Q He arXiv preprint arXiv:2312.13691, 2023 | 15 | 2023 |
DEADiff: An Efficient Stylization Diffusion Model with Disentangled Representations T Qi, S Fang, Y Wu, H Xie, J Liu, L Chen, Q He, Y Zhang Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2024 | 11 | 2024 |
Sounding video generator: A unified framework for text-guided sounding video generation J Liu, W Wang, S Chen, X Zhu, J Liu IEEE Transactions on Multimedia 26, 141-153, 2023 | 8 | 2023 |
Magicvideo-v2: Multi-stage high-aesthetic video generation W Wang, J Liu, Z Lin, J Yan, S Chen, C Low, T Hoang, J Wu, JH Liew, ... arXiv preprint arXiv:2401.04468, 2024 | 7 | 2024 |
Mm21 pre-training for video understanding challenge: Video captioning with pretraining techniques S Chen, X Zhu, D Hao, W Liu, J Liu, Z Zhao, L Guo, J Liu Proceedings of the 29th ACM International Conference on Multimedia, 4853-4857, 2021 | 7 | 2021 |
Ed-t2v: An efficient training framework for diffusion-based text-to-video generation J Liu, W Wang, W Liu, Q He, J Liu 2023 International Joint Conference on Neural Networks (IJCNN), 1-8, 2023 | 6 | 2023 |
WL-MSR: Watch and Listen for Multimodal Subtitle Recognition J Liu, H Wang, W Wang, X He, J Liu ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023 | 1 | 2023 |