Yingqing He

Hi 👋😋. I am currently a Ph.D. student at HKUST supervised by Prof. Qifeng Chen.

My main research focus is text-to-video generation, video diffusion models and related downstream applications. I am also working at other AIGC-related research💖.

Email  /  Google Scholar  /  Github  /  LinkedIn

profile photo

News

- [03/2024] Invited talk at Meituan on the topic of "Recent Advance of Text-to-Video Generation".
- [03/2024] 1 paper was accepted to CVPR 2024.
- [02/2024] 1 paper was accepted to TVCG 2024.
- [01/2024] 2 papers were accepted to ICLR 2024 (including 1 Spotlight paper).
- [12/2023] 1 paper was accepted to AAAI 2024.
- [11/2023] Released VideoCrafter 1.
- [08/2023] 1 paper was accepted to SIGGRAPH Asia 2023.
- [06/2023] Invited talk at LOVEU Workshop at CVPR 2023 on the topic of "Crafting Your Videos: From Unconditional to
                            Controllable Video Diffusion Models"
.
- [04/2023] Released VideoCrafter 0.9.
- [08/2021] 1 paper was accepted to ACM MM 2021 as an Oral paper.

Publications

Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts
Yue Ma*, Yingqing He*, Hongfa Wang, Andong Wang, Chenyang Qi, Chengfei Cai, Xiu Li, Zhifeng Li, Heung-Yeung Shum, Wei Liu, Qifeng Chen,
arXiv, 2024
Project page / arXiv / Github

Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion Latent Aligners
Yazhou Xing*, Yingqing He*, Zeyue Tian,   Xintao Wang, Qifeng Chen,
CVPR, 2024
Project page / arXiv / Github

Make a Cheap Scaling: A Self-Cascade Diffusion Model for Higher-Resolution Adaptation
Lanqing Guo*, Yingqing He*, Haoxin Chen, Menghan Xia, Xiaodong Cun, Yufei Wang, Siyu Huang,   Yong Zhang, Xintao Wang, Qifeng Chen, Ying Shan, Binhan Wen,
arXiv, 2024
Project page / arXiv / Github

MagicStick: Controllable Video Editing via Control Handle Transformations
Yue Ma, Xiaodong Cun, Yingqing He, Chenyang Qi, Xintao Wang, Ying Shan, Xiu Li,   Qifeng Chen,
arXiv, 2023
Project page / arXiv / Github

Scalecrafter: Tuning-free higher-resolution visual generation with diffusion models
Yingqing He*, Shaoshu Yang,   Haoxin Chen,   Xiaodong Cun, Menghan Xia, Yong Zhang, Xintao Wang, Ran He,   Qifeng Chen, Ying Shan
ICLR, 2024   (Spotlight)
Project page / arXiv / Github

Generating 16x higher-resolution images and 4x higher-resolution videos without any extra data and training effort.

Freenoise: Tuning-free longer video diffusion via noise rescheduling
Haonan Xiu, Menghan Xia, Yong Zhang, Yingqing He, Xintao Wang, Ying Shan, Ziwei Liu,
ICLR, 2024
Project page / arXiv / Github

Follow your pose: Pose-guided text-to-video generation using pose-free videos
Yue Ma*, Yingqing He*, Xiaodong Cun, Xintao Wang, Siran Chen,   Ying Shan, Xiu Li,   Qifeng Chen,
AAAI, 2024
Project page / arXiv / Github

Make-Your-Video: Customized Video Generation Using Textual and Structural Guidance
Jinbo Xing, Menghan Xia, Yuxin Liu,   Yuechen Zhang, Yong Zhang, Yingqing He, Hanyuan Liu, Haoxin Chen,   Xiaodong Cun, Xintao Wang, Ying Shan, Tien-Tsin Wong
TVCG, 2024
Project page / arXiv

Given text description and video structure (depth), our approach can generate temporally coherent and high-fidelity videos. Its applications include dynamic 3d-scene-to-video creation, real-life scene to video, and video rerendering.

TaleCrafter: Interactive Story Visualization with Multiple Characters
Yuan Gong,   Youxi Pang,   Xiaodong Cun, Menghan Xia, Yingqing He, Haoxin Chen,   Longyue Wang,   Yong Zhang, Xintao Wang, Ying Shan, Yujiu Yang1,
SIGGRAPH Asia, 2023
Project page / arXiv / Github

Videocrafter1: Open diffusion models for high-quality video generation
Haoxin Chen*, Menghan Xia*, Yingqing He*, Yong Zhang, Xiaodong Cun, Shaoshu Yang,   Jinbo Xing, Yaofang Liu,   Qifeng Chen, Xintao Wang, Chao Weng,   Ying Shan
arXiv, 2023
Project page / arXiv / Github

An open-sourced foundational text-to-video and image-to-video diffusion model for high-quality video generation.

Animate-A-Story: Storytelling with Retrieval-Augmented Video Generation
Yingqing He, Menghan Xia, Haoxin Chen,   Xiaodong Cun, Yuan Gong,   Jinbo Xing, Yong Zhang, Xintao Wang, Chao Weng,   Ying Shan, Qifeng Chen
arXiv, 2023  
Project page / arXiv / Github

A novel story-to-video pipeline with both structure and character controls, facilitating the generation of a vlog for a teddy bear.

Latent Video Diffusion Models for High-Fidelity Long Video Generation
Yingqing He, Tianyu Yang, Yong Zhang, Ying Shan, Qifeng Chen,
arXiv, 2022
Project page / arXiv / Github

Interpreting class conditional GANs with channel awareness
Yingqing He, Zhiyi Zhang,   Jiapeng Zhu,   Yujun Shen, Qifeng Chen,
arXiv, 2022
Project page / arXiv / Github

we propose an unsupervised method for portrait shadow removal, leveraging the facial priors from StyleGAN2. Our approach also supports facial tattoo and watermark removal.

Unsupervised portrait shadow removal via generative priors
Yingqing He*, Yazhou Xing,   Tianjia Zhang,   Qifeng Chen,
ACM MM, 2021   (Oral)
arXiv / Github

we propose an unsupervised method for portrait shadow removal, leveraging the facial priors from StyleGAN2. Our approach also supports facial tattoo and watermark removal.


Webpage templete is borrowed from this