Hello! I am Yingqing He. Nice to meet you!
π¨βπ»β I am currently a PhD student at HKUST. My research focuses on text-to-video generation and multimodal generation.
π« How to reach me: yhebm@connect.ust.hk
π£ Our lab is hiring engineering-oriented research assistants (RA). If you would like to apply, feel free to reach out with your CV!
π§ Recent projects:
- π VideoTuna: An all-in-one codebase for text-to-video applications. Github
- π Awesome-LLMs-meet-Multimodal-Generation: A curated list of papers on LLMs-based multimodal generation (image, video, 3D, and audio). Github
- π Awesome Diffusion Models in High-Resolution Synthesis Github
- π [CVPR 2024] Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion Latent Aligners. Github
- π [ECCV 2024] Make a Cheap Scaling: A Self-Cascade Diffusion Model for Higher-Resolution Adaptation. Github