Yu Kong

yukong.jpg

Office: EB 3573

Email: yukong@cse.msu.edu

Dr. Yu Kong is an Assistant Professor of the Department of Computer Science and Engineering, Michigan State University. He is now directing the ACTION Lab. He serves Pattern Recognition and IEEE Transaction as an Associate Editor. He also serves CVPR and ICCV as an Area Chair. His research has been supported by NSF and DoD. He is an IEEE Senior Member.

Dr. Kong is interested in video understanding, vision-language modeling, video diffusion, open world vision, and their applications in Embodied AI, Biometrics, and Cybersecurity. More specifically, he works on action recognition, visual forecasting, video question-answering, video grounding, open-set recognition, and meta-learning.

Openings: We have openings for Fall 2025 for Ph.D. students, Visiting Scholars, and Research Assistants. Due to large volume of emails, I am not able to track and respond to every applicant. Please fill in the external contact form, and I will review applications periodically. How to apply?




Research Areas

  • Video Understanding
    • Action Recognition: A survey [IJCV-22], DEAR: Deep Evidential Action Recognition [ICCV-21], Max-Margin Heterogenous Information Machine [IJCV 2017]
    • Action Detection: Open vocabulary action detection [WACV 2025], GateHUB [CVPR 2022], OpenTAL: open-set temporal action localization [CVPR 2022]
    • Visual Forecasting: Egocentric 3D Trajectory Forecasting [ICCV-23], Adversarial Action Prediction Networks [T-PAMI 2020], DRIVE [ICCV 2021], Group Activity Prediction [ECCV 2020], Uncertainty-based Traffic Accident Anticipation [MM 2020]
  • Vision-language Modeling: Visual Reasoning [MM-23], Video Entailment [ICCV-21], Visual Grounding [MM-20]
  • Open-world Recognition: Open-set Action Recognition [ICCV 2021], Open-set Action Localization [CVPR 2022], Generalized Open-set Recognition [WACV-23], Compositional Grounding [ECCV-24], Compositional Zero-shot Learning [ECCV-24]

[Recent work:] survey on visual large language model, instructional video generation, action analysis, and vision foundation models.


News

Feb, 2025 We have one paper accepted by CVPR 2025. Congratulations to Zhanbo on his first CVPR paper at ACTION Lab!
Dec, 2024 I will be serving ICCV 2025 as an Area Chair.
Oct, 2024 We have one paper accepted by WACV 2025. Congrats to all the authors!
Sep, 2024 ACTION Lab members were attending Midwest Computer Vision Workshop at Indiana University Bloomington. [Photo]
Sep, 2024 I will be serving as an Area Chair for CVPR 2025.
Sep, 2024 We have received a grant from DoD/ARO ($300K) to develop Dynamic Scene Graphs (DSGs) over large-scale multimodal time series data for representation learning, enabling learning for a variety of computer vision tasks in complex and dynamic environments.
Aug, 2024 We have received a grant from NSF ($1.5M) to develop a single comprehensive tool Varuna to perform media forensics with Dr. Matthew Wright (Rochester Institute of Technology, lead institute) and Dr. Andrea E Hickerson (University of Mississippi).
Aug, 2024 Two new Ph.D. students, Zhongyi Cai and Wenliang Guo, are joining the lab. Welcome!
Jul, 2024 Wentao Bao successfully defended his dissertation, and will be joining Meta as a research scientist. Congratulations Dr. Bao!
Jul, 2024 We have 4 papers accepted at ECCV 2024. Congratulations to Wentao, Yifan, and Yujiang!