top of page

March 19, 2026
A great article by UNC Research that discusses the background of our group's research on skill learning and video understanding. It also briefly describes my journey from being a basketball player to an AI researcher.

March 6, 2026
Gave a keynote talk "From Perception to Agency in Strategic Video Understanding" at the WACV 2026 Computer Vision for Winter Sports workshop held in Tucson, Arizona.

February 25, 2026
Gave a talk at Ai2 on our recently developed video benchmark for Strategic Video Intelligence. Stay tuned for more details and the public release!

February 9, 2026
Yulu showcasing his sports video understanding system to Chancellor Lee Roberts during the UNC Core AI event.

February 9, 2026
Gave a talk "From Pixels to Coaches: AI That Sees, Understands and Teaches" describing how our group's video understanding framework could give everyone access to world-class coaching by providing personalized feedback on skills from basketball to piano.

December 22, 2025
UNC published an article on our research to develop AI assistants that can understand human skills and help humans learn those skills more effectively.

December 4, 2025
Han presented his paper "ExAct: A Video-Language Benchmark for Expert Action Analysis" at NeurIPS'25 in San Diego.

November 21, 2025
Yan-Bo successfully defended his PhD thesis "Efficient Audio-Visual Video Modeling and Its Generative Applications." Congrats, Yan-Bo!

November 5, 2025
Our proposal to develop an AI Video Analyst has been funded by NCS LAS. Thanks for the support!

October 19, 2025
Gave a keynote talk "Beyond Recognition: Challenges in Expert-Level Skill Analysis" at the ICCV 2025 AI-driven Skilled Activity Understanding workshop in Honolulu.

October 6, 2025
Emon successfully defended his PhD thesis "Efficient Models for Long-Range Video Understanding." Congrats, Emon!

September 15, 2025
Gave a talk "Paradoxes of Language-Driven Video Understanding" at the Video AI Symposium in Paris, France.

June 26, 2025
Our paper "SiLVR: A Simple Language-based Video Reasoning Framework" won the first place award at Multi-Discipline Lecture Understanding at CVPR 2025 Multimodal Video Agent Workshop. Congrats, Ce and Yan-Bo!

June 14, 2025
Emon's paper "Video ReCap: Recursive Captioning of Hour-Long Videos" won the Distinguished Egocentric Vision Paper Award at CVPR 2025 EgoVis workshop. Congrats, Emon!

June 14, 2025
Emon's developed state-space video model, BIMBA, won the first place in the EgoSchema challenge in CVPR 2025.

June 13, 2025
Our group presented four papers at CVPR 2025 in Nasvhille, TN.

June 12, 2025
Wrapped up with the two of our organized workshops, Transformers for Vision and Multimodal Video Agents, at CVPR 2025

June 11, 2025
Presented "SiLVR: A Simple Language-based Video Reasoning Framework" at the LOVEU workshop in CVPR 2025 in Nashville, TN.

June 3, 2025
Gave a talk "From Cooking to Basketball: Skill Learning in the Age of AI" at UNC's Friday Center.

March 24, 2025
Received SONY Focused Research Award. Thank you for supporting our research, SONY!

March 12, 2025
Gave a talk on sports-based video understanding using multimodal large language models at the University of Tubingen.

February 26, 2025
4 papers accepted to CVPR. Congrats Yulu, Mohaiminul, Tanveer and Ziyang!

December 20, 2024
We will be organizing our 4th Transformers for Vision workshop and a workshop on Multimodal Video Agents at CVPR 2025.

December 14, 2024
Gave an invited talk "Complex Video Understanding using Language ... and Video" in the first Video-Language Models workshop at NeurIPS 2024 in Vancouver.

October 29, 2024
Two papers accepted to WACV 2025. Congrats to Yan-Bo and Feng!
bottom of page