About
Experiences
Education
News
- Oct 2025 We release VideoNSA, a hardware-aware native sparse attention mechanism for video understanding.
- Sep 2025 Invited talk at Lambda AI titled From Seeing to Thinking.
- Sep 2025 One paper accepted by ICCV 2025 KnowledgeMR Workshop.
- Aug 2025 Our paper MovieChat+: Question-aware Sparse Memory for Long Video Question Answering is accepted by IEEE TPAMI.
- Jul 2025 Our paper Video-MMLU: A Massive Multi-Discipline Lecture Understanding Benchmark is accepted by ICCV 2025 Findings.
Selected Publications and Manuscripts
* Equal contribution.
Also see Google Scholar.

VideoNSA: Native Sparse Attention Scales Video Understanding
Preprint, 2025
VideoNSA delivers hardware-aware native sparse attention primitives for efficient video understanding systems.

Video-MMLU: A Massive Multi-Discipline Lecture Understanding Benchmark
ICCVW, 2025
Video-MMLU is a massive benchmark designed to evaluate the capabilities of LMMs in understanding Multi-Discipline Lectures.

AuroraLong: Bringing RNNs Back to Efficient Open-Ended Video Understanding
ICCV, 2025
Video-MMLU uses a linear RNN language model that handles input sequence of arbitrary length with constant-size hidden states to solve long video understanding tasks.
AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark
ICLR, 2025
AuroraCap is a multimodal LLM designed for image and video detailed captioning. We also release VDC, the first benchmark for detailed video captioning.

MovieChat: From Dense Token to Sparse Memory for Long Video Understanding
CVPR, 2024
MovieChat achieves state-of-the-art performace in extra long video (more than 10K frames) understanding by introducing memory mechanism.
Teaching Assistant
     Teaching Assistant (TA), with Prof. Gaoang Wang
Selected Honors & Awards
- Lambda AI Cloud Credits Grant Sponsorship, 2025
- National Scholarship, 2025 (Zhejiang University)
- National Scholarship, 2024 (Zhejiang University)
- National Scholarship, 2021 (Dalian University of Technology)
Top