Video Captioning: Bridging Video and Language with Deep Learning

Video Captioning: Bridging Video and Language with Deep Learning

【主讲人】 Dr. Tao Mei

Lead Researcher, Microsoft Research Asia

【时 间】 2016年01月21日(周四)下午15:00

【地 点】 北京大学 计算机所大楼 106报告厅

Abstract:

The recent advances in deep learning have boosted the research on video analysis. For example, convolutional neural networks have demonstrated the superiority on modeling high-level visual concepts, while recurrent neural networks have been proven to be good at modeling mid-level temporal dynamics in the video data. We present a few recent advances for understanding video content using deep learning techniques. Specifically, this talk will focus on translating video to sentence with joint embedding and translation, which achieves the best to-date performance in this nascent vision task. We will also talk about future directions for video captioning.

Biography:

Dr. Tao Mei is a Lead Researcher with Microsoft Research Asia. His current research interests include multimedia information retrieval and computer vision. He has authored or co-authored over 150 papers (4,000+ citations) in journals and conferences and holds 15 U.S. granted patents. Tao was the recipient of several paper awards from prestigious multimedia journals and conferences, including the IEEE T-CSVT Best Paper Award in 2014, the IEEE TMM Prize Paper Award in 2013, and the Best Paper Awards at ACM Multimedia in 2009 and 2007, etc.

He is an Associate Editor of IEEE Trans. on Multimedia (TMM), ACM Trans. on Multimedia Computing, Communications, and Applications (TOMM), and Multimedia Systems (MMSJ). He is the General Co-chair of ACM ICIMCS 2013, the Program Co-chair of ACM MM 2018, IEEE ICME 2015, IEEE MMSP 2015 and MMM 2013. He received the B.E. degree in automation and the Ph.D. degree in pattern recognition and intelligent systems from the University of Science and Technology of China, Hefei, China, in 2001 and 2006, respectively.

CLOSE

上一篇 下一篇