Innovation Forums: The Imminent Trends of Video Compression

Moderator: Shan Liu

Jim Bankoski
Director at Google
Talk Title: Media Compression with Machine Learning


Jim Bankoski is a Distinguished Engineer at Google responsible for Chrome's video, audio, mesh and image compression efforts. He is the former CTO of On2 Technologies and a technical contributor to all of On2 / Google's video codecs from Tm2x through VP9 and AV1.

Benjamin Bross
Project Manager at the Video Coding & Analytics Department
of the Fraunhofer Institute for Telecommunications - Heinrich Hertz Institute, Berlin
Talk Title: What is cooking in video compression?


Benjamin Bross is a Project Manager at the Video Coding & Analytics Department of the Fraunhofer Institute for Telecommunications - Heinrich Hertz Institute, Berlin and a part-time lecturer at the HTW University of Applied Sciences Berlin. Since the development of the H.265 | MPEG-H High-Efficiency Video Coding (HEVC) Standard, which started in 2010, Benjamin was very actively involved in the standardization process as a technical contributor and coordinator of core experiments. In July 2012, Benjamin was appointed as a co-chair of the editing Ad Hoc Group and became the chief editor of the HEVC video coding standard. At the Heinrich Hertz Institute, he is currently responsible for the development of HEVC encoder- and decoder software as well es investigating new video coding techniques for the emerging Versatile Video Coding (VVC) standard. Besides giving talks about recent video coding technologies, Benjamin Bross is an author or co-author of several fundamental HEVC-related publications, and an author of two book chapters on HEVC and Inter-Picture Prediction Techniques in HEVC. He received the IEEE Best Paper Award at the 2013 IEEE International Conference on Consumer Electronics – Berlin in 2013, the SMPTE Journal Certificate of Merit in 2014 and an Emmy Award at the 69th Engineering Emmy Awards in 2017 as part of the Joint Collaborative Team on Video Coding for its development of HEVC.

Jill M. Boyce
Intel Fellow and Chief Media Architect at Intel
Talk Title: MPEG-I Immersive Media 3DoF+

The MPEG-I activity includes multiple projects related to Immersive Media, including 3 DoF+, which extends 360-degree 3 Degree of Freedom (Dof) video to add support for head-scale motion parallax. 3DoF+ provides a form of limited range 6 DoF video, where viewers can view content from any (X, Y, Z) position and any (yaw, pitch, roll) orientation with a limited range. The MPEG-I 3DoF+ project utilizes the HEVC codec to code texture plus depth from multiple nearly spaced cameras, plus view interpolation to form intermediate views. An update will be provided of the latest activities in the 3DoF+ project, including a planned Call for Proposals.

Jill M. Boyce is Intel Fellow and Chief Media Architect at Intel. She represents Intel at the Joint Collaborative Team on Video Coding (JCT-VC) and Joint Video Exploration Team (JVET) of ITU-T SG16 and ISO/IEC MPEG. She serves as Associate Rapporteur of ITU-T VCEG, and was an editor of the Scalability High Efficiency Video Coding extension (SHVC). She received a B.S. in Electrical Engineering from the University of Kansas in 1988 and an M.S.E. in Electrical Engineering from Princeton University in 1990. She was formerly Director of Algorithms at Vidyo, Inc. where she led video and audio coding and processing algorithm development. She was formerly VP of Research and Innovation Princeton for Technicolor, formerly Thomson. She was formerly with Lucent Technologies Bell Labs, AT&T Labs, and Hitachi America. She was Associate Editor from 2006 to 2010 of IEEE Transactions on Circuits and Systems for Video Technology. She is the inventor of more than 150 granted U.S. patents, and has published more than 40 papers in peer-reviewed conferences and journals.

Ali Tabatabai
Consultant and technical advisor for Sony US Research Center and Sony Tokyo R&D HQ
Talk Title: Video Compression Trend: Current and Future?

The emergence of digital TV is followed by ever-expanding efforts toward creating higher visual quality contents with lifelike sense of realism. These include development of higher spatial/temporal resolution (HD -> 4K -> 8K)/(30 -> 240 frame rate), higher color contrast and dynamic range (HDR) and multi-view stereo capture and display devices. In addition and in parallel efforts are being made toward direct capture and volumetric display of geometry and texture of 3D data points for a fully immersive and interactive experience (VR/AR/MR , 2.5 DoF -> 6 DoF). These industry trends have in turn been the key factor driving the development of 2D/3D video compression standards such as AVC/H.264, HEVC/H.265 and MVC as well as ongoing versatile video coding (VVC) and point cloud compression (PCC) activities. In this presentation we will briefly review these activities and point out the common characteristics and differences between traditional image based 2D/3D vs. geometry and texture based coding approaches. Finally, some thoughts about trends for future video compression technology direction will be presented.

Ali Tabatabai is currently a consultant and technical advisor to Sony US Research Center and Sony Tokyo R&D HQ. In his last position as a VP at Sony US Research Center he was responsible for research activities related to VR/AR capture and processing, camera signal processing and next generation video compression technologies. He was appointed as a Sony DE (Distinguished Engineer), a prestigious designation assigned to a very limited number of Sony worldwide employees, during 2016 – 2017. Before joining Sony he worked at Tektronix Laboratories and Bellcore/Bell Laboratories, respectively. Dr. Tabatabai is a Fellow of the IEEE and he has received many professional awards for his contributions to video coding standardization activities. He is the author and co-author of more than 60 awarded patents. He has also served as associate and guest editor of IEEE Trans. on Circuits and Systems for Video Technology (CSVT) and he is past chair of picture coding symposium (PCS).