Member of Technical Staff , Video Codec Specialist
Dr. Debargha Mukherjee received his M.S./Ph.D. degrees in ECE from University of California Santa Barbara in 1999. Thereafter, through 2009 he was with Hewlett Packard Laboratories, conducting research on video/image coding and processing. Since 2010 he has been with Google Inc., where he is currently involved with open-source video codec research and development. Prior to that he was responsible for video quality control and 2D-3D conversion on YouTube. Debargha has authored/co-authored more than 100 papers on various signal processing topics, and holds more than 60 US patents, with many more pending. He has delivered many workshops and talks on Google's royalty-free line of codecs and recently AV1 since 2012. He currently serves as an Associate Editor of the IEEE Trans. on Circuits and Systems for Video Technology and has previously served as Associate Editor of the IEEE Trans. on Image Processing; He is also a member of the IEEE Image, Video, and Multidimensional Signal Processing Technical Committee (IVMSP TC).
演讲:A Technical Overview of the coding tools in AV1
2018-10-20 09:00 - 09:45
The Alliance for Open Media - a consortium of major Internet companies formed in 2016 with the mission to develop open media formats for the web - closed their first video codec AV1 in June of 2018. AV1 is the best standardized video codec available today that is also royalty-free. This talk will provide a high level overview of the coding tools in AV1, with special attention to the tools and features that are industry-first in standardized codecs. Results on standard tests sets will be provided.
Faouzi Kossentini
The world is today witnessing a revolutionary transformation in the areas of media delivery, processing and consumption. Not only the internet traffic is already dominated by the exchange of visual information but also the visual cloud has become of the center of such traffic. Towards a fast-growing visual cloud, with underlying mostly-Xeon-populated homogeneous data centers, Intel Corporation is now leading the way with an Open Source SW-centric strategy. Intel is already open sourcing its SVT-HEVC encoder in Q3’2018, and it also plans to open source the SVT-AV1 encoder in Q2’2019. Such will increase adoption and decrease the cost of HEVC and AV1 by our visual cloud customers, helping them to accelerate the growth of their visual cloud applications. We will discuss briefly Intel’s current visual cloud SW strategy as well as provide detailed descriptions of the being-open-sourced SVT-HEVC encoder and the soon-to-be-open-sourced SVT-AV1 encoder. We will also present results that will illustrate the performance-quality tradeoffs of each of the SVT-HEVC and SVT-AV1 encoders. Finally, we will invite our visual cloud customers to participate in the development and growth of the new SVT-HEVC and SVT-AV1 Open Source communities.
Point cloud data has been emerging and popular for various applications including virtual reality (VR), augmented reality (AR), plus many other computer vision problems. Comparing to conventional image/video data, several fundamental challenges need to be addressed, e.g., its large data rate, irregular sampling structure, varying geometric as well as associated attributes. In this talk, we first review recent progress in point cloud compression and an ongoing MPEG standardization attempt. A framework to conduct fast point cloud resampling is then presented based graph signal processing, which could be utilized to scale down the complexity of many point cloud processing tasks. Last, deep neural networks (DNNs) are generalized for native supporting on point clouds as a new type of signal. For one example, a folding-based autoencoder (AE) architecture FoldingNet is introduced with great potentials for point cloud classification tasks. In short, fundamental techniques are highlighted how point clouds could be successful with more efficient compression, processing and understanding.