Project Details
Video Coding for Deep Learning-Based Machine-to-Machine Communication
Applicant
Professor Dr.-Ing. André Kaup
Subject Area
Image and Language Processing, Computer Graphics and Visualisation, Human Computer Interaction, Ubiquitous and Wearable Computing
Communication Technology and Networks, High-Frequency Technology and Photonic Systems, Signal Processing and Machine Learning for Information Technology
Communication Technology and Networks, High-Frequency Technology and Photonic Systems, Signal Processing and Machine Learning for Information Technology
Term
from 2019 to 2022
Project identifier
Deutsche Forschungsgemeinschaft (DFG) - Project number 426084215
In this research project, the video coding for machines (VCM) task shall be researched with a strong focus on the recently emerged neural compression networks (NCNs). There, autoencoder networks are trained to reconstruct the input image from a latent space requiring as as little bitrate as possible to transmit. For the follow-up project, state-of-the-art NCNs optimized for the human shall serve as a starting point to base own optimizations and designs for the VCM task on. To that end, the project is divided into two phases. First, intra compression is researched by developing novel training losses representing the evaluation network at the decoder side in order to improve the coding gains for VCM. Thereby, it will be differentiated between generally optimizing for multiple evaluation networks and when the specific evaluation network is known and available before encoding. Second, the found methods from the first phase are supposed to be used to further optimize suitable NCN architectures designed for inter coding. However, suitable pristine labeled video data has to be acquired for an appropriate evaluation before. Besides, also test cases requiring video data such as tracking shall be considered to measure the coding efficiency.
DFG Programme
Research Grants