Transcoded Video Restoration by Temporal Spatial Auxiliary Network
DOI:
https://doi.org/10.1609/aaai.v36i3.20192Keywords:
Computer Vision (CV)Abstract
In most video platforms, such as Youtube, Kwai, and TikTok, the played videos usually have undergone multiple video encodings such as hardware encoding by recording devices, software encoding by video editing apps, and single/multiple video transcoding by video application servers. Previous works in compressed video restoration typically assume the compression artifacts are caused by one-time encoding. Thus, the derived solution usually does not work very well in practice. In this paper, we propose a new method, temporal spatial auxiliary network (TSAN), for transcoded video restoration. Our method considers the unique traits between video encoding and transcoding, and we consider the initial shallow encoded videos as the intermediate labels to assist the network to conduct self-supervised attention training. In addition, we employ adjacent multi-frame information and propose the temporal deformable alignment and pyramidal spatial fusion for transcoded video restoration. The experimental results demonstrate that the performance of the proposed method is superior to that of the previous techniques. The code is available at https://github.com/icecherylXuli/TSAN.Downloads
Published
2022-06-28
How to Cite
Xu, L., He, G., Zhou, J., Lei, J., Xie, W., Li, Y., & Tai, Y.-W. (2022). Transcoded Video Restoration by Temporal Spatial Auxiliary Network. Proceedings of the AAAI Conference on Artificial Intelligence, 36(3), 2875-2883. https://doi.org/10.1609/aaai.v36i3.20192
Issue
Section
AAAI Technical Track on Computer Vision III