Spatial-Temporal Residue Network Based In-Loop Filter for Video Coding

Chuanmin Jia, Shiqi Wang, Xinfeng Zhang, Shanshe Wang, Siwei Ma

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

73 Citations (Scopus)

Abstract

Deep learning has demonstrated tremendous break through in the area of image/video processing. In this paper, a spatial-temporal residue network (STResNet) based in-loop filter is proposed to suppress visual artifacts such as blocking, ringing in video coding. Specifically, the spatial and temporal information is jointly exploited by taking both current block and co-located block in reference frame into consideration during the processing of in-loop filter. The architecture of STResNet only consists of four convolution layers which shows hospitality to memory and coding complexity. Moreover, to fully adapt the input content and improve the performance of the proposed in-loop filter, coding tree unit (CTU) level control flag is applied in the sense of rate-distortion optimization. Extensive experimental results show that our scheme provides up to 5.1% bit-rate reduction compared to the state-of-the-art video coding standard.
Original languageEnglish
Title of host publication2017 IEEE Visual Communications and Image Processing (VCIP)
PublisherIEEE
Number of pages4
ISBN (Print)9781538604625, 9781538604632
DOIs
Publication statusPublished - Dec 2017
Event2017 IEEE Visual Communications and Image Processing (VCIP 2017) - St. Petersburg, United States
Duration: 10 Dec 201713 Dec 2017

Conference

Conference2017 IEEE Visual Communications and Image Processing (VCIP 2017)
PlaceUnited States
CitySt. Petersburg
Period10/12/1713/12/17

Research Keywords

  • High Efficiency Video Coding
  • In-loop Filter
  • Spatial-Temporal Network

Fingerprint

Dive into the research topics of 'Spatial-Temporal Residue Network Based In-Loop Filter for Video Coding'. Together they form a unique fingerprint.

Cite this