BlockGraphTransforms

From WikiBiron
Revision as of 11:16, 7 September 2012 by Sunil (talk | contribs) (Created page with '<P><B>Edge-adaptive transforms for efficient depth map coding </B></P> <p> In this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standar…')
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Edge-adaptive transforms for efficient depth map coding

In this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standard DCTs used in image and video coding applications. These transforms avoid ?ltering across edges in each image block, thus, they avoid creating large high frequency coef?cients. These transforms are then combined with the DCT in H.264/AVC and a transform mode selection algorithm is used to choose between DCT and EAT in an RD-optimized manner. These transforms are applied to coding depth maps used for view synthesis in a multi-view video coding system, and provides up to 29% bit rate reduction for a ?xed quality in the synthesized views.

Edge adaptive transform (EAT) design

We now describe how to construct our proposed EATs. This process consists of three steps, i.e., (i) edge detection is applied on the residual block to ?nd edge locations, (ii) a graph is constructed based on the edge map, then (iii) an EAT is constructed and EAT coef?cients are computed. The EAT coef?cients are then quantized using a uniform scalar quantizer and the same run-length coding used for DCT coef?cients is applied. The 2 × 2 sample block in Fig. 1 is used to illustrate the main ideas. We describe the encoder operation when applied to blocks of prediction residuals, though the same ideas can be easily applied to original pixel values.