Distribution of Wavelet Coefficients and the Mid-Treat Quantizer of JPEG2000

DOI : 10.17577/IJERTV2IS100968

Download Full-Text PDF Cite this Publication

Text Only Version

Distribution of Wavelet Coefficients and the Mid-Treat Quantizer of JPEG2000

Distribution of Wavelet Coefficients and the Mid-Treat Quantizer of JPEG2000

Arul.A 1, Mohamed Nizar.S 2, John Dhanaseely.A 3, Vinodh Kumar.B 4

1,4M.E (Applied Electronics)-IFET COLLEGE OF ENGINEERING 2,3Associate Professor-IFET COLLEGE OF ENGINEERING Villupuram.

Abstract The expectation of rate becomes ever more rapid in proportion to the growing image sizes, visually lossless coding is increasingly being considered as an alternative to numerically lossless coding, which has limited compression ratios and reduce the bit per pixels (bpp). This paper presents a method of encodes a colour image or video taking human perception in a visually of without loss manner using JPEG2000 and increase the Peak-Signal-to-Noise ratio (PSNR). In order to hide coding artifacts caused by quantization, visibility thresholds (VTs) are measured to the capacity of human eye to see fine detail and used for uniform quantization of subband signals are employed in JPEG2000 Part 1. The VTs are experimentally determined from statistically modelled uniform quantizer for rate distortion, which is based on the distribution of wavelet coefficients and the mid-treat quantizer of JPEG2000. The resulting VTs are adjusted of visibility of a signal due to presents of backgrounds through a visual masking model of image sharpening technique, and then used to determine the minimum number of coding passes to be included in the final codestream for visually lossless quality under the desired viewing conditions. Core Codestreams produced by this scheme are fully JPEG2000 Part-I compliant.

Keywords Contrast sensitivity function (CSF), human visual system(HVS), JPEG2000, visibility threshold (VT), visually lossless coding.

  1. INTRODUCTION

    JPEG-2000 part 1 has recently been approved as a new international standard for still image compression. This standard, formally known as ISO/IEC15444-1[1], provides a rich feature set including support for both lossy lossless encoding. One of the core technologies employed by the JPEG-2000 codec is the discrete wavelet transform. Thus, in order to build a high-quality JPEG-2000 encoder or decoder, one must able to construct an effective of distribution wavelet transform in dead-zone quantizer. This observation motivated us to examine numerical lossless encoding for practical issues related to implementation of DWTs in the context of JPEG- 2000.

    The remainder of this paper is structured as follows. First, we begin with a brief introduction to JPEG-2000 part I. Then, we study various issues related the implementation of wavelet transforms. For example, normalized Cohen-Daubechies- Feauveau (CDF) 9/7 DWT [3] designed for efficient implementation. Finally, this proposes a visually lossless

    JPEG2000 encoder for color images. In YCbCr color space, VTs are measured for a realistic quantization distortion model, which takes into account the statistical characteristics of wavelet coefficients as well as the dead-zone quantizer. The proposed [3] scheme is implemented without violating the JPEG2000 part I standard and automatically produces visually lossless imagery at lower bitrates than other visually lossless schemes in literature.

  2. JPEG2000 FUNDAMENTALS

    The fundamentals building blocks [4] of a JPEG2000 encoder are shown in the figure 1.

    Pre- processing

    Pre- processing

    Discrete Wavelet Transform

    Discrete Wavelet Transform

    Uniform Quantizer with deadzone

    Uniform Quantizer with deadzone

    Entropy coding

    Entropy coding

    Original Image

    Codestream image

    Rate/distortion optimisation

    Rate/distortion optimisation

    Figure1. JPEG2000 fundamentals building blocks.

    These components include preprocessing, DWT, uniform quantization with dead zone quantizer and entropy encoder. The input image of JPEG2000 may contain one or more components. Although a typical color image would have three components, (e.g.RGB or YCbCr), upto 16384(214) components can be specified for an input image to accommodate multispectral or other types of imagery [1,24]. Given a sample with a color-depth or bit-depth is the number of bits used in a single pixels of B bits, the unsigned representation would correspond to the range [0,2bit-depth -1], while the signed representation would correspond to the range [-2bit-depth-1,2bit-depth-1-1]. The bit-depth, resolution, and signed versus unsigned specification can vary for each components. If the components have different bit-depths, the most significant bits of the components should be aligned when estimating the distortion at the encoder.

    1. Preprocessing:

      The first step in preprocessing is to partition input image into rectangular and non-overlapping tiles of equal size. The

      tile size is arbitrary and can be as [5] large as the original image itself (i.e. only one tile) or as small as a single pixel. Each tile is compressed independently using its own set of specified compression parameters. Tiling is particularly useful for application [6] where the amount of available memory is limited compared to the image size.

      Next, unsigned sample values in each component are level shifted (DC offset) by subtracting a fixed value of 2B-1 from each sample to make its value symmetric around zero. Signed sample values are not level shifted. Finally, the level-shifted values can be subjected to a forward point-wise intercomponent transformation to decorrelate the color data. One restricted on apply the intercomponent must have identical bit-depths and dimensions. Two transformation choices are allowed in part-1, where both transforms operate on the first three components of an image tile with the implicit assumption that these components correspond to RGB. One transform is the irreversible color transform (ICT), which is identical to the traditional RGB to YCbCr color transformation and only be used for lossy coding. The forward ICT is identified as

      (1)

      full-frame nature of the transform decorrelates the image across a large scale and eliminates blocking artifacts at high compression ratios. Finally, the use of integer DWT filters allows for both lossless and lossy compression within a single compressed bit stream. We consider a (1-D) DWT for simplicity, and then extended the concepts to two dimensions.

      Original image Wavelet Compressed

      Coefficients image

      Figure2. Wavelet transform: effect of frequency filtering on a gray scale image [22].

      For the forward DWT the standard uses a (1-D) subband decomposition of a 1-D set of samples into low-pass samples, representing reducing the sampling rate of low-resolution version of the original set, and high-pass samples, representing a reduce the size of data version of the original set, needed for the perfect reconstruction of the original set, needed for the perfect reconstruction of the original set from the wavelet operation. In general, any user supplied wavelet filter bank may be used Part II of the standard). The DWT can be irreversible or reversible.

      This can alternatively be written as

      While the inverse ICT is given by

      (2)

      (3)

      TABLE I

      Daubechies 9/7 analysis filter coefficients [8]

      Analysis Filter coefficients

      i

      Lowpass Filter hL(i)

      Lowpass Filter hH(i)

      0

      0.6029490182363579

      1.115087052456994

      ±1

      0.266864118428723

      -0.5912717631142470

      ±2

      -0.07822326652898785

      -0.05754352622849957

      ±3

      -0.01686411844287495

      0.09127176311424948

      ±4

      0.02674875741080976

      TABLE II

      Synthesis Filter coefficients

      i

      Lowpass Filter gL(i)

      Lowpass Filter gH(i)

      0

      1.115087052456994

      0.6029490182363579

      ±1

      -0.5912717631142470

      0.2668641184428723

      ±2

      -0.05754352622849957

      -0.07822326652898785

      ±3

      0.09127176311424948

      -0.01686411844287495

      ±4

      0.02674875741080976

      Synthesis Filter coefficients

      i

      Lowpass Filter gL(i)

      Lowpass Filter gH(i)

      0

      1.115087052456994

      0.6029490182363579

      ±1

      -0.5912717631142470

      0.2668641184428723

      ±2

      -0.05754352622849957

      -0.07822326652898785

      ±3

      0.09127176311424948

      -0.01686411844287495

      ±4

      0.02674875741080976

      Daubechies 9/7 synthesis filter coefficients [8]

      The other transform is the reversible color transform (RCT), while integer-to-integer transform that approximates the ICT for color decorrelation and can be used for both lossless and lossy coding. The forward RCT is defined as

      (4)

      Where () denotes the largest integer that is smaller than or equal to .

    2. Discrete Wavelet Transform:

      The block DCT transformation in baseline JPEG has been replaced with the efficient highly intuitive framework of DWT in JPEG2000. The DWT has several transformation that make it suitable for fulfilling some of the requirements set forth by the JPEG2000 committee. For example, a multiresolution image representation is inherent to the DWT. Furthermore, the

      TABLE III

      Analysis Filter coefficients

      i

      Lowpass Filter hL(i)

      Lowpass Filter hH(i)

      0

      6/8

      1

      ±1

      2/8

      -1/2

      ±2

      -1/8

      Analysis Filter coefficients

      i

      Lowpass Filter hL(i)

      Lowpass Filter hH(i)

      0

      6/8

      1

      ±1

      2/8

      -1/2

      ±2

      -1/8

      Daubechies 5/3 analysis filter coefficients [8]

      TABLE IV

      Synthesis Filter coefficients

      i

      Lowpass Filter gL(i)

      Lowpass Filter gH(i)

      0

      1

      6/8

      ±1

      -1/2

      2/8

      ±2

      -1/8

      Synthesis Filter coefficients

      i

      Lowpass Filter gL(i)

      Lowpass Filter gH(i)

      0

      1

      6/8

      ±1

      -1/2

      2/8

      ±2

      -1/8

      Daubechies 5/3 synthesis filter coefficients [8]

      The default irreversible transform is implemented by means of the Daubechies 9-tap/7-tap filter [8].The analysis and the corresponding synthesis filter coefficients are given in Table I and II. The default reversible transformation is implemented by means of the 5-tap/3-tap filter, the coefficients of which are given in Table III and Table IV [9]. Two filtering modes are supported by the standard: the convolution-based and the lifting-based [9, 10].

    3. Uniform Quantizer with dead Zone:

      The JPEG baseline system employs a uniform quantizer and an inverse quantization process that reconstructs the quantized coefficient to the mid-point of the quantization is called Mid-treat quantizer. A different step size is allowed for each DCT coefficient to take advantage of the sensitivity of the HVS, and these step sizes are conveyed to the decoder via an 8 x 8 quantization table. One difference is in the incorporation of a central dead zone in the quantizer. It was shown in [11] that the rate-distortion optimal quantizer for a continuous signal with Laplacian probability density is a uniform quantizer with a central dead zone. The size of the optimal dead zone as a fraction of the step size increases as the variance of the Laplacian distribution decreases.

      b

      b

      In JPEG2000 part I, the dead zone can be parameterized to have a different value for each subband. Part-I adopted the dead zone with twice the step size as depicted in Figure.3, due optimal embedded structure [12]. This means that if an Mb-bit quantizer index significant bit (MSB) and proceeding to the LSB, the resulting index after decoding only Nb bits is identical to obtained by using a similar quantizer with a step size of b2M -Nb. This property allows for Signal-to Noise Ratio scalability, which in its optimal sense means that the decoder can cease decoding at any truncation point in the code-stream and still produce exactly the same image would have been encoded at the bit-rate corresponding to the truncated code-stream. This allows an original image to be compressed with JPEG2000 to the higher quality required by a given set of clients and the disseminated to each client according to the specific image quality requirement without the need to decompress and recompress the existing code- stream.

      b

      b

      b

      b

      2b

      b b b b

      -4

      -3

      -2

      -1

      0

      +1 +2 +3 +4

      Figure3. Uniform quantizer with a central dead zone with step size b.

      In JPEG baseline, a simple reconstruction strategy has been shown to improve the decoded image PSNR by approximately 0.25db [13]. Similar gains can be expected with the biased reconstruction of wavelet coefficients in JPEG2000.

      Quantization at the encoder, for each subband b,a quantizer step size b can be driven by the perceptual importance of each subband on HVS data [14] or it can be driven by other

      consideration such as rate control. The quantizer maps a wavelet coefficient yb (u, v) in subband b to a quantized index value qb (u,v) as shown in figure.3. The quantization operation is an encoder issue and can be implemented in any desired manner. However, it is most efficiently performed according to

      (5)

      The step size b is represented with a total of two bytes, an 11-bit mantissa µb and a 5-bit exponent b, according to the relationship:

      (6)

      Where Rb is the number of bits representing the nominal dynamic range of the subband b. For the irreversible (9,7) wavelet transform, two modes of signalling the value of b to the decoder are possible. In one mode, which is similar to the q-table specification used in the current JPEG, the (b, µb) value for every subband is explicitly transmitted. This is referred to as expounded quantization. The quantized values that must be chosen to take into account he HVS properties

      [15] and/or the L2-norm of each subband in order to align the bit-planes of the quantizer indices according to their true effort to the MSE. In another mode, referred to as derived quantization, a single (0, µ0) is sent for the LL subband and the (b, µb) values for each subband are derived by scaling the

      0 value by some power of two, depending on the level of decomposition associate with that subband. In particular to the given equation,

      (b, µb)= (0–NL+nb, µ0) (7)

      where NL is the total number of decomposition levels and nb is the decomposition level corresponding to subband b. It is easy to show that the equation (7) scale the step sizes for each subband according to a power of two that best approximates the L2-norm of a subband relative to the LL subband fixed value for 0.67188 in(9,7) filter.

    4. Entropy coding :

      The coding models used by the JPEG 2000 entropy [16]coder employ 18 coding contexts, in addition to a uniform context, according to the following assignment. Contexts 08 are used for significance coding during the significance propagation and cleanup passes, contexts 913 are used for sign coding, contexts 1416 are used during the refinement pass, and an additional context is used for run coding during the cleanup pass. Each code-block employs its own MQ-coder to generate an arithmetic code-stream for the entire code-block. In the default mode, the coding contexts for each code-block are initialized at the start of the coding process and are not reset at any time during the encoding process. Furthermore, the resulting codeword can only be truncated at the coding pass

      boundaries to include a different number of coding passes from each code-block in the final code-stream. All contexts are initialized [17] to uniform probabilities except for the zero context (all insignificant neighbours) and the run context, where the initial less probable symbol (LPS) probabilities are set to 0.030053 and 0.063012, respectively.

    5. Rate-Distortion optimization :

    i i

    i i

    The embedded bit stream of a single codeblock [18] has several potential truncation points. i.e., each codeblock has a separate rate distortion (RD) function. The goal of an encoder is an arrange the bitstream data of all codeblocks in an RD optimal manner, i.e. to find the truncation points which minimize the distortion for a given rate. The most common algorithm for JPEG2000 is PCRD-Optimization (post- compression-rate-distortion). A truncation[4] point of the codeblock Bi is denoted by ni, all truncation points by n. Te embedded bitstream of the codeblock Bi can be truncated to a rate Rni (for a given truncation point n ).

    The rate constraint is then

    5bpp for monochrome images. The monochrome images have a minimum PSNR of 37.1db and a maximum of 49.3 db.These results demonstrate that bitrate and PSNR are effective criteria for determining visual lossless quality [21].

    This tableVI reports bitrates obtained by each method for encoding five 512 x 512 8-bit digitized radiographs used in [22]. The method proposed here provides significantly lower bitrates. These lower bitrates result in a lower average PSNR, but do not result in visual artifacts. The Kakadu implementation of JPEG2000 provides a CSF-based visual weighting option that can enhance the visual quality of compressed imagery.

    TABLEV

    BITRATES AND PSNRS FOR THE PROPOSED VISUALLY LOSSLESS JPEG2000 ENCODER FOR 8-BIT MONOCHROME IMAGE- BARBARA(512 X 512)

    S.

    No

    Bits per pixel

    Lossless method (bpp)

    Proposed method (bpp)

    Comp ressed Ratio

    MSE

    Y PSNR

    (db)

    1

    1

    1

    0.126

    1:8.0

    12.7

    37.1

    2

    2

    2

    0.51

    1:4.0

    3.2

    43.0

    3

    3

    3

    1.111

    1:2.7

    1.1

    47.9

    4

    4

    4

    2.13

    1.2.0

    0.8

    47.9

    5

    5

    5

    3.127

    1:1.6

    0.8

    49.3

    S.

    No

    Bits per pixel

    Lossless method (bpp)

    Proposed method (bpp)

    Comp ressed Ratio

    MSE

    Y PSNR

    (db)

    1

    1

    1

    0.126

    1:8.0

    12.7

    37.1

    2

    2

    2

    0.51

    1:4.0

    3.2

    43.0

    3

    3

    3

    1.111

    1:2.7

    1.1

    47.9

    4

    4

    4

    2.13

    1.2.0

    0.8

    47.9

    5

    5

    5

    3.127

    1:1.6

    0.8

    49.3

    (8)

    i

    i

    The distortion of each codeblock [19]Ri for a truncation points ni is given by Dni . Given an additive distortion measure, the distortion D of the compressed image is derived by

    (9)

    An optimal solution (minimizing D) of truncation points for this constrained problem can be found by solving the corresponding unconstrained problem.

    (10)

  3. PROPOSED METHOD AND CODING RESULT

    The proposed encoder was implemented in Kakadu V7.2 demo [20]. In that software, we the modified and change the value of encoder. And we cannot disturb the decoder in our operation. All reported results were generated with this Kakadu_expand, Kakadu_compress, Kakadu_transcode and unmodified decoder.

    Images marked with an [20] asterisk in Table V were used during masking model calibration. For monochrome images, the numerically lossless coding method of JPEG2000 yields an average bitrate of 5 bpp visually lossless coding method achieves an average bit rate 3.127bpp and improvement in compression ratio of 1:1.6, without perceivable quality degradation.

    Also the peak-signal-to-noise-ratios for the luminance component(Y PSNRs) vary widely depending on the image. In particular, the resulting bitrates range from 1 to

    TABLEVI

    BITRATES AND PSNRS FOR THE PROPOSED VISUALLY LOSSLESS JPEG2000 ENCODER FOR 8-BIT 512 X 512 DIGITIZED RADIOGRAPHS[158208]

    S.

    No

    Bits per pixel

    Lossless method (bpp)

    Proposed method (bpp)

    Comp ressed Ratio

    MSE

    Y PSNR

    (db)

    1

    1

    1

    0.125

    1:8.0

    1.8

    45.7

    2

    2

    2

    0.465

    1:4.3

    0.9

    48.4

    3

    3

    3

    1.074

    1:2.8

    0.9

    48.4

    4

    4

    4

    2.014

    1:2.0

    0.9

    48.4

    1. (b)

      Fig. 1 (a) Original monochrome image [Barbara] and its pixels (512 x 512) value. (b) Visually lossless image encoded by the proposed method. After decompression to avoid rescaling during display. Reconstructed image of Barbara and its pixels (512 x 512) value.

  4. CONCLUSIONS

In this paper, an efficient visually lossless HVS image compression metho is proposed that is compatible in JPEG2000 part I. this method employs YCbCr in Contrast sensitivity function based weighting in the discrete wavelet transform to reduce the imperceptible information. The proposed methods offer visually lossless encoding [23] of quality at a significantly reduced the bit per pixels value and increase the peak- to-signal-ratio (PSNR) value and limits the compression ratios numerically lossless coding.

REFERENCES

  1. ISO/IEC, ISO/IEC 15444-1: Information technology-JPEG2000 image coding system- part1: core coding system, 2001.

  2. A. Cohen, I.Daubechies and J.C.Feauveau, Biorthogonal bases of compactly supported wavelets, Common. Pure Appl. Math., vol.45, no.5, PP-485-560, 1992.

  3. H.oh, A.Bilgin, and M.W.Marcellin,Visually lossless JPEG2000 using adaptive visibility thresholds and visual masking effects, in proc. Asilomor Conf. Signals, Syt. Comput., Pacific Grove, CA, Nov.2009, pp.563-567.

  4. D.Taubman and M.Marcellin,JPEG2000:Image compression fundamentals standard and practice, Boston and Kluwer Academic publishers,Nov-2001

  5. M.W.Marcellin, M.Gormish, A.Bilgin, M.P.Boliek,An overview of JPEG2000, pp-523-544 in proc. Of the Data Compression conference, snowbird (UT), Marcp000.

  6. William B. Pennabaker, Joan L. Mitchell, JPEG: Still image data compression standard, Kluwer Academic Publishers, sep-1992.

  7. D. Wu, D. M.Tan, M. Baird, J. DeCampo, C. White, and H. R. Wu, Perceptually lossless medical image coding, IEEE Trans. Med. Imag., vol. 25, no. 3, pp. 335344, Mar. 2006.

  8. M.Antonini, M.Barlaud P.Mathieu and I.Daubechies:"Image coding using the wavelet transforms", IEEE Trans. Image Proc., pp205-220, April1992.

  9. A.R.Calder bank, I.Daubechies, W.Sweldens and B.L.Yeo: Lossless Image Compression using integer to integer wavelet transforms,Proc.IEEE Int. conf. on Image Processing(ICIP97), Santra Barbara,CA, oct1997.

  10. ISO/IEC JTC1/SC29/WG1 N1646: JPEG2000 Final committee Draft v1.0, March 16, 2000.

  11. Sullivan, G (1996) Efficient scalar quantization of exponential and Laplacian variables, IEEE Trans. On Information Theory, 42(5), September, 1365-1374.

  12. M.W. Marcellin, M.A.Lepley, A.Bilgin, T.J.Flohr,T.T.Chinen and J.H.Kasner, An overview of quantization in JPEG2000,signal processing: Image compressions,17(1),pp73-84,2002.

  13. J.R.Price and M.Rabbani, Biased reconstruction of JPEGdecoding,

    signal processing letters, 6(12), pp297-299, Dec. 1999.

  14. M.Albanesi,S.Bertoluzza ,Human vision model and wavelets for high quality image compression, in proceedings of the 5th International Conference on Image Processing and its Applications, vol.410,Edinburgh,UK,pp 311-315, July1995.

  15. W.Zeng, S.Daly, S.Lei,An overview of visual optimization tools in JPEG2000, signal processing: Image communications, 17(1), pp 85- 105, January 2002.

  16. M.Marcellin, T. Flohr, A. Bilgin, D. Taubman, E. Ordentlich, M.Weinberger, G. Seroussi, C.Chrysafis, T. Fischer,B. Banister,

    M. Rabbani, and R.Joshi, (1999) Reduced Complexity Entropy Coding, ISO/IEC JTC1/SC29/WG1 Doc. N1312, June 1999.

  17. Z.Liu, L.J. Karam, and A.BWatson, JPEG 2000 encoding with perceptual distortion control, IEEE Transactions on Image Processing, 15(7), 17631778, July 2006.

  18. R.R.Coifman and M.V.Wickerhauser, Entropy based methods for best basis selection,IEEE Trans. On Information Theory, vol.38, no.2,pp. 719-746,1992.

  19. Li, J. and Lei, S. (1999) An embedded still image coder with rate-distortion optimization, IEEE Transactions on Image Processing, 8(7), July, 913924.

  20. Kakadu Software [Online]. Available: http://www.kakadusoftware.com.

  21. D. M. Chandler, N. L. Dykes, and S. S. Hemami, Visually lossless compression of digitized radiographs based on contrast sensitivity and visual masking, Proc. SPIE, vol. 5749, pp. 359372, Mar. 2005.

  22. Matlab toolbox, generate wavemenu for (2-D) Wavelet.

  23. Han Oh, Ali Bilgin, and Michael W. Marcellin, Visually lossless encoding for JPEG2000 Department of Electrical and Computer Engineering, Biomedical EngineeringUniversity of Arizona, Tucson, pp189-201, January 2013.

  24. Supplemental Images [Online]. Available: http://www.spacl.ece.arizona.edu/ohhan/visually_lossless/

Leave a Reply