9base Paperbase

download 9base Paperbase

of 12

Transcript of 9base Paperbase

  • 7/28/2019 9base Paperbase

    1/12

    IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012 2641

    Hierarchical Oriented Predictions for ResolutionScalable Lossless and Near-Lossless Compression

    of CT and MRI Biomedical ImagesJonathan Taquet and Claude Labit

    AbstractWe propose a new hierarchical approach to resolution

    scalable lossless and near-lossless (NLS) compression. It combines

    the adaptability of DPCM schemes with new hierarchical orientedpredictors to provide resolution scalability with better compressionperformances than the usual hierarchical interpolation predictoror the wavelet transform. Because the proposed hierarchical ori-ented prediction (HOP) is not really efficient on smooth images, wealso introduce new predictors, which are dynamically optimized

    using a least-square criterion. Lossless compression results, whichare obtained on a large-scale medical image database, are more

    than 4% better on CTs and 9% better on MRIs than resolutionscalable JPEG-2000 (J2K) and close to nonscalable CALIC. The

    HOP algorithm is also well suited for NLS compression, providingan interesting ratedistortion tradeoff compared with JPEG-LS

    and equivalent or a better PSNR than J2K for a high bit rate onnoisy (native) medical images.

    Index TermsHierarchical prediction, image coding, losslesscompression, medical imaging, near-lossless (NLS) compression.

    I. INTRODUCTION

    E VEN if numerous studies have been developed in this fieldof interest, compression of biomedical images [1] remainsan important issue. Since the emergence of digital acquisition inmedical imaging, the data production is continuously growing.In recent years, it has been subject to a quasi-exponential in-

    crease, in particular, because of an extensive use of MRI imagesand, even more, computed tomography(CT). These are both vo-lumic modalities that can be viewedas a sequence of 2-Dimages

    (slices). The successive improvements of acquisition equipmenttends to amplify the resolution of those images, which intensi-fies the mass of data to archive [2]. All this makes them reallymuch more cumbersome than other imaging modalities. This iswhy we focused on CT and MRI.

    Medical images may require to be saved for periods of over30 years. They are stored in picture archiving and communi-

    cation systems for which efficient compression algorithms areof great interest. They could also benefit from functionalities

    Manuscript received May 25, 2011; revised September 21, 2011 and De-cember 16, 2011; accepted January 04, 2012. Date of publication January 26,2012; date of current version April 18, 2012. This work was supported in partby the Brittany Council under Grant 4591. The associate editor coordinatingthe review of this manuscript and approving it for publication was Prof. HitoshiKiya.

    The authors are with INRIA, Centre Inria Rennes Bretagne Atlan-tique, IRISA, 35042 Rennes, France (e-mail: [email protected];[email protected]).

    Color versions of one or more of the figures in this paper are available onlineat http://ieeexplore.ieee.org.

    Digital Object Identifier 10.1109/TIP.2012.2186147

    allowing an intelligent navigation through a network that may

    reduce the data transmissions.Legally, the diagnostic information must be kept in the same

    state as during the initial diagnosis stage to allow their reconsid-

    eration in case of judicial proceedings. Therefore, if some lossesappear as compression consequences, the radiologists would

    have to study the degraded images when doing their diagnosis.Although recent research efforts are focused on lossy compres-sion of biomedical images, such as in [3] or [4] (where lossy

    compression is obtained using vector quantization), and even ifavant-garde radiologists advocate using lossy compression [5]

    and recommend compression ratio that ranges with JPEG and

    JPEG-2000 (J2K) standards for various imaging modalities [6](emphasizing that these are only guidelines), it is difficult to

    convince all medical specialists who really do not want to lose

    any important information. Therefore, for ethical reasons, med-ical images are most often stored without any loss, even if they

    always contain unnecessary noisy information that could be re-

    moved by using a less drastic lossy compression that can ensurea control on the losses, such as near-lossless (NLS) algorithms,

    to preserve a visually lossless quality.Focusing on 2-D algorithms, the best lossless compression re-

    sults are usually obtained with efficient DPCM schemes. They

    follow a row-scan-ordered prediction and use adaptive methodsexploiting causal information. JPEG-LS (JLS) standard [7] and

    CALIC [8] are often used as references [9]. Such coders lacka progressive model, which is important for distant access ofbiomedical images. In lossless context, resolution or rate scala-

    bility coding is allowed by HIP approaches such as hierarchicalinterpolation (HINT) [10] and derivative works such as inter-

    leaved HINT (IHINT) [11], as well as by an integer wavelettransform (IWT) used for example in the J2K standard [12].

    HIPs use subsampling to extract a low-resolution image that

    is employed for an interpolated prediction of the missing sam-

    ples with finite support filters. To provide resolution or rate scal-ability, the residual data can be compressed by a subband bit-

    plane coder such as EBCOT [13] or SPIHT [14], or by anyspecific -ary dictionary specialized entropy coder that will be

    more optimal for resolution only scalability.

    Lifting schemes commonly used for IWT act similar to an in-terpolated prediction but also concentrate energy in low bands.

    Such integer wavelet bases provide antialiased multiresolution

    representations and have better PSNRthan HIPs when used withrate scalable coders, but it is necessary to perform an additional

    optimization scheme to exploit them efficiently for NLS com-

    pression (see Section VI).In [15], the authors introduced a new hierarchical oriented

    prediction (HOP) that combines DPCM with HIP. It provides

    1057-7149/$31.00 2012 IEEE

  • 7/28/2019 9base Paperbase

    2/12

    2642 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    TABLE IDATABASES COMPOSITION

    only resolution scalability but exploits already coded pixels ofthe same subband to improve decorrelation, compared withIWT

    or HIP. In this paper, the algorithm is more detailed and in-cludes some minors changes with no significant impact on theresults. These modifications concern the previously used

    threshold that is removed from the definition of the static pre-dictor, and the simplification of the context selection for biascancelation. New experiments are added to validate three main

    contributions: 1) the HOP approach; 2) the sequential context-based error correction; and 3) the entropy coding technique.To improve the lack of efficiency of the predictors on smooth

    images, two new extensions, which exploit dynamically con-structed predictors using a least square optimization, are alsoproposed. The hierarchical approach of IHINT, which performs

    better than HINT, was retained for lossless and NLS compar-isons with the proposed methods. Then, the used implementa-

    tion of IHINT will also be detailed throughout this paper.In Section II, the data sets used for the experiments are

    introduced. The hierarchical decomposition of IHINT andthe proposed one for HOP are described in Section III. TheHOP prediction approaches are then presented in Section IV,

    which details the static predictors (see Section IV-A) and thedynamically least-square-estimated ones (see Section IV-B).

    The proposed sequential prediction bias cancelation is ex-plained in Section V. The NLS extensions of IHINT and HOPare exposed in Section VI; then, the proposed coding scheme

    that is used with both algorithms is described in Section VII.Finally, lossless and NLS results on CT and MRI images arecompared with usual 2-D reference algorithms in Section VIII.

    II. DATA SETS USED FOR EXPERIMENTS

    A very large amount of data has been used for this paper.

    Experiments were carried out on more than 6000 CT and 3000

    MRI slices. The database used contains 12-bit images: native

    CT and MRI slices of the National Library of Medicines Vis-

    ible Human Project,1 resulting from the acquisitions of two full

    bodies (VHP-Male/Female) and a head (VHP-Harvard); Med-

    ical Database for the Evaluation of Image and Signal Processing

    Algorithms (MeDEISA)2; and 8-bit images from Mallinckrodt

    Institute of Radiology, Image Proc. Lab., available at CIPR.3

    The structure of the whole data set is further detailed in Table I.

    From the VHP-Harvard data set, because it has given

    atypical results compared with the other MRI (see results in

    Section VIII), a special 3-D MRI volume has been removed

    1http://www.nlm.nih.gov/research/visible/visible_human.html

    2http://www.medeisa.net

    3http://www.cipr.rpi.edu/resource/sequences/sequence01.html

    and is considered separately (Harvard-3D). It contains textured

    noise, which has correlated information, and well-pronounced

    Gibbs reconstruction artefacts. The CIPR data set contains

    postprocessed images for which values are remapped to 8 bits;

    therefore, most of the noise is removed. However, it is often

    used as a reference in the literature of volumic medical images

    compression, such as in [16] and [17]. MeDEISA is composed

    of postfiltered CT slices, native and postprocessed MRI slices,and 3-D MRI slices. Postprocessed MRI slices have parts of

    the background noise (regions of no interest) set to 0. The

    VHP data sets are native acquisitions coming directly from the

    acquisition material; therefore, the images are the most noisy

    of the whole database.

    For validation experiments, three subsets show the different

    kinds of results that may be obtained. Smooth images results are

    showed by those obtained on the CT slices of MeDEISA, and

    noisy native CT/MRI slices by those obtained on VHP-Male

    CT/MRI data sets. Three slice examples, shown in Fig. 1, are

    also extracted for rate/distortion comparisons.

    III. HIERARCHICAL DECOMPOSITIONS

    A. IHINT Decomposition

    To hierarchically decompose an image, a prediction level

    of IHINT [11] can be summarized in the two prediction steps

    showed in Fig. 2. Let be the set of horizontally even indexed

    pixel values, and let be the set of horizontally odd indexed

    pixel values; the first step (HStep) consists of predicting the

    pixels of using an interpolative finite impulse response filter

    on . then contains the residual values of the prediction. The

    second step (VStep) is the mathematical transposition of HStep

    applied independently on to obtain two sets and , and

    on to obtain and .The set then contains a subsampled lower resolu-

    tion image that can be recursively decomposed. The pixel

    values/residual of each set can be reorganized to obtain a

    dyadic representation.

    For comparisons to be held in the following, the IHINT pre-

    diction of an odd indexed value is the average of its two (even

    indexed) neighboring values. This high-pass filter was retained

    because of its similarity with predictors used by HOP [see (3)

    in Section IV-A] and has been well known to be an efficient

    predictor.

    B. Proposed Decomposition

    Similar to IHINT, a prediction level of the HOP scheme [15]is performed in two steps (see Fig. 3). The first one (HStep) con-

    sists of predicting horizontally odd indexed pixels values with

    the aid of already known pixels: the proposed approach uses

    not only the even indexed ones but it can also take advantage of

    any previously predicted ones (which are now causal values).

    Thus, we can obtain a horizontally subsampled image and the

    residual of the odd predicted pixels. The second step (VStep) is

    the mathematical transposition of HStep and acts on the lower

    resolution image.

    With the HOPs prediction approach (see next section), the

    subset is sufficiently decorrelated and does not require to be

    decomposed once again.

    Similar to IHINT, HOP can be computed in the same

    memory space as the image, but it requires that the current row

  • 7/28/2019 9base Paperbase

    3/12

    TAQUET AND LABIT: HOPs FOR LOSSLESS AND NLS COMPRESSION OF CT AND MRI BIOMEDICAL IMAGES 2643

    Fig. 1. 12-bit-slice examples: contrast LUT has been adjusted for 8-bit displayand printing. (a) Not native CT of MeDEISA [CT_data_1(91)]. (b) Native CTof VHP-Male [normalCT(248)]. (c) Native MRI of VHP-Male [pd-1(21)].

    Fig. 2. One prediction level of the IHINT algorithm.

    Fig. 3. One prediction level of the HOP algorithm.

    (or column) and the previous ones that are necessary for the

    prediction are buffered (it represents here a total of three rows).

    IV. PROPOSED HOPS

    A. Static Predictor

    HOP is mostly designed for noisy images containing struc-

    tured objects with sharp edges (contrasted data), which is the

    case for most of native medical images.

    Inspired by the efficient gradient-adjusted predictor of

    CALIC, an orientation estimation is done using the pattern

    presented in Fig. 4. First, the absolute value of the local dif-ferences is computed for each orientation belonging to set

    , i.e.,

    Card(1)

    where is the image, , and is defined by the

    set of linked pixels in Fig. 4. Then, using a noise threshold

    (defined later in Section IV-C), the diagonal gra-

    dient , and the horizontal/vertical

    gradient , the most favorable orienta-

    tion is selected for the

    prediction

    if

    else

    withif

    else.(2)

    It allows a choice between five predictors: non-oriented ,

    vertical , horizontal , diagonal, and diag-

    onal. The prediction value is then given as follows:

    if

    if

    if

    if

    (3)

  • 7/28/2019 9base Paperbase

    4/12

    2644 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    Fig. 4. (left) Contextual prediction pattern and (right) linked pixels used forgradient estimation.

    Fig.5. Sets ofcausal pixelsusedfor(left)HOP-LSE, and for (leftand right)

    HOP-LSE .

    to perform an oriented prediction along edges, and

    if (4)

    for estimating the value by a more robust filter in homogeneous

    areas that mainly contain noise.

    The local activity for the predictors orientation

    if

    else(5)

    will be used in the next sections (Sections IV-B and V).

    B. Least-Square-Estimated Dynamic Predictors

    The HOPs prediction is not really effective on smooth

    images mainly because the small prediction support size (3) is

    not adequate for the decorrelation of diffuse information. Be-

    cause higher order predictors would result in better estimation

    on such images (using a wider prediction support size), we

    also introduce two new predictive approaches: HOP-LSE and

    HOP-LSE . To build , they both exploit the extended sets of

    causal pixels compared with HOP (see Fig. 5). For an efficient

    TABLE IIIMPACT OF THE THRESHOLD ON THE COMPRESSION (bpp)

    use of these extended sets, the predictors are dynamically built

    using least square estimations, giving a better adaptation to the

    specific characteristics of each image.

    For HOP-LSE , the following optimization is always done,

    whereas for HOP-LSE, if , the nonoriented static pre-

    dictor (4) of the previous section is used. It reduces the com-

    plexity of the algorithm by computing the least square optimiza-

    tion only in the nonhomogeneous areas, which represent about

    30% of the pixels on the whole data sets and then allows the di-

    vision of the computation time, for a nonoptimized implemen-

    tation, by around 3 compared with HOP-LSE .

    To compute different predictors depending on the local ac-

    tivity, (5) is logarithmically quantized to

    (6)

    where gives the index of the highest bit of the integer

    part of . A prediction context that combines both

    the local activity and the local orientation is then selected for

    the prediction.

    Let be the value of the current pixel to be predicted,

    where is the context selected for the prediction and is the

    number of times that has been previously selected as a pre-diction context. Let be the column vector containing the

    values of the causal pixels belonging to the set (see Fig. 5).

    The prediction will be

    (7)

    where is the vector containing the least-square-optimized

    linear prediction coefficients. It is classically computed by

    (8)

    where and.

    Because this least square optimization needs to be non-

    singular and because it requires a certain number of learning

    and to be sufficiently effective, as long as ( was fixed

    to 64 for experiments), and whenever is not far from sin-

    gularity, the static predictors associated to (2) and depicted in

    the previous section are used.

    To avoid keeping incrementally large matrix and

    vector for each context, the implementation is done by

    iteratively updating the positive definite symmetric matrix

    AC (noncentered autocorrelation of the

    prediction pixels) as follows:

    AC AC (9)

  • 7/28/2019 9base Paperbase

    5/12

    TAQUET AND LABIT: HOPs FOR LOSSLESS AND NLS COMPRESSION OF CT AND MRI BIOMEDICAL IMAGES 2645

    TABLE IIICOMPRESSION RESULTS FOR DIFFERENT LENGTH OF THE SEQUENTIAL CONTEXT ERROR CORRECTION

    Best results are in bold, whereas worst are underlined.

    and the vector CC (noncentered cross

    correlation between prediction pixels and predicted values) asfollows:

    CC CC (10)

    with in lossless mode and being its

    reconstructed value if any quantization is used (see Section VI).

    C. Threshold

    In order not to spend too much time optimizing the

    threshold, it is set to a noise estimation (using the Donohos

    thresholding formula [18]) computed on the highest frequencies

    of an orthonormal Haar transform of the image to compress.For HOP and HOP-LSE , even if it is not always optimal in

    terms of coding efficiency, it allows a slight improvement of

    the compression, up to about 1% for HOP on noisy data sets

    and for HOP-LSE on smooth data sets. However, on noisy

    CTs, HOP-LSE compression is not improved because the

    noise in CTs has strong location dependence and then cannot

    be captured by the used least square optimization process.

    Table II shows results obtained with and without the threshold.

    Results of HOP-LSE are not presented because HOP-LSE has

    no sense without the threshold: It would then be equivalent

    to HOP-LSE , the predictors optimization would be done for

    almost every pixel, and then the statically defined nonoriented

    predictor characterizing HOP-LSE would not be used.

    For practical uses, depending on the material and acquisition

    parameters, the threshold could be optimized offline.

    V. PREDICTION BIAS CANCELATION

    A. Overview

    To avoid systematic errors, also called biases, generally oc-

    curring with context-based static predictors, a common tech-

    nique is to correct prediction in a context with the av-

    erage of the previous errors that occurred in the same

    context

    (11)

    can bedefined from any causal computation. For example,

    JLS is quantizing the gradient on the causal neighborhood ofthe predicted value, and CALIC is also using the prediction

    value itself to define a binary texture pattern. In [19], the au-

    thors extended this approach by using a cumulated prediction

    error correction corresponding to the average of weighted errors

    from different contexts selections as follows:

    (12)

    and showed some compression improvements.

    B. SCEC

    We propose to extend (12) by using a sequential context-

    based error correction (SCEC). This is done by sequentially ap-plying the following correction scheme for to :

    (13)

    with . The final prediction value is , with

    as the rounding to the nearest integer operator. The prediction

    error is then used to update each .

    When the context selection does not depend on , (13) is

    equivalent to (12), but when it depends on it, the prediction

    is successively refined and allows the selection of more and

    more accurate contexts. For experiments, the coefficients are

    simply fixed to be .

    The context selection used with HOP is now described.Since oriented approaches are not always effective on textured

    areas, the , depending of , are defined by retaining the

    texture capture process of CALIC. A local texture information

    nn ww is binarized

    to define a 5-bit number , with if

    , or else, . is combined with the local

    activity quantization (6) and with the employed orientation

    numbers as follows: 0 for noise, 1 for horizontal, 2 for vertical,

    3 for diagonal, and 4 for diagonal.

    C. SCEC Validation

    Some SCEC results, obtained when varies, are reported

    in Table III. The analysis is done for HOP, HOP-LSE, and

    HOP-LSE in lossless mode PAE and NLS mode

  • 7/28/2019 9base Paperbase

    6/12

    2646 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    (PAE , see next section). The predictors performances will

    be discussed later in Section VIII.

    With static predictors (HOP), SCEC always allows the

    asymptotic improvement of the compression, with further ame-

    liorations for smooth images. As an example, on the smooth

    MeDEISA CT slices, the lossless compression rate is reduced

    (compared with no correction: ) by 3.1% using only

    one correction context , whereas it is reduced by3.55% using a sequence of five corrections . Then,

    on noisy native VHP-Male MRI images, the improvement is

    of 0.64% when and 0.91% when , and on noisy

    native VHP-Male CT images, where the kind of noise is much

    more location dependent, the rate is only reduced from 0.51%

    to 0.76% . In NLS mode, the compression

    amelioration is better again, with a rate reduced by up to

    6.25% for MeDEISA CT images, 1.35% for CT

    images, and 1.87% for MRI images of VHP-Male. In terms of

    distortions in NLS mode, only the first context correction seems

    to improve the PSNR; then, both the PSNR and the rate slightly

    decrease, keeping the ratedistortion tradeoff approximatelysimilar.

    With dynamic predictors (HOP-LSE and HOP-LSE ), the

    same observations can be made on smooth images and native

    CT images. However, on native MRI images, when ,

    the bias cancelation seems to fail. This is due to the smaller

    size of the images (256 256) and the highly contrasted na-

    ture of MRI. For the smallest resolutions, the images are aliased

    due to the subsampling: Pixels are less correlated, and the least

    square estimation is more difficult. The prediction is therefore

    less efficient, and the bias cancelation is not effective enough in

    small resolutions due to a lack of data for correctly estimating

    correction statistics. This could be improved by disabling the

    contextual correction for a given context while the number ofprevious occurrences in the same context is smaller than a cer-

    tain threshold. However, since allows the bypass of this

    drawback (by averaging the errors of different contexts) and im-

    proves results, no threshold was used.

    Results show that SCEC allows the reduction of compressed

    image size in both compression modes. In NLS, it has a small

    impact on the ratedistortion tradeoff with static predictors, but

    with dynamically optimized ones, would be better being null

    or greater than 1. For the experiments that follow, is always

    fixed to 3.

    VI. NLS EXTENSIONS

    A. Brief Review

    When introducing slight compression losses in medical

    imaging, a frequently used error control is the peak of absolute

    error (PAE). Its value is the maximum absolute difference

    between the pixels of the original and the ones of the recon-

    structed images. A PAE-constrained lossy compression is often

    referred to as NLS compression.

    With predictive coding, NLS compression is performed by

    predicting pixels from the NLS causal reconstructed values. The

    residual error is then generally quantized by a uniform quanti-

    zation as follows:

    (14)

    which ensures a PAE less or equal to when reconstructing

    (15)

    Note that it is verified that

    (16)

    This is useful for NLS hierarchical predictions (see

    Sections VI-B and VI-C).

    JLS has adopted a slightly different approach, which pro-

    motes run length encoding (RLE) whenever the RLE value sat-

    isfies the PAE and uses the quantized residual if otherwise.

    For high PAE, it generates line artefacts but allows an effi

    cientcompression without employing arithmetic coding.

    When using transform coding for NLS compression, one of

    the best known way is to use a lossy compression, followed by a

    residual coding of the quantized error image [20], [21]. This

    kind of approach requires an optimization of the coding cost

    of joint lossy residual representation. Some research efforts

    were also carried out on wavelet-based scalable -oriented

    compression [22].

    In terms of PSNR, the transform approach is more effective

    than predictive coding for high PAE because the lossy compres-

    sion tends to minimize the MSE and thus preserves the most

    important energy information. It also results in good compres-

    sion properties of the residual probability distribution function(PDF). However, for small PAE, when really good quality must

    be preserved, as in medical imaging, the predictive approach is

    currently recognized as performing an equivalent or a better job.

    B. IHINT NLS Extension

    Let be the residual of when predicted from .

    For NLS extension of IHINT (see Section III-A and Fig. 2),

    the approach starts, with the pyramidal dyadic reorganization

    of the pixels, by approximating the lowest resolution image

    . Then, a pyramidal descent is done, starting

    from this lowest resolution until the full-resolution NLS approx-

    imation of the image is obtained: is used to predict andto get their approximated residuals and . The

    prediction of is reverted to build and allows the com-

    putation of . By reverting

    the prediction steps, and are also built. Then, , ,

    , and data are reorganized to obtain an upper resolution

    NLS image that can be used as the band of the next resolu-

    tion level.

    When the full-resolution NLS image is obtained, the hierar-

    chical prediction (which will be the exactly the same as during

    the pyramidal descent) is done in inverse order (from highest

    to lowest resolution, as during a lossless compression). The

    residual values and the lowest resolution image are quan-

    tized again to obtain exactly the same quantized residual values,

    as during the pyramidal descent.

  • 7/28/2019 9base Paperbase

    7/12

    TAQUET AND LABIT: HOPs FOR LOSSLESS AND NLS COMPRESSION OF CT AND MRI BIOMEDICAL IMAGES 2647

    C. HOP NLS Extension

    The NLS extension of HOP follows the merged guidelines

    of HIP and DPCM: it starts by approximating the lowest

    resolution image used to predict higher resolution ones, and al-

    ways performs its successive predictions from previously ap-

    proximated values. For each increasing resolution, it starts by

    a VStep (see Fig. 3) to reconstruct a completely NLS band,which is then used for the NLS prediction of the band.

    Similar to IHINT, when the full-resolution NLS image is ob-

    tained, the decomposition is done once again, in the same order

    as during lossless compression, and the residual values and the

    lowest resolution image are quantized again.

    For both IHINT and HOP, the NLS decomposition requires

    the NLS image estimation (pyramidal descent), followed by

    the decomposition (pyramidal ascent). The time complexity is

    then around two times the one of the lossless decomposition,

    which only requires the pyramidal ascent decomposition, but the

    memory consumption stays the same. Another implementation

    using temporary storage of the residual data obtained during the

    pyramidal descent would allow it to perform with the same timecomplexity as lossless but with a loss of memory of the size of

    the image.

    VII. CODING SCHEME

    A. Residual Remapping

    The residual remapping is often used by predictive coders to

    reduce the alphabet size (by a factor of 2 compared with the

    full-residual-range values) to make the entropy coding easier.

    Knowing and (with

    being the image), the prediction value can be forced to lie

    in . If and, respectively, , it is

    forced to be resp. . The residual prediction

    is then bounded by

    and . Now, it can be remapped using a

    bijective function as follows:

    (17)

    To obtain a PDF that tends to be decreasing, positive and nega-

    tive residual values are interlaced using the following:

    if andif and

    else if

    else if

    (18)

    with , , and

    (see also Fig. 6).

    The remapping is employed with both IHINT and HOP.

    B. Resolution Scalable Coding

    For the following results, the image decomposition is fully

    done for both IHINT and HOP. The value of the lowest resolu-

    tion image, which is then a single pixel, is first transmitted.

    To allow a resolution-progressive transmission, the decom-

    position is compressed starting from the lowest resolution.

    Fig. 6. Residual remapping.

    For IHINT, each remapped quantized residual band ,

    , and then ; for HOP, each remapped quantized

    residual band and then are sequentially coded

    row by row and column by column. A coding context CC

    is determined from quantized local error energy, modeled by

    causal neighbors, and already coded lower band hierarchical

    neighbors of the current symbol (remapped quantized residual

    value). CC allows an entropy coding of , using a PDF that isiteratively computed each time a symbol has previously been

    coded in the same context.

    From the minimal resolution of 32 32 (width 32 and

    height 32), these statistics and the associated coders are reset

    to their initial states before the coding of the next resolution.

    This allows the independent decoding of a small image and of

    each successive expansion.

    This multiresolution scalable coding method can be em-

    ployed, e.g., in quick view for a fast navigation approach: In a

    huge set of images, it can be used to quickly decode/transmit

    subresolution images; a selection can then be easily made, and

    only the information required to reconstruct upper resolutionsneed to be decoded/transmitted.

    To be effective with the resolution progressive constraints,

    the entropy coder needs to estimate quickly and efficiently the

    PDF of the data that must be compressed. In the literature, there

    are fast adaptation entropy coders exploiting usual PDF mod-

    eling framework (generalized Gaussian, Laplacian, geometric,

    etc.), and we need only adjust a small number of parameters

    for the PDF estimation. However, obviously, they are not as ef-

    ficient as if a finer PDF estimation was used. The problem of

    a finer PDF estimation is that it requires more parameters and,

    therefore, more input data to correctly adjust to it.

    The approach proposed for HOP is a combination of the two

    previously mentioned ones. It uses a small set of parameters toquickly estimate the data distribution and to have a rapid coding

    efficiency, and a wider set of parameters for a slower but better

    fit to the true distribution, which allows a more powerful com-

    pression. This is done with the aid of two distinct coders used

    in a two-stage coding shown in Fig. 7. A reduced-parameter en-

    tropy coder (stage-0 coder S0C) allows us to quickly approxi-

    mate the data distribution and to compress the first coefficients,

    whereas another coder (stage-1 coder S1C) is learning finer dis-

    tribution statistics. When the S1C coder is estimated to have suf-

    ficiently learned the statistics, it is used instead of S0C for the

    next coefficients compression. S1C is continuing to update the

    data distribution. When a coefficient value is unknown by S1C,

    an escape symbol is output, and S0C is employed to compress

    this value. The algorithm details are shown in the Appendix.

  • 7/28/2019 9base Paperbase

    8/12

    2648 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    TABLE IVRESOLUTION SCALABILITY RESULTS

    Uncompressed images have size with 12 bits per pixel. The bpap notation means bits per added pixels and correspondsto the bit rate used for coding the pixels required to get this resolution from the previous one.Best results are in bold, whereas the worst results are underlined.

    Fig. 7. Two-stage coder.

    C. Resolution Scalability Validation

    Table IV shows some coding results obtained for various res-

    olution decoding of the three images in Fig. 1. The efficiency of

    HOP, HOP-LSE, and HOP-LSE is compared with IHINT and

    J2K. Comparing all results to J2K, which is a multiscale refer-ence coder, it shows that the proposed coder is also well suited

    for scalability.

    For lossless mode, even if it is difficult to compare J2K with

    IHINT because of the energy compaction of wavelets, the pre-

    sented entropy coder seems to have equivalent or better perfor-

    mance in terms of scalable compression. For the small resolu-

    tions, in both lossless and NLS modes, oriented predictive ap-

    proaches (HOP, HOP-LSE, and HOP-LSE ) are more efficient

    than the nonoriented IHINT. HOP also gives better results than

    HOP-LSE and HOP-LSE , which can be explained by a lack of

    data for least square estimations of aliased (subsampled only)

    subresolution images. This remark could be taken into account

    to improve the compression by disabling the optimization on

    lowest subresolutions.

    VIII. EXPERIMENTAL RESULTS

    A. Lossless Compression

    HOP, HOP-LSE, and HOP-LSE are compared with the de-

    scribed implementation of the IHINT algorithm, with the stan-

    dards JLS4 and J2K5 and with the reference software of SPIHT6

    (quality scalability) and CALIC.7 Table V reports the lossless

    results averaged by the data set for the whole database.

    On CTs, CALIC always gives the best compression perfor-

    mances, except for smooth images data set (MeDEISA), where

    least square dynamically optimized predictors perform better.

    On MRI, except on the smooth data set (Harvard-3D), it often

    performs equivalent compression to HOP.

    For scalable coders only, J2K is most often the worst or not far

    from the worst coding algorithm, leaving out MeDEISA CT im-

    ages for which HOP is not efficient. However, except on smooth

    data sets, HOP is always better than SPIHT, J2K, and IHINT.

    IHINT obtains results similar to J2K on CT images but is com-

    petitive with CALIC and HOP on MRI.

    The proposed least square optimization of the predictors al-

    lows us to bypass the inefficiency of HOP on smooth images.

    Both HOP-LSE and HOP-LSE give the best compression re-

    sults on the smooth data sets. They also outperform HOP most

    of the time.

    4http://www.hpl.hp.com/loco/

    5http://www.kakadusoftware.com/

    6http://www.cipr.rpi.edu/research/SPIHT/spiht3.html (linux version)

    7ftp://ftp.csd.uwo.ca/pub/from_wu/v.arith (sun version)

  • 7/28/2019 9base Paperbase

    9/12

    TAQUET AND LABIT: HOPs FOR LOSSLESS AND NLS COMPRESSION OF CT AND MRI BIOMEDICAL IMAGES 2649

    TABLE VLOSSLESS RATES AVERAGES

    Best results are in bold, whereas worst are underlined. Cells of the best results by category (scalable/nonscalable)are grayed.

    TABLE VI

    NLS RATES AND PSNR AVERAGES FOR A PAE OF 4

    Best rate averages and best PSNR averages for NLS coders are in bold. For each data set, the PSNR obtained by the best ratealgorithm (BRA) is compared with the average of PSNR obtained by J2K-9/7: The bit rate obtained by the BRA on each slice isused to constrain the J2K-9/7 compression. The best PSNR average between the BRA and J2K are underlined.

    On MeDEISA CT images, HOP is 3% worse than J2K or

    IHINT, but with HOP-LSE and HOP-LSE , 3.3% and 6.8%

    of the space required by J2K and IHINT is saved. Moreover,

    HOP, HOP-LSE, and HOP-LSE improve the compression by

    more than 3% on the others CT images. On 8-bit CT images,

    HOP-LSE also reduces space by more than 10% compared

    with J2K.

    On common MRI, the least square optimization of the pre-

    dictors improves HOP compression by up to around 2%, which

    leads to improvements of 5% to more than 7% compared withJ2K, and around 3% compared with IHINT. On 8-bit MRI im-

    ages, HOP, HOP-LSE, and HOP-LSE improve compression

    by 8.4%, 11.1%, and 12.5% compared with J2K; and 2.9%,

    5.7%, and 7.2% compared with IHINT. On smooth MRIs (Har-

    vard-3D) HOP-LSE saves 17.8% of the space used by J2K and

    13.7% of the IHINT requirements.

    Averaging the results for all the database, HOP, HOP-LSE,

    and HOP-LSE are 2.24%, 3.87% and 4.28% better than J2K

    on CT images; and 4.22%, 7.45% and 9.13% better on MRI

    images, respectively.

    B. NLS Compression

    For NLS comparisons, only IHINT and JLS are kept because

    the other previously used software do not provide PAE-con-

    strained compression. However, PSNR performances of the

    PAE constrained algorithms are compared with the lossy ver-

    sion of J2K (rate constrained), using real valued biorthogonal

    wavelets (9/7 kernel). Averaged rates and PSNR obtained by

    JLS, IHINT, HOP, HOP-LSE, and HOP-LSE for a slight

    PAE of 4 are given in Table VI. The J2K PSNR results are

    obtained with the Kakadu v6.0 software, using a quantization

    step (Qstep) set to , with being the bit depth of the

    original image (8 or 12).

    The images presented in Fig. 1 are used in Fig. 8 to showratedistortion curves that may be obtained with JLS, IHINT,

    and the best HOP HOP-LSE . For better readability, HOP and

    HOP-LSE results are not plotted. On the CT of the MeDEISA

    data set, which shows results on smooth images and is one of

    the worst losslessly coded by HOP (compared with the other

    algorithms), HOP would be about 0.21 bpp less efficient than

    IHINT when PAE (the PSNR is estimated to be 1.92 dB

    worst) and 0.06 bpp less efficient when PAE (PSNR is

    estimated to be 0.96 dB worst). The HOP-LSE results would

    be between IHINT and HOP-LSE . On the two others images,

    which give examples for the compression of noisy native CT

    images and MRI images, HOP and HOP-LSE results would be

    similar to HOP-LSE .

  • 7/28/2019 9base Paperbase

    10/12

    2650 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    Fig. 8. NLS rate/distortions results.

    A PAE of 4 allows a compression improved by a factor

    greater than 2 compared with lossless mode and has no vi-

    sual impact. HOP, HOP-LSE, and HOP-LSE results show

    that, in terms of PSNR, they allows us to obtain equivalent or

    better performances than J2K on noisy native images. PSNR

    is slightly worse on 8-bit data sets and on smooth images.

    However, the NLS approaches enjoy the PAE quality control,

    which is not the case of J2K.

    Similar in lossless mode, the three HOP approaches

    have relatively equivalent results on most noisy images.

    The HOP oriented approach is, again, better than the

    nonoriented IHINT one, except on the smoothest data sets

    where HOP-LSE and HOP-LSE make the difference with

    the least square optimization, which improves HOP results

    by more than 10% and up to 20% on Harvard-3D set.

    Compared with the JLS standard, HOPs improves both

  • 7/28/2019 9base Paperbase

    11/12

    TAQUET AND LABIT: HOPs FOR LOSSLESS AND NLS COMPRESSION OF CT AND MRI BIOMEDICAL IMAGES 2651

    TABLE VIIEXECUTION TIME (SECS CODER/SECS DECODER) FOR IMAGES IN FIG. 1

    rates and PSNR and, moreover, provides the resolution

    scalability.

    C. Indicative Execution Time

    Just for information, we give some execution timesin Table VII obtained with an Intel(R) Core(TM)2 DuoE6750 CPU at 2.66 GHz. IHINT and HOP applications aremonothreaded and are not optimized. The main optimizationsthat could be performed are related to the entropy coder and to

    the least square optimization.

    IX. CONCLUSION AND PERSPECTIVES

    A new HOP approach, and two variants using least squareoptimization (HOP-LSE and HOP-LSE ) have been presentedin the context of resolution scalable lossless and NLS biomed-ical image compression. A new sequential context-based biascancelation method was proposed and analyzed to improve theprediction efficiency. The last original contribution was an en-tropy coding technique based on a two-stage coder designed toimprove the compression in the resolution scalable context.

    This paper has shown that, even if providing resolution scal-ability, some compression improvements could be obtained on

    noisy native medical images both in lossless and NLS modescompared with the reference algorithms. The least square opti-mization has allowed us to boost the prediction on smooth im-ages, where HOP was not really efficient.

    For small distortions, HOP obtains equivalent or better PSNRthan J2K on noisy native images and moreover enjoys the PAEquality control, which is not the case of J2K. An NLS compres-sion using a PAE of 4 allows a division by 2 of the archivalrequirements, compared with lossless compression, with no vi-sual impact.

    As an extension of this paper, complementary experimentshave been carried out, particularly on virtual microscopy slides.These are really huge anatomical pathology images that require

    from hundreds of megabytes to tens of gigabytes. Some pre-liminary tests on those images have given promising results.HOP obtained 10% lossless compression improvements com-pared with CALIC. The uses and needs for this emerging imagemodality are considerable, which is why we will focus our fu-ture work on it.

    APPENDIX

    DETAILS ON THE TWO-STAGE ENTROPY CODING

    The first-stage coder (S0C) performs compression with a lim-

    ited-parameter infinite alphabet adaptive coder. Each symbol

    is decomposed in

    (19)

    (20)

    with being the number of bits required to represent

    . This Golomb-Rice-like decomposition is then

    coded by adaptive binary arithmetic coders: one for each

    bit of (mainly noise), and one for the unary representation

    of (for IHINT, is fixed to 0. Since no predictors

    adaptation is done and is not used during prediction,

    better compression results were obtained like this).

    The binary arithmetic coders, used for , independentlymaintain the frequencies of the symbols 0 and 1 .

    When reaches a threshold , and are di-

    vided by 2. It allows a better adaptation to local statistics. The

    binary arithmetic coder used for the unary representation of

    compresses a sequence of ones followed by one zero, and

    also continuously maintains the frequencies of the symbols 0

    and 1. From experiments, since it is less dependent on local sta-

    tistics, the threshold only serves to ensure that the integer

    capacities are not reached. Some details on this unary arithmetic

    coding approach can be found in [23], where it is originally de-

    signed for Laplacian distributions sources.

    During thefi

    rst-stage coding, as described before, the statis-tics of a size-adaptive reduced alphabet, which includes an es-

    cape symbol, are learned. Starting with an alphabet that only

    contains the escape symbol (ESC) with frequency ESC

    , the size-adaptive reduced alphabet is designed as follows.

    When a symbol is encountered, ESC is incre-

    mented by one, is set to , and .

    When a symbol is encountered, is incre-

    mented by one.

    When SF ESC is greater than

    threshold , frequencies are reset: The frequency

    of each symbol ESC is set to , and

    ESC is set to ESC . Every symbol

    with null frequency is removed from .The switching to the second stage occurs when the most

    frequent symbols have been taken (referring to escape symbol

    probability). This is estimated by simply comparing ESC

    to threshold SF . If ESC is lower, it means that, most

    of the time, the encountered symbols are present in . Then, be-

    cause frequent symbols are in , ESC will be required less, and

    the ESC will probably decrease after each new frequency

    reset. Instead of waiting for it (and because ESC will

    induce a higher coding cost of other symbols in ),

    is empirically reduced to .

    Once the second-stage coding (S1C) is started, the next sym-

    bols are compressed by an adaptive arithmetic coder guided by

    the reduced alphabet , and the same rules used during its de-

    sign are used to continue to update thestatistics. When an escape

    symbol is encountered, S0C is retained to code escaped symbol

    value by Card . Because this remapping

    trick (reducing coding cost) principally induces modifications

    in the binary coder statistics used for the unary representation,

    when the coder switching occurs, the frequencies of 0 and 1

    symbols are reinitialized to 1.

    In Table VIII, some results obtained with the proposed en-

    tropy coding approach are compared with those obtained using

    the stage-0-only entropy coding and the stage-1 only entropy

    coding. It shows that S0C alone can give interesting results.

    Since S1C alone gives slightly less efficient results but the pro-posed combination of the two coders gives the best results, it

  • 7/28/2019 9base Paperbase

    12/12

    2652 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 21, NO. 5, MAY 2012

    TABLE VIIIHOP COMPRESSION RATES (bpp) USING THE STAGE-0-ONLY CODER, THE

    STAGE-1-ONLY CODER, AND THE PROPOSED COMBINATION

    can be deduced that the dynamic size alphabet allows a good

    PDF approximation once frequency tables are fed enough, and

    that a two-stage approach is relevant, particularly for resolu-

    tion progressive coding: S0C is more suited for small dimen-

    sions/resolutions images, whereas S1C boosts the compression

    of the wider residual bands.

    For all experiments of this paper, the thresholds/constants

    have been empirically fixed to , , and

    .

    ACKNOWLEDGMENT

    The authors would like to thank the reviewers and C. Hay-

    ward who helped to improve this paper.

    REFERENCES

    [1] Compression of Biomedical Images and Signals, A. Nat-Ali and C.Cavaro-Mnard, Eds.. London, U.K.: ISTE/Wiley, 2008.

    [2] P. M. A. van Ooijen, P. J. M. ten Bhomer, A. Broekema, and M.Oudkerk, Shifting storage requirements due to modality changes insix years of PACS, in Proc. 24th Int. EuroPACS Conf., Trondheim,Norway, Jun. 1517, 2006.

    [3] Y. Gaudeau and J.-M. Moureaux, Lossy compression of volumetricmedical images with 3D dead-zone lattice vector quantization, Ann.Telecommun. , vol. 64, no. 5/6, pp. 359367, Jun. 2009.

    [4] B. P. Nguyen, C.-K. Chui, S.-H. Ong, and S. Chang, An efficientcompression scheme for 4-D medical images using hierarchical vectorquantization and motion compensation, Comput. Biol. Med., vol. 41,no. 9, pp. 843856, Sep. 2011.

    [5] D. Koff, P. Bak, P. Brownrigg, D. Hosseinzadeh, A. Khademi, A. Kiss,L. Lepanto, T. Michalak, H. Shulman, and A. Volkening, Pan-cana-dian evaluation of irreversible compression ratios (lossy compres-sion) for development of national guidelines, J. Digit. Imag., vol. 22,no. 6, pp. 569578, Dec. 2009.

    [6] CAR Standards for Irreversible Compression in Digital Diag-nostic Imaging Within Radiology, Jun. 2011. [Online]. Available:http://www.car.ca/uploads/standardsguidelines/201106_EN_Stan-dard_Lossy_Compression.pdf.

    [7] M. J. Weinberger, G. Seroussi, and G. Sapiro, The LOCO-I losslessimage compression algorithm: Principles and standardization intoJPEG-LS, IEEE Trans. Image Process., vol. 9, no. 8, pp. 13091324,Aug. 2000.

    [8] X. Wu and N. Memom, Context-based, adaptative, lossless imagecoding,IEEE Trans. Commun., vol. 45, no.4, pp.437444, Apr. 1997.

    [9] D. A. Clunie, Losslesscompression of grayscale medical imagesEf-fectiveness of traditional andstateof theart approaches,in Proc. SPIE(Med. Imag.)PACS Design Evaluation, Engineering Clinical Issues,San Diego, CA, Feb. 2000, vol. 3980, pp. 7484.

    [10] P. Roos, M. A. Viergever, M. C. A. van Dijke, and J. H. Peters, Re-versible intraframe compression of medical images,IEEE Trans. Med.

    Imag., vol. 7, no. 4, pp. 328336, Dec. 1988.

    [11] A. Abrardo, L. Alparone, and F. Bartolini, Encoding-interleavedhierarchical interpolation for lossless image compression, Signal

    Process., vol. 56, no. 3, pp. 321328, Feb. 1997.[12] D. Taubman and M. Marcellin, JPEG2000: Image Compression

    Fundamentals, Standards and Practice. New York: Springer-Verlag,2001.

    [13] D. Taubman, High performance scalable image compression withEBCOT, IEEE Trans. Image Process., vol. 9, no. 7, pp. 11581170,Jul. 2000.

    [14] A. Said and W. A. Pearlman, A new, fast, and efficient image codecbased on set partitioning in hierarchical trees, IEEE Trans. CircuitsSyst. Video Technol., vol. 6, no. 3, pp. 243250, Jun. 1996.

    [15] J. Taquet and C. Labit, Near-lossless and scalable compression formedical imaging using a new adaptive hierarchical oriented predic-tion, in Proc. 17th IEEE ICIP, 2010, pp. 481484.

    [16] A. Bilgin, G. Zweig, and M. W. Marcellin, Three-dimensional imagecompression with integer wavelet transforms, Appl. Opt., vol. 39, no.11, pp. 17991814, Apr. 2000.

    [17] X. Wu and T. Qiu, Wavelet coding of volumetric medical images forhigh throughput and operability, IEEE Trans. Med. Imag., vol. 24, no.6, pp. 719727, Jun. 2005.

    [18] D. L. Donoho, De-noising by soft-thresholding, IEEE Trans. Inf.Theory, vol. 41, no. 3, pp. 613627, May 1995.

    [19] G. Ulacha and R. Stasinski, A new simple context lossless imagecoding algorithm based on adaptive context arithmetic coder, in Proc.

    IWSSIP, Jun. 2008, pp. 4548.[20] A. Krivoulets, A method for progressive near-lossless image com-

    pression, in Proc. IEEE Int. Conf. Image Process., 2003, vol. 2, pp.185188.

    [21] S. Yea and W. A. Pearlman, A wavelet-based two-stage near-losslesscoder, IEEE Trans. Image Process., vol. 15, no. 11, pp. 34883500,Nov. 2006.

    [22] A. Alecu, A. Munteanu, J. P. H. Cornelis, and P. Schelkens, Wavelet-based scalable L-infinity-oriented compression, IEEE Trans. ImageProcess., vol. 15, no. 9, pp. 24992512, Sep. 2006.

    [23] A. Krivoulets, Efficient entropy coding for image compression, ITUniv. Copenhagen, Copenhagen, Denmark, Tech. Rep. TR-2002-13,Feb. 2002.

    Jonathan Taquet was born in France in 1983. Hereceived the S.M. degree in computer science from

    the University of La Rochelle, La Rochelle, France,in 2007 and the Ph.D. degree in signal proces singand telecommunications from the University ofRennes 1, Rennes, France, in 2011, after a studyon the compression of biomedical images with theINRIA-Rennes research center.

    His major research interests include signal pro-cessing, analysis and synthesis, image compressionand communication for multimedia or medical

    applications, and machine learning.

    Claude Labit was born in France in 1956. He re-ceived the S.M. degree in computer science/telecom-munications or tfrom Ecole Nationale Suprieure desTlcommunications, Paris, France, in 1979 and thePh.D. degree in computer science and the DocteurdEtat degree from University of Rennes 1, Rennes,France, in 1982 and 1988, respectively.

    Since 1982, he has been with INRIA. In 1985, hecreated and managed the TEMIS research team foradvanced image sequence processing. From 1999 to2007, he was a Director of the INRIA-Rennes/IRISA

    Research Center. He is currently the Vice President of the scientific board atUniversity of Rennes 1 and the President of RennesAtalante Science Park. Hismajor research interests include motion analysis; image compression; image se-quence coding, both for digital video and multimedia service; and 3-D TV orlarge-scale image databases for medical or satellite applications.