We introduce a new wavelet image coding framework using context-based zerotree quantization, where an unique and efficient method for optimization of zerotree quantization is proposed. Because of the localization properties of wavelets, when a wavelet coefficient is to be quantized, the best quantizer is expected to be designed to match the statistics of the wavelet coefficients in its neighborhood, that is, the quantizer should be adaptive both in space and frequency domain. Previous image coders tended to design quantizers in a band or a class level, which limited their performances as it is difficult for the localization properties of wavelets to be exploited. Contrasting with previous coders, we propose to trace the localization properties with the combination of the tree-structured wavelet representations and adaptive models which are spatial-varying according to the local statistics. In the paper, we describe the proposed coding algorithm, where the spatial-varying models are estimated from the quantized causal neighborhoods and the zerotree pruning is based on the Lagrangian cost that can be evaluated from the statistics nearby the tree. In this way, optimization of zerotree quantization is no longer a joint optimization problem as in SFQ. Simulation results demonstrate that the coding performance is competitive, and sometimes is superior to the best results of zerotree-based coding reported in SFQ.
The copyright of the original papers published on this site belongs to IEICE. Unauthorized use of the original or translated papers is prohibited. See IEICE Provisions on Copyright for details.
Copy
Kai YANG, Hiroyuki KUDO, Tsuneo SAITO, "Wavelet Image Coding with Context-Based Zerotree Quantization Framework" in IEICE TRANSACTIONS on Information,
vol. E83-D, no. 2, pp. 211-222, February 2000, doi: .
Abstract: We introduce a new wavelet image coding framework using context-based zerotree quantization, where an unique and efficient method for optimization of zerotree quantization is proposed. Because of the localization properties of wavelets, when a wavelet coefficient is to be quantized, the best quantizer is expected to be designed to match the statistics of the wavelet coefficients in its neighborhood, that is, the quantizer should be adaptive both in space and frequency domain. Previous image coders tended to design quantizers in a band or a class level, which limited their performances as it is difficult for the localization properties of wavelets to be exploited. Contrasting with previous coders, we propose to trace the localization properties with the combination of the tree-structured wavelet representations and adaptive models which are spatial-varying according to the local statistics. In the paper, we describe the proposed coding algorithm, where the spatial-varying models are estimated from the quantized causal neighborhoods and the zerotree pruning is based on the Lagrangian cost that can be evaluated from the statistics nearby the tree. In this way, optimization of zerotree quantization is no longer a joint optimization problem as in SFQ. Simulation results demonstrate that the coding performance is competitive, and sometimes is superior to the best results of zerotree-based coding reported in SFQ.
URL: https://global.ieice.org/en_transactions/information/10.1587/e83-d_2_211/_p
Copy
@ARTICLE{e83-d_2_211,
author={Kai YANG, Hiroyuki KUDO, Tsuneo SAITO, },
journal={IEICE TRANSACTIONS on Information},
title={Wavelet Image Coding with Context-Based Zerotree Quantization Framework},
year={2000},
volume={E83-D},
number={2},
pages={211-222},
abstract={We introduce a new wavelet image coding framework using context-based zerotree quantization, where an unique and efficient method for optimization of zerotree quantization is proposed. Because of the localization properties of wavelets, when a wavelet coefficient is to be quantized, the best quantizer is expected to be designed to match the statistics of the wavelet coefficients in its neighborhood, that is, the quantizer should be adaptive both in space and frequency domain. Previous image coders tended to design quantizers in a band or a class level, which limited their performances as it is difficult for the localization properties of wavelets to be exploited. Contrasting with previous coders, we propose to trace the localization properties with the combination of the tree-structured wavelet representations and adaptive models which are spatial-varying according to the local statistics. In the paper, we describe the proposed coding algorithm, where the spatial-varying models are estimated from the quantized causal neighborhoods and the zerotree pruning is based on the Lagrangian cost that can be evaluated from the statistics nearby the tree. In this way, optimization of zerotree quantization is no longer a joint optimization problem as in SFQ. Simulation results demonstrate that the coding performance is competitive, and sometimes is superior to the best results of zerotree-based coding reported in SFQ.},
keywords={},
doi={},
ISSN={},
month={February},}
Copy
TY - JOUR
TI - Wavelet Image Coding with Context-Based Zerotree Quantization Framework
T2 - IEICE TRANSACTIONS on Information
SP - 211
EP - 222
AU - Kai YANG
AU - Hiroyuki KUDO
AU - Tsuneo SAITO
PY - 2000
DO -
JO - IEICE TRANSACTIONS on Information
SN -
VL - E83-D
IS - 2
JA - IEICE TRANSACTIONS on Information
Y1 - February 2000
AB - We introduce a new wavelet image coding framework using context-based zerotree quantization, where an unique and efficient method for optimization of zerotree quantization is proposed. Because of the localization properties of wavelets, when a wavelet coefficient is to be quantized, the best quantizer is expected to be designed to match the statistics of the wavelet coefficients in its neighborhood, that is, the quantizer should be adaptive both in space and frequency domain. Previous image coders tended to design quantizers in a band or a class level, which limited their performances as it is difficult for the localization properties of wavelets to be exploited. Contrasting with previous coders, we propose to trace the localization properties with the combination of the tree-structured wavelet representations and adaptive models which are spatial-varying according to the local statistics. In the paper, we describe the proposed coding algorithm, where the spatial-varying models are estimated from the quantized causal neighborhoods and the zerotree pruning is based on the Lagrangian cost that can be evaluated from the statistics nearby the tree. In this way, optimization of zerotree quantization is no longer a joint optimization problem as in SFQ. Simulation results demonstrate that the coding performance is competitive, and sometimes is superior to the best results of zerotree-based coding reported in SFQ.
ER -