Taeyoung JUNG Hyuk-Ju KWON Joonku HAHN Sung-Hak LEE
We propose image synthesizing using luminance adapted range compression and detail-preserved blending. Range compression is performed using the correlated visual gamma then image blending is performed by local adaptive mixing and selecting method. Simulations prove that the proposed method reproduces natural images without any increase in noise or color desaturation.
Chanho JUNG Sanghyun JOO Do-Won NAM Wonjun KIM
In this paper, we aim to investigate the potential usefulness of machine learning in image quality assessment (IQA). Most previous studies have focused on designing effective image quality metrics (IQMs), and significant advances have been made in the development of IQMs over the last decade. Here, our goal is to improve prediction outcomes of “any” given image quality metric. We call this the “IQM's Outcome Improvement” problem, in order to distinguish the proposed approach from the existing IQA approaches. We propose a method that focuses on the underlying IQM and improves its prediction results by using machine learning techniques. Extensive experiments have been conducted on three different publicly available image databases. Particularly, through both 1) in-database and 2) cross-database validations, the generality and technological feasibility (in real-world applications) of our machine-learning-based algorithm have been evaluated. Our results demonstrate that the proposed framework improves prediction outcomes of various existing commonly used IQMs (e.g., MSE, PSNR, SSIM-based IQMs, etc.) in terms of not only prediction accuracy, but also prediction monotonicity.
Antonio CEDILLO-HERNANDEZ Manuel CEDILLO-HERNANDEZ Francisco GARCIA-UGALDE Mariko NAKANO-MIYATAKE Hector PEREZ-MEANA
A visible watermarking technique to provide copyright protection for portrait images is proposed in this paper. The proposal is focused on real-world applications where a portrait image is printed and illegitimately used for commercial purposes. It is well known that this is one of the most difficult challenges to prove ownership through current watermark techniques. We propose an original approach which avoids the deficiencies of typical watermarking methods in practical scenarios by introducing a smart process to automatically detect the most suitable region of the portrait image, where the visible watermark goes unnoticed to the naked eye of a viewer and is robust enough to remain visible when printed. The position of the watermark is determined by performing an analysis of the portrait image characteristics taking into account several conditions of their spatial information together with human visual system properties. Once the location is set, the watermark embedding process is performed adaptively by creating a contrast effect between the watermark and its background. Several experiments are performed to illustrate the proper functioning of the proposed watermark algorithm on portrait images with different characteristics, including dimensions, backgrounds, illumination and texture, with the conclusion that it can be applied in many practical situations.
Yazhong ZHANG Jinjian WU Guangming SHI Xuemei XIE Yi NIU Chunxiao FAN
Reduced-reference (RR) image quality assessment (IQA) algorithm aims to automatically evaluate the distorted image quality with partial reference data. The goal of RR IQA metric is to achieve higher quality prediction accuracy using less reference information. In this paper, we introduce a new RR IQA metric by quantifying the difference of discrete cosine transform (DCT) entropy features between the reference and distorted images. Neurophysiological evidences indicate that the human visual system presents different sensitivities to different frequency bands. Moreover, distortions on different bands result in individual quality degradations. Therefore, we suggest to calculate the information degradation on each band separately for quality assessment. The information degradations are firstly measured by the entropy difference of reorganized DCT coefficients. Then, the entropy differences on all bands are pooled to obtain the quality score. Experimental results on LIVE, CSIQ, TID2008, Toyama and IVC databases show that the proposed method performs highly consistent with human perception with limited reference data (8 values).
Hang ZHANG Yong DING Peng Wei WU Xue Tong BAI Kai HUANG
Visual quality evaluation is crucially important for various video and image processing systems. Traditionally, subjective image quality assessment (IQA) given by the judgments of people can be perfectly consistent with human visual system (HVS). However, subjective IQA metrics are cumbersome and easily affected by experimental environment. These problems further limits its applications of evaluating massive pictures. Therefore, objective IQA metrics are desired which can be incorporated into machines and automatically evaluate image quality. Effective objective IQA methods should predict accurate quality in accord with the subjective evaluation. Motivated by observations that HVS is highly adapted to extract irregularity information of textures in a scene, we introduce multifractal formalism into an image quality assessment scheme in this paper. Based on multifractal analysis, statistical complexity features of nature images are extracted robustly. Then a novel framework for image quality assessment is further proposed by quantifying the discrepancies between multifractal spectrums of images. A total of 982 images are used to validate the proposed algorithm, including five type of distortions: JPEG2000 compression, JPEG compression, white noise, Gaussian blur, and Fast Fading. Experimental results demonstrate that the proposed metric is highly effective for evaluating perceived image quality and it outperforms many state-of-the-art methods.
A specification for digital cinema systems which deal with movies digitally from production to delivery as well as projection on the screens is recommended by DCI (Digital Cinema Initiative), and the systems based on this specification have already been developed and installed in theaters. The parameters of the systems that play an important role in determining image quality include image resolution, quantization bit depth, color space, gamma characteristics, and data compression methods. This paper comparatively discusses a relation between required bit depth and gamma quantization using both of a human visual system for grayscale images and two color difference models for color images. The required bit depth obtained from a contrast sensitivity function against grayscale images monotonically decreases as the gamma value increases, while it has a minimum value when the gamma is 2.9 to 3.0 from both of the CIE 1976 L* a* b* and CIEDE2000 color difference models. It is also shown that the bit depth derived from the contrast sensitivity function is one bit greater than that derived from the color difference models at the gamma value of 2.6. Moreover, a comparison between the color differences computed with the CIE 1976 L* a* b* and CIEDE2000 leads to a same result from the view point of the required bit depth for digital cinema systems.
Luong Pham VAN Hoyoung LEE Jaehwan KIM Byeungwoo JEON
Blocking artifacts are introduced in many block-based coding systems, and its reduction can significantly improve the subjective quality of compressed video. The H.264/AVC uses an in-loop deblocking filter to remove the blocking artifacts. The filter considers some coding conditions in its adaptive deblocking filtering such as coded block pattern (CBP), motion vector, macroblock type, etc. for inter-predicted blocks, however, it does not consider much for intra-coded blocks. In this paper, we utilize the human visual system (HVS) characteristic and the local characteristic of image blocks to modify the boundary strength (BS) of the intra-deblocking filter in order to gain improvement in the subjective quality and also to reduce the complexity in filtering intra coded slices. In addition, we propose a low-complexity deblocking method which utilizes the correlation between vertical and horizontal boundaries of a block in inter coded slices. Experimental results show that our proposed method achieves not only significant gain in the subjective quality but also some PSNR gain, and reduces the computational complexity of the deblocking filter by 36.23% on average.
Xin LIAO Qiaoyan WEN Jie ZHANG
In this letter, a novel steganographic method with four-pixel differencing and exploiting modification direction is proposed. Secret data are embedded into each four-pixel block by adaptively applying exploiting modification direction technique. The difference value of the four-pixel block is used to judge whether the pixels in edge areas can tolerate larger changes than those in smooth areas. The readjustment guarantees to extract the secret data exactly and to minimize the embedding distortion. Since the proposed method processes non-overlapping 22 pixels blocks instead of two consecutive pixels, the features of edge can be considered sufficiently. Compared with the previous method, experimental results show that the proposed method provides better performance, i.e., larger embedding capacity and better image quality.
Recent advances in 3-D technologies draw an interest on the just noticeable difference in depth (JNDD) that describes a perceptual threshold of depth differences. In this letter, we address a new application of the JNDD to the depth image enhancement. In the proposed algorithm, a depth image is first segmented into multiple layers and then the depth range of the layer is expanded if the depth difference between adjacent layers is smaller than the JNDD. Therefore, viewers can effectively perceive the depth differences between layers and thus the human depth perception can be improved. The proposed algorithm can be applied to any depth-based 3-D display applications.
Joo Myoung SEOK Junggon KO Younghun LEE Doug Young SUH
For the panoramic video streaming service, this letter proposes a visual perception-based view navigation trick mode (VP-VNTM) that reduces bandwidth requirements by adjusting the quality of transmitting views in accordance with the view navigation velocity without decreasing the user's visual sensitivity. Experiments show that the proposed VP-VNTM reduces bandwidth requirements by more than 44%.
In this paper, we propose a spatially adaptive gradient-projection algorithm for the H.264 video coding standard to remove coding artifacts using local statistics. A hybrid method combining a new weighted constrained least squares (WCLS) approach and the projection onto convex sets (POCS) approach is introduced, where weighting components are determined on the basis of the human visual system (HVS) and projection set is defined by the difference between adjacent pixels and the quantization index (QI). A new visual function is defined to determine the weighting matrices controlling the degree of global smoothness, and a projection set is used to obtain a solution satisfying local smoothing constraints, so that the coding artifacts such as blocking and ringing artifacts can be simultaneously removed. The experimental results show the capability and efficiency of the proposed algorithm.
Miao SONG Keizo SHINOMORI Shiyong ZHANG
Visual adaptation is a universal phenomenon associated with human visual system. This adaptation affects not only the perception of low-level visual systems processing color, motion, and orientation, but also the perception of high-level visual systems processing complex visual patterns, such as facial identity and expression. Although it remains unclear for the mutual interaction mechanism between systems at different levels, this issue is the key to understand the hierarchical neural coding and computation mechanism. Thus, we examined whether the low-level adaptation influences on the high-level aftereffect by means of cross-level adaptation paradigm (i.e. color, figure adaptation versus facial identity adaptation). We measured the identity aftereffects within the real face test images on real face, color chip and figure adapting conditions. The cross-level mutual influence was evaluated by the aftereffect size among different adapting conditions. The results suggest that the adaptation to color and figure contributes to the high-level facial identity aftereffect. Besides, the real face adaptation obtained the significantly stronger aftereffect than the color chip or the figure adaptation. Our results reveal the possibility of cross-level adaptation propagation and implicitly indicate a high-level holistic facial neural representation. Based on these results, we discussed the theoretical implication of cross-level adaptation propagation for understanding the hierarchical sensory neural systems.
Objective assessment of image and video quality should be based on a correct understanding of subjective assessment by human observers. Previous models have incorporated the mechanisms of early visual processing in image quality metrics, enabling us to evaluate the visibility of errors from the original images. However, to understand how human observers perceive image quality, one should also consider higher stages of visual processing where perception is established. In higher stages, the visual system presumably represents a visual scene as a collection of meaningful components such as objects and events. Our recent psychophysical studies suggest two principles related to this level of processing. First, the human visual system integrates shape and color signals along perceived motion trajectories in order to improve visibility of the shape and color of moving objects. Second, the human visual system estimates surface reflectance properties like glossiness using simple image statistics rather than by inverse computation of image formation optics. Although the underlying neural mechanisms are still under investigation, these computational principles are potentially useful for the development of effective image processing technologies and for quality assessment. Ideally, if a model can specify how a given image is transformed into high-level scene representations in the human brain, it would predict many aspects of subjective image quality, including fidelity and naturalness.
Kwanghyun LEE Suyoung PARK Sanghoon LEE
For the acquisition of visual information, the nonuniform sampling process by photoreceptors on the retina occurs at the earliest stage of visual processing. From objects of interest, the human eye receives high visual resolution through nonuniform distribution of photoreceptors. Therefore, this paper proposes auto exposure and focus algorithms for the real-time video camera system based on the visual characteristic of the human eye. For given moving objects, the visual weight is modeled for quantifying the visual importance and the associated auto exposure and focus parameters are derived by applying the weight to the traditional numerical expression, i.e., the DoM (Difference of Median) and Tenengrad methods for auto focus.
In this paper, a perceptually adaptive watermarking scheme for color images is proposed in order to achieve robustness and transparency. A new just noticeable distortion (JND) estimator for color images is first designed in the wavelet domain. The key issue of the JND model is to effectively integrate visual masking effects. The estimator is an extension to the perceptual model that is used in image coding for grayscale images. Except for the visual masking effects given coefficient by coefficient by taking into account the luminance content and the texture of grayscale images, the crossed masking effect given by the interaction between luminance and chrominance components and the effect given by the variance within the local region of the target coefficient are investigated such that the visibility threshold for the human visual system (HVS) can be evaluated. In a locally adaptive fashion based on the wavelet decomposition, the estimator applies to all subbands of luminance and chrominance components of color images and is used to measure the visibility of wavelet quantization errors. The subband JND profiles are then incorporated into the proposed color image watermarking scheme. Performance in terms of robustness and transparency of the watermarking scheme is obtained by means of the proposed approach to embed the maximum strength watermark while maintaining the perceptually lossless quality of the watermarked color image. Simulation results show that the proposed scheme with inserting watermarks into luminance and chrominance components is more robust than the existing scheme while retaining the watermark transparency.
Wavelet tree based watermarking algorithms are using the wavelet coefficient energy difference for copyright protection and ownership verification. WTQ (Wavelet Tree Quantization) algorithm is the representative technique using energy difference for watermarking. According to the cryptanalysis on WTQ, the watermark embedded in the protected image can be removed successfully. In this paper, we present a novel differential energy watermarking algorithm based on the wavelet tree group modulation structure, i.e. WTGM (Wavelet Tree Group Modulation). The wavelet coefficients of host image are divided into disjoint super trees (each super tree containing two sub-super trees). The watermark is embedded in the relatively high-frequency components using the group strategy such that energies of sub-super trees are close. The employment of wavelet tree structure, sum-of-subsets and positive/negative modulation effectively improve the drawbacks of the WTQ scheme for its insecurity. The integration of the HVS (Human Visual System) for WTGM provides a better visual effect of the watermarked image. The experimental results demonstrate the effectiveness of our algorithm in terms of robustness and imperceptibility.
Jong-Hwan OH Byoung-Ju YUN Se-Yun KIM Kil-Houm PARK
The TFT-LCD image has non-uniform brightness that is the major difficulty of finding the visible defect called Mura in the field. To facilitate Mura detection, background signal shading should level off and Mura signal must be amplified. In this paper, Mura signal amplification and background signal flattening method is proposed based on human visual system (HVS). The proposed DC normalized contrast sensitivity function (CSF) is used for the Mura signal amplification and polynomial regression (PR) is used to level off the background signal. In the enhanced image, tri-modal thresholding segmentation technique is used for finding Dark and White Mura at the same time. To select reliable defect, falsely detected invisible region is eliminated based on Weber's Law. By the experimental results of artificially generated 1-d signal and TFT-LCD image, proposed algorithm has novel enhancement results and can be applied to real automated inspection system.
Jong-Hwan OH Woo-Seob KIM Chan-Ho HAN Kil-Houm PARK
The thin film transistor liquid crystal display (TFT-LCD) image has nonuniform brightness, which is a major difficulty in finding the Mura defect region. To facilitate Mura segmentation, globally widely varying background signal must be flattened and then Mura signal must be enhanced. In this paper, Mura signal enhancement and background-signal-flattening methods using wavelet coefficient processing are proposed. The wavelet approximation coefficients are used for background-signal flattening, while wavelet detail coefficients are employed to magnify the Mura signal on the basis of an adapted contrast sensitivity function (CSF). Then, for the enhanced image, trimodal thresholding segmentation technique and a false-region elimination method based on the human visual system (HVS) are employed for reliable Mura segmentation. The experimental results show that the proposed algorithms produce promising results and can be applied to automated inspection systems for finding Muras in a TFT-LCD image.
This paper proposes a design method for representing monochrome medical X-ray images on an electronic display. The required quantizing resolution of the input density and output voltage are theoretically clarified. The proposed method makes it easier to determine the required quantizing resolution which is important in a X-ray diagnostic system.
Lijie WANG Takahiko HORIUCHI Hiroaki KOTERA
Adaptation process of retina helps human visual system to see a high dynamic range scene in real world. This paper presents a simple static local adaptation method for high dynamic range image compression based on a retinal model. The proposed simple model aims at recreating the same sensations between the real scene and the range compressed image on display device when viewed after reaching steady state local adaptation respectively. Our new model takes the display adaptation into account in relation to the scene adaptation based on the retinal model. In computing local adaptation, the use of nonlinear edge preserving bilateral filter presents a better tonal rendition in preserving the local contrast and details while avoiding banding artifacts normally seen in local methods. Finally, we demonstrate the effectiveness of the proposed model by estimating the color difference between the recreated image and the target visual image obtained by trial and error method.