1-4hit |
Yoshitaka INOUE Takamichi MIYATA Yoshinori SAKAI
Recently, a novel approach to color image compression based on colorization has been presented. The conventional method for colorization-based image coding tends to lose the local oscillation of chrominance components that the original images had. A large number of color assignments is required to restore these oscillations. On the other hand, previous studies suggest that an oscillation of a chrominance component correlates with the oscillation of a corresponding luminance component. In this paper, we propose a new colorization-based image coding method that utilizes the local correlation between texture components of luminance and chrominance. These texture components are obtained by a total variation regularized energy minimization method. The local correlation relationships are approximated by linear functions, and their coefficients are extracted by an optimization method. This key idea enables us to represent the oscillations of chrominance components by using only a few pieces of information. Experimental results showed that our method can restore the local oscillation and code images more efficiently than the conventional method, JPEG, or JPEG2000 at a high compression rate.
Shunsuke ONO Takamichi MIYATA Isao YAMADA Katsunori YAMAOKA
Solving image recovery problems requires the use of some efficient regularizations based on a priori information with respect to the unknown original image. Naturally, we can assume that an image is modeled as the sum of smooth, edge, and texture components. To obtain a high quality recovered image, appropriate regularizations for each individual component are required. In this paper, we propose a novel image recovery technique which performs decomposition and recovery simultaneously. We formulate image recovery as a nonsmooth convex optimization problem and design an iterative scheme based on the alternating direction method of multipliers (ADMM) for approximating its global minimizer efficiently. Experimental results reveal that the proposed image recovery technique outperforms a state-of-the-art method.
Takamichi MIYATA Tomonobu YOSHINO Sei NAITO
Ultra high definition (UHD) imaging systems have attracted much attention as a next generation television (TV) broadcasting service and video streaming service. However, the state of the art video coding standards including H.265/HEVC has not enough compression rate for streaming, broadcasting and storing UHD. Existing coding standard such as H.265/HEVC normaly use RGB-YCbCr color transform before compressing RGB color image since that procedure can decorrelate color components well. However, there is room for improvement on the coding efficiency for color image based on an observation that the luminance and chrominance components changes in same locations. This observation inspired us to propose a new post-processing method for compressed images by using weighted least square (WLS) filter with coded luminance component as a guide image, for refining the edges of chrominance components. Since the computational cost of WLS tends to superlinearly increase with increasing image size, it is difficult to apply it to UHD images. To overcome this problem, we propose slightly overlapped block partitioning and a new variant of WLS (constrained WLS, CWLS). Experimental results of objective quality comparison and subjective assessment test using 4K images show that our proposed method can outperform the conventional method and reduce the bit amount for chrominance component drastically with preserving the subjective quality.
Hiromi TSURUMI Takamichi MIYATA Katsunori YAMAOKA Yoshinori SAKAI
Many content distribution systems such as CDN and P2P file sharing have been developed. In these systems, file-type contents require downloads to be completed before they can be played and they have no value before the download finishes. Therefore, a user's satisfaction depends on the length of the service latency. That is, the length of time from when the user issued a request until the user received an entire file. Reducing the sum of that time is necessary for the whole delivery system to satisfy users and maintain dependability on system performance. We discuss a hop-by-hop file delivery system suitable for delivering file contents whereby the sum of service latency is reduced by using the request conditions. Moreover, we propose a file delivery scheduling algorithm for a one-link model given that the content request frequency is unknown. The algorithm is based on a local optimal strategy. We performed a characteristic analysis by computer simulation. The results showed that our algorithm performs at nearly the theoretical efficiency limit of the hop-by-hop system when the request frequency distribution of each content has a deviation.