The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] quantization(222hit)

21-40hit(222hit)

  • Rapid Generation of the State Codebook in Side Match Vector Quantization

    Hanhoon PARK  Jong-Il PARK  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2017/05/16
      Vol:
    E100-D No:8
      Page(s):
    1934-1937

    Side match vector quantization (SMVQ) has been originally developed for image compression and is also useful for steganography. SMVQ requires to create its own state codebook for each block in both encoding and decoding phases. Since the conventional method for the state codebook generation is extremely time-consuming, this letter proposes a fast generation method. The proposed method is tens times faster than the conventional one without loss of perceptual visual quality.

  • A High-Throughput and Compact Hardware Implementation for the Reconstruction Loop in HEVC Intra Encoding

    Yibo FAN  Leilei HUANG  Zheng XIE  Xiaoyang ZENG  

     
    PAPER-Integrated Electronics

      Vol:
    E100-C No:6
      Page(s):
    643-654

    In the newly finalized video coding standard, namely high efficiency video coding (HEVC), new notations like coding unit (CU), prediction unit (PU) and transformation unit (TU) are introduced to improve the coding performance. As a result, the reconstruction loop in intra encoding is heavily burdened to choose the best partitions or modes for them. In order to solve the bottleneck problems in cycle and hardware cost, this paper proposed a high-throughput and compact implementation for such a reconstruction loop. By “high-throughput”, it refers to that it has a fixed throughput of 32 pixel/cycle independent of the TU/PU size (except for 4×4 TUs). By “compact”, it refers to that it fully explores the reusability between discrete cosine transform (DCT) and inverse discrete cosine transform (IDCT) as well as that between quantization (Q) and de-quantization (IQ). Besides the contributions made in designing related hardware, this paper also provides a universal formula to analyze the cycle cost of the reconstruction loop and proposed a parallel-process scheme to further reduce the cycle cost. This design is verified on the Stratix IV FPGA. The basic structure achieved a maximum frequency of 150MHz and a hardware cost of 64K ALUTs, which could support the real time TU/PU partition decision for 4K×2K@20fps videos.

  • Improved Block Truncation Coding Using Multimode Color Conversion to Reduce Frame Memory in LCD Overdrive

    Moonki CHO  Yungsup YOON  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:1
      Page(s):
    251-258

    The overdrive technique is widely used to eliminate motion blur in liquid-crystal displays (LCDs). However, this technique requires a large frame memory to store the previous frame. A reduction in the frame memory requires an image compression algorithm suitable for real-time data processing. In this paper, we present an algorithm based on multimode-color-conversion block truncation coding (MCC-BTC) to obtain a constant output bit rate and high overdrive performance. The MCC-BTC algorithm uses four compression methods, one of which is selected. The four compression modes either use the single-bitmap-generation method or the subsampling method for chrominance. As shown in the simulation results, the proposed algorithm improves the performance of both coding (up to 2.73dB) and overdrive (up to 2.61dB), and the visual quality is improved in comparison to other competing algorithms in literature.

  • A Low-Power VLSI Architecture for HEVC De-Quantization and Inverse Transform

    Heming SUN  Dajiang ZHOU  Shuping ZHANG  Shinji KIMURA  

     
    PAPER

      Vol:
    E99-A No:12
      Page(s):
    2375-2387

    In this paper, we present a low-power system for the de-quantization and inverse transform of HEVC. Firstly, we present a low-delay circuit to process the coded results of the syntax elements, and then reduce the number of multipliers from 16 to 4 for the de-quantization process of each 4x4 block. Secondly, we give two efficient data mapping schemes for the memory between de-quantization and inverse transform, and the memory for transpose. Thirdly, the zero information is utilized through the whole system. For two memory parts, the write and read operation of zero blocks/ rows/ coefficients can all be skipped to save the power consumption. The results show that up to 86% power consumption can be saved for the memory part under the configuration of “Random-access” and common QPs. For the logical part, the proposed architecture for de-quantization can reduce 77% area consumption. Overall, our system can support real-time coding for 8K x 4K 120fps video sequences and the normalized area consumption can be reduced by 68% compared with the latest work.

  • Speeding up Deep Neural Networks in Speech Recognition with Piecewise Quantized Sigmoidal Activation Function

    Anhao XING  Qingwei ZHAO  Yonghong YAN  

     
    LETTER-Acoustic modeling

      Pubricized:
    2016/07/19
      Vol:
    E99-D No:10
      Page(s):
    2558-2561

    This paper proposes a new quantization framework on activation function of deep neural networks (DNN). We implement fixed-point DNN by quantizing the activations into powers-of-two integers. The costly multiplication operations in using DNN can be replaced with low-cost bit-shifts to massively save computations. Thus, applying DNN-based speech recognition on embedded systems becomes much easier. Experiments show that the proposed method leads to no performance degradation.

  • Linear Programming Phase Feeding Method for Phased-Array Scanning

    Yi ZHANG  Guoqiang ZHAO  Houjun SUN  Mang HE  Qiang CHEN  

     
    BRIEF PAPER-Electromagnetic Theory

      Vol:
    E99-C No:7
      Page(s):
    892-894

    Digital phase shifters are widely used to achieve space scanning in phased array antenna, and beam pointing accuracy depends on the bit number and resolution of the digital phase shifter. This paper proposes a novel phase feeding method to reduce the phase quantization error effects. A linear formula for the beam pointing deviation of a linear uniform array in condition of phase quantization error is derived, and the linear programming algorithm is introduced to achieve the minimum beam pointing deviation. Simulations are based on the pattern of the phased array, which gives each element a certain quantization phase error to find the beam pointing deviation. The novel method is then compared with previous methods. Examples show that a 32-element uniform linear array with 5-bit phase shifters using the proposed method can achieve a higher beam-steering accuracy than the same array with 11-bit phase shifters.

  • Method of Spread Spectrum Watermarking Using Quantization Index Modulation for Cropped Images

    Takahiro YAMAMOTO  Masaki KAWAMURA  

     
    PAPER-Data Engineering, Web Information Systems

      Pubricized:
    2015/04/16
      Vol:
    E98-D No:7
      Page(s):
    1306-1315

    We propose a method of spread spectrum digital watermarking with quantization index modulation (QIM) and evaluate the method on the basis of IHC evaluation criteria. The spread spectrum technique can make watermarks robust by using spread codes. Since watermarks can have redundancy, messages can be decoded from a degraded stego-image. Under IHC evaluation criteria, it is necessary to decode the messages without the original image. To do so, we propose a method in which watermarks are generated by using the spread spectrum technique and are embedded by QIM. QIM is an embedding method that can decode without an original image. The IHC evaluation criteria include JPEG compression and cropping as attacks. JPEG compression is lossy compression. Therefore, errors occur in watermarks. Since watermarks in stego-images are out of synchronization due to cropping, the position of embedded watermarks may be unclear. Detecting this position is needed while decoding. Therefore, both error correction and synchronization are required for digital watermarking methods. As countermeasures against cropping, the original image is divided into segments to embed watermarks. Moreover, each segment is divided into 8×8 pixel blocks. A watermark is embedded into a DCT coefficient in a block by QIM. To synchronize in decoding, the proposed method uses the correlation between watermarks and spread codes. After synchronization, watermarks are extracted by QIM, and then, messages are estimated from the watermarks. The proposed method was evaluated on the basis of the IHC evaluation criteria. The PSNR had to be higher than 30 dB. Ten 1920×1080 rectangular regions were cropped from each stego-image, and 200-bit messages were decoded from these regions. Their BERs were calculated to assess the tolerance. As a result, the BERs were less than 1.0%, and the average PSNR was 46.70 dB. Therefore, our method achieved a high image quality when using the IHC evaluation criteria. In addition, the proposed method was also evaluated by using StirMark 4.0. As a result, we found that our method has robustness for not only JPEG compression and cropping but also additional noise and Gaussian filtering. Moreover, the method has an advantage in that detection time is small since the synchronization is processed in 8×8 pixel blocks.

  • Audio Watermarking Based on Eigenvalue Decomposition

    Pranab KUMAR DHAR  Tetsuya SHIMAMURA  

     
    LETTER-Cryptography and Information Security

      Vol:
    E97-A No:12
      Page(s):
    2658-2661

    This letter presents a new blind audio watermarking scheme using eigenvalue decomposition (EVD). Initially, the original audio is divided into frames and the samples of each frame are arranged into a square matrix. EVD is applied to each of these matrices. Watermark data is embedded into the largest eigenvalue of each diagonal matrix by quantization. Data embedding rate of the proposed scheme is 172.39bps. Simulation results confirm the imperceptibility of the proposed scheme and its higher robustness against various attacks compared to the state-of-the-art watermarking methods available in the literature.

  • Region-Based Distributed Estimation Using Quantized Data

    Yoon Hak KIM  

     
    PAPER-Information Network

      Vol:
    E97-D No:12
      Page(s):
    3155-3162

    In this paper, we consider distributed estimation where the measurement at each of the distributed sensor nodes is quantized before being transmitted to a fusion node which produces an estimate of the parameter of interest. Since each quantized measurement can be linked to a region where the parameter is found, aggregating the information obtained from multiple nodes corresponds to generating intersections between the regions. Thus, we develop estimation algorithms that seek to find the intersection region with the maximum likelihood rather than the parameter itself. Specifically, we propose two practical techniques that facilitate fast search with significantly reduced complexity and apply the proposed techniques to a system where an acoustic amplitude sensor model is employed at each node for source localization. Our simulation results show that our proposed algorithms achieve good performance with reasonable complexity as compared with the minimum mean squared error (MMSE) and the maximum likelihood (ML) estimators.

  • Performance of Partitioned Vector Quantization with Optimized Feedback Budget Allocation

    Mirza Golam KIBRIA  Hidekazu MURATA  Susumu YOSHIDA  

     
    PAPER-Terrestrial Wireless Communication/Broadcasting Technologies

      Vol:
    E97-B No:6
      Page(s):
    1184-1194

    This study analyzes the performance of a downlink beamformer with partitioned vector quantization under optimized feedback budget allocation. A multiuser multiple-input single-output downlink precoding system with perfect channel state information at mobile stations is considered. The number of feedback bits allocated to the channel quality indicator (CQI) and the channel direction indicator (CDI) corresponding to each partition are optimized by exploiting the quantization mean square error. In addition, the effects of equal and unequal partitioning on codebook memory and system capacity are studied and elucidated through simulations. The results show that with optimized CQI-CDI allocation, the feedback budget distributions of equal or unequal partitions are proportional to the size ratios of the partitioned subvectors. Furthermore, it is observed that for large-sized partitions, the ratio of optimal CDI to CQI is much higher than that for small-sized partitions.

  • Adaptive Spectral Masking of AVQ Coding and Sparseness Detection for ITU-T G.711.1 Annex D and G.722 Annex B Standards

    Masahiro FUKUI  Shigeaki SASAKI  Yusuke HIWASAKI  Kimitaka TSUTSUMI  Sachiko KURIHARA  Hitoshi OHMURO  Yoichi HANEDA  

     
    PAPER-Speech and Hearing

      Vol:
    E97-D No:5
      Page(s):
    1264-1272

    We proposes a new adaptive spectral masking method of algebraic vector quantization (AVQ) for non-sparse signals in the modified discreet cosine transform (MDCT) domain. This paper also proposes switching the adaptive spectral masking on and off depending on whether or not the target signal is non-sparse. The switching decision is based on the results of MDCT-domain sparseness analysis. When the target signal is categorized as non-sparse, the masking level of the target MDCT coefficients is adaptively controlled using spectral envelope information. The performance of the proposed method, as a part of ITU-T G.711.1 Annex D, is evaluated in comparison with conventional AVQ. Subjective listening test results showed that the proposed method improves sound quality by more than 0.1 points on a five-point scale on average for speech, music, and mixed content, which indicates significant improvement.

  • Broadcast Control of Multi-Agent Systems with Quantized Measurements

    Yosuke TANAKA  Shun-ichi AZUMA  Toshiharu SUGIE  

     
    PAPER-Systems and Control

      Vol:
    E97-A No:3
      Page(s):
    830-839

    This paper addresses a broadcast control problem of multi-agent systems with quantized measurements, where each agent moves based on the common broadcasted signal and tries to minimize a given quadratic performance index. The problem is solved by introducing dither type random movements to the agents' action which reduce the degradation caused by quantized measurements. A broadcast controller is derived and it is proven that the controller approximately achieves given tasks with probability 1. The effectiveness of the proposed controller is demonstrated by numerical simulation.

  • Discrete Abstraction of Stochastic Nonlinear Systems

    Shun-ichi AZUMA  George J. PAPPAS  

     
    PAPER

      Vol:
    E97-A No:2
      Page(s):
    452-458

    This paper addresses the discrete abstraction problem for stochastic nonlinear systems with continuous-valued state. The proposed solution is based on a function, called the bisimulation function, which provides a sufficient condition for the existence of a discrete abstraction for a given continuous system. We first introduce the bisimulation function and show how the function solves the problem. Next, a convex optimization based method for constructing a bisimulation function is presented. Finally, the proposed framework is demonstrated by a numerical simulation.

  • An Efficient Compression of Amplitude-Only Images for the Image Trading System

    Shenchuan LIU  Wannida SAE-TANG  Masaaki FUJIYOSHI  Hitoshi KIYA  

     
    LETTER-Image Processing and Video Processing

      Vol:
    E97-D No:2
      Page(s):
    378-379

    This letter proposes an efficient compression scheme for the copyright- and privacy-protected image trading system. The proposed scheme multiplies pseudo random signs to amplitude components of discrete cosine transformed coefficients before the inverse transformation is applied. The proposed scheme efficiently compresses amplitude-only image which is the inversely transformed amplitude components, and the scheme simultaneously improves the compression efficiency of phase-only image which is the inversely transformed phase components, in comparison with the conventional systems.

  • A 5.83pJ/bit/iteration High-Parallel Performance-Aware LDPC Decoder IP Core Design for WiMAX in 65nm CMOS

    Xiongxin ZHAO  Zhixiang CHEN  Xiao PENG  Dajiang ZHOU  Satoshi GOTO  

     
    PAPER-High-Level Synthesis and System-Level Design

      Vol:
    E96-A No:12
      Page(s):
    2623-2632

    In this paper, we propose a synthesizable LDPC decoder IP core for the WiMAX system with high parallelism and enhanced error-correcting performance. By taking the advantages of both layered scheduling and fully-parallel architecture, the decoder can fully support multi-mode decoding specified in WiMAX with the parallelism much higher than commonly used partial-parallel layered LDPC decoder architecture. 6-bit quantized messages are split into bit-serial style and 2bit-width serial processing lines work concurrently so that only 3 cycles are required to decode one layer. As a result, 12∼24 cycles are enough to process one iteration for all the code-rates specified in WiMAX. Compared to our previous bit-serial decoder, it doubles the parallelism and solves the message saturation problem of the bit-serial arithmetic, with minor gate count increase. Power synthesis result shows that the proposed decoder achieves 5.83pJ/bit/iteration energy efficiency which is 46.8% improvement compared to state-of-the-art work. Furthermore, an advanced dynamic quantization (ADQ) technique is proposed to enhance the error-correcting performance in layered decoder architecture. With about 2% area overhead, 6-bit ADQ can achieve the error-correcting performance close to 7-bit fixed quantization with improved error floor performance.

  • An Improved Quantization Scheme for Lattice-Reduction Aided MIMO Detection Based on Gram-Schmidt Orthogonalization

    Wei HOU  Tadashi FUJINO  Toshiharu KOJIMA  

     
    PAPER-Communication Theory

      Vol:
    E96-A No:12
      Page(s):
    2405-2414

    Lattice-reduction (LR) technique has been adopted to improve the performance and reduce the complexity in MIMO data detection. This paper presents an improved quantization scheme for LR aided MIMO detection based on Gram-Schmidt orthogonalization. For the LR aided detection, the quantization step applies the simple rounding operation, which often leads to the quantization errors. Meanwhile, these errors may result in the detection errors. Hence the purpose of the proposed detection is to further solve the problem of degrading the performance due to the quantization errors in the signal estimation. In this paper, the proposed quantization scheme decreases the quantization errors using a simple tree search with a threshold function. Through the analysis and the simulation results, we observe that the proposed detection can achieve the nearly optimal performance with very low complexity, and require a little additional complexity compared to the conventional LR-MMSE detection in the high Eb/N0 region. Furthermore, this quantization error reduction scheme is also efficient even for the high modulation order.

  • Comparative Study on Required Bit Depth of Gamma Quantization for Digital Cinema Using Contrast and Color Difference Sensitivities

    Junji SUZUKI  Isao FURUKAWA  

     
    PAPER-Image

      Vol:
    E96-A No:8
      Page(s):
    1759-1767

    A specification for digital cinema systems which deal with movies digitally from production to delivery as well as projection on the screens is recommended by DCI (Digital Cinema Initiative), and the systems based on this specification have already been developed and installed in theaters. The parameters of the systems that play an important role in determining image quality include image resolution, quantization bit depth, color space, gamma characteristics, and data compression methods. This paper comparatively discusses a relation between required bit depth and gamma quantization using both of a human visual system for grayscale images and two color difference models for color images. The required bit depth obtained from a contrast sensitivity function against grayscale images monotonically decreases as the gamma value increases, while it has a minimum value when the gamma is 2.9 to 3.0 from both of the CIE 1976 L* a* b* and CIEDE2000 color difference models. It is also shown that the bit depth derived from the contrast sensitivity function is one bit greater than that derived from the color difference models at the gamma value of 2.6. Moreover, a comparison between the color differences computed with the CIE 1976 L* a* b* and CIEDE2000 leads to a same result from the view point of the required bit depth for digital cinema systems.

  • Bit-Plane Coding of Lattice Codevectors

    Wisarn PATCHOO  Thomas R. FISCHER  

     
    LETTER-Coding Theory

      Vol:
    E96-A No:8
      Page(s):
    1817-1820

    In a sign-magnitude representation of binary lattice codevectors, only a few least significant bit-planes are constrained due to the structure of the lattice, while there is no restriction on other more significant bit-planes. Hence, any convenient bit-plane coding method can be used to encode the lattice codevectors, with modification required only for the lattice-defining, least-significant bit-planes. Simple encoding methods for the lattice-defining bit-planes of the D4, RE8, and Barnes-Wall 16-dimensional lattices are described. Simulation results for the encoding of a uniform source show that standard bit-plane coding together with the proposed encoding provide about the same performance as integer lattice vector quantization when the bit-stream is truncated. When the entire bit-stream is fully decoded, the granular gain of the lattice is realized.

  • A Generation Method of Amplitude-Only Images with Low Intensity Ranges

    Wannida SAE-TANG  Masaaki FUJIYOSHI  Hitoshi KIYA  

     
    PAPER

      Vol:
    E96-A No:6
      Page(s):
    1323-1330

    In this paper, 1) it is shown that amplitude-only images (AOIs) have quite wide intensity ranges (IRs), and 2) an IR reduction method for AOIs is proposed. An AOI is the inversely transformed amplitude spectra of an image, and it is used in the privacy- and copyright-protected image trading system because of its invisibility. Since an AOI is the coherent summation of cosine waves with the same phase, the IR of the AOI is too large to be stored and/or transmitted. In the proposed method, random signs are applied to discrete Fourier transformed amplitude coefficients to obtained AOIs with significantly lower IRs without distortion while keeping the invisibility of images. With reasonable processing time, high correct watermark extracting rates, inversely quantized AOIs with low mean squared errors, and reconstructed images with high peak signal-to-noise ratios are obtained by a linear quantizer in the proposed method.

  • Development of a Robust and Compact On-Line Handwritten Japanese Text Recognizer for Hand-Held Devices

    Jinfeng GAO  Bilan ZHU  Masaki NAKAGAWA  

     
    PAPER-Pattern Recognition

      Vol:
    E96-D No:4
      Page(s):
    927-938

    The paper describes how a robust and compact on-line handwritten Japanese text recognizer was developed by compressing each component of an integrated text recognition system including a SVM classifier to evaluate segmentation points, an on-line and off-line combined character recognizer, a linguistic context processor, and a geometric context evaluation module to deploy it on hand-held devices. Selecting an elastic-matching based on-line recognizer and compressing MQDF2 via a combination of LDA, vector quantization and data type transformation, have contributed to building a remarkably small yet robust recognizer. The compact text recognizer covering 7,097 character classes just requires about 15 MB memory to keep 93.11% accuracy on horizontal text lines extracted from the TUAT Kondate database. Compared with the original full-scale Japanese text recognizer, the memory size is reduced from 64.1 MB to 14.9 MB while the accuracy loss is only 0.5% from 93.6% to 93.11%. The method is scalable so even systems of less than 11 MB or less than 6 MB still remain 92.80% or 90.02% accuracy, respectively.

21-40hit(222hit)