The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] ATI(18690hit)

3361-3380hit(18690hit)

  • Image Watermarking Method Satisfying IHC by Using PEG LDPC Code

    Nobuhiro HIRATA  Takayuki NOZAKI  Masaki KAWAMURA  

     
    PAPER

      Pubricized:
    2016/10/07
      Vol:
    E100-D No:1
      Page(s):
    13-23

    We propose a digital image watermarking method satisfying information hiding criteria (IHC) for robustness against JPEG compression, cropping, scaling, and rotation. When a stego-image is cropped, the marking positions of watermarks are unclear. To detect the position in a cropped stego-image, a marker or synchronization code is embedded with the watermarks in a lattice pattern. Attacks by JPEG compression, scaling, and rotation cause errors in extracted watermarks. Against such errors, the same watermarks are repeatedly embedded in several areas. The number of errors in the extracted watermarks can be reduced by using a weighted majority voting (WMV) algorithm. To correct residual errors in output of the WMV algorithm, we use a high-performance error-correcting code: a low-density parity-check (LDPC) code constructed by progressive edge-growth (PEG). In computer simulations using the IHC ver. 4 the proposed method could a bit error rate of 0, the average PSNR was 41.136 dB, and the computational time for synchronization recovery was less than 10 seconds. The proposed method can thus provide high image quality and fast synchronization recovery.

  • Efficient Balanced Truncation for RC and RLC Networks

    Yuichi TANJI  

     
    PAPER-Circuit Theory

      Vol:
    E100-A No:1
      Page(s):
    266-274

    An efficient balanced truncation for RC and RLC networks is presented in this paper. To accelerate the balanced truncation, sparse structures of original networks are considered. As a result, Lyapunov equations, the solutions of which are necessary for making the transformation matrices, are efficiently solved, and the reduced order models are efficiently obtained. It is proven that reciprocity of original networks is preserved while applying the proposed method. Passivity of the reduced RC networks is also guaranteed. In the illustrative examples, we will show that the proposed method is compatible with PRIMA in efficiency and is more accurate than PRIMA.

  • Secret Sharing with Cheaters Using Multi-Receiver Authentication

    Rui XU  Kirill MOROZOV  Tsuyoshi TAKAGI  

     
    PAPER

      Vol:
    E100-A No:1
      Page(s):
    115-125

    We introduce two cheater identifiable secret sharing (CISS) schemes with efficient reconstruction, tolerating t

  • Deep Nonlinear Metric Learning for Speaker Verification in the I-Vector Space

    Yong FENG  Qingyu XIONG  Weiren SHI  

     
    LETTER-Speech and Hearing

      Pubricized:
    2016/10/04
      Vol:
    E100-D No:1
      Page(s):
    215-219

    Speaker verification is the task of determining whether two utterances represent the same person. After representing the utterances in the i-vector space, the crucial problem is only how to compute the similarity of two i-vectors. Metric learning has provided a viable solution to this problem. Until now, many metric learning algorithms have been proposed, but they are usually limited to learning a linear transformation. In this paper, we propose a nonlinear metric learning method, which learns an explicit mapping from the original space to an optimal subspace using deep Restricted Boltzmann Machine network. The proposed method is evaluated on the NIST SRE 2008 dataset. Since the proposed method has a deep learning architecture, the evaluation results show superior performance than some state-of-the-art methods.

  • Analyzing Fine Motion Considering Individual Habit for Appearance-Based Proficiency Evaluation

    Yudai MIYASHITA  Hirokatsu KATAOKA  Akio NAKAMURA  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2016/10/18
      Vol:
    E100-D No:1
      Page(s):
    166-174

    We propose an appearance-based proficiency evaluation methodology based on fine-motion analysis. We consider the effects of individual habit in evaluating proficiency and analyze the fine motion of guitar-picking. We first extract multiple features on a large number of dense trajectories of fine motion. To facilitate analysis, we then generate a histogram of motion features using a bag-of-words model and change the number of visual words as appropriate. To remove the effects of individual habit, we extract the common principal histogram elements corresponding to experts or beginners according to discrimination's contribution rates using random forests. We finally calculate the similarity of the histograms to evaluate the proficiency of a guitar-picking motion. By optimizing the number of visual words for proficiency evaluation, we demonstrate that our method distinguishes experts from beginners with an accuracy of about 86%. Moreover, we verify experimentally that our proposed methodology can evaluate proficiency while removing the effects of individual habit.

  • Efficient Algorithm for Sentence Information Content Computing in Semantic Hierarchical Network

    Hao WU  Heyan HUANG  

     
    LETTER-Natural Language Processing

      Pubricized:
    2016/10/18
      Vol:
    E100-D No:1
      Page(s):
    238-241

    We previously proposed an unsupervised model using the inclusion-exclusion principle to compute sentence information content. Though it can achieve desirable experimental results in sentence semantic similarity, the computational complexity is more than O(2n). In this paper, we propose an efficient method to calculate sentence information content, which employs the thinking of the difference set in hierarchical network. Impressively, experimental results show that the computational complexity decreases to O(n). We prove the algorithm in the form of theorems. Performance analysis and experiments are also provided.

  • Low Computational Complexity Direction-of-Arrival Estimation of Wideband Signal Sources Based on Squared TOPS

    Hirotaka HAYASHI  Tomoaki OHTSUKI  

     
    PAPER

      Vol:
    E100-A No:1
      Page(s):
    219-226

    We propose a new direction-of-arrival (DOA) estimation method of wideband signals. In several decades, many approaches to estimate DOA of wideband signal sources have been proposed. Test of orthogonality of projected subspaces (TOPS) and Squared TOPS are the estimation algorithms to realize high resolution performance of closely spaced signal sources. These methods, however, are not suitable for DOA estimation of multiple signal sources, because the spatial spectrum calculated by Squared TOPS has some false peaks. Therefore, the authors have proposed the weighted squared TOPS (WS-TOPS) to suppress these false peaks by modifying the orthogonality evaluation matrix, WS-TOPS also achieves better DOA estimation accuracy than that of Squared TOPS. On the other hand, WS-TOPS has a drawback, it requires high computational complexity. Our new method can realize good DOA estimation performance, which is better than that of Squared TOPS, with low computational complexity by reducing the size of orthogonality evaluation matrix and the number of subspaces to be used. Simulation results show that the new method can provide high resolution performance and high DOA estimation accuracy with low computational complexity.

  • Improvement of Artificial Auscultation on Hemodialysis Stenosis by the Estimate of Stenosis Site and the Hierarchical Categorization of Learning Data

    Hatsuhiro KATO  Masakazu KIRYU  Yutaka SUZUKI  Osamu SAKATA  Mizuya FUKASAWA  

     
    PAPER-Biocybernetics, Neurocomputing

      Vol:
    E100-D No:1
      Page(s):
    175-180

    Many hemodialysis patients undergo plasitc surgery to form the arterio-venous fistula (AVF) in their forearm to improve the vascular access by shunting blood flows. The issue of AVF is the stenosis caused by the disturbance of blood flows; therefore the auscultation system to assist the stenosis diagnosis has been developed. Although the system is intended to be used as a steady monitoring for stenosis assessment, its efficiency was not always high because it cannot estimate where the stenosis locates. In this study, for extracting and estimating the stenosis signal, the shunt murmurs captured by many microphones were decomposed by the principal component analysis (PCA). Furthermore, applying the hierarchical categorization of the recursive subdivision self-organizing map (rs-SOM), the modelling of the stenosis signal was proposed to realise the effective stenosis assessment. The false-positive rate of the stenosis assessment was significantly reduced by using the improved auscultation system.

  • Iterative Constructions of Orthogonal Arrays of Strength t and Orthogonal Partitions

    Shanqi PANG  Ying WANG  Jiao DU  Wenju XU  

     
    LETTER-Coding Theory

      Vol:
    E100-A No:1
      Page(s):
    308-311

    Orthogonal arrays and orthogonal partitions have great significance in communications and coding theory. In this letter, by using a generalized orthogonal partition, Latin squares and orthogonal Latin squares, we present an iterative construction method of orthogonal arrays of strength t and orthogonal partitions. As an application of the method, more orthogonal arrays of strength t and orthogonal partitions than the existing methods can be constructed.

  • Joint Maximum Likelihood Detection in Far User of Non-Orthogonal Multiple Access

    Kenji ANDO  Yukitoshi SANADA  Takahiko SABA  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2016/07/29
      Vol:
    E100-B No:1
      Page(s):
    177-186

    Non-orthogonal multiple access (NOMA) enables multiple mobile devices to share the same frequency band. In a conventional NOMA scheme, the receiver of a far user detects its desired signal without canceling the signal for a near user. However, the signal for the near user acts as interference and degrades the accuracy of likelihood values for the far user. In this paper, a joint maximum likelihood detection scheme for the far user of the NOMA downlink is proposed. The proposed scheme takes the interference signal into account in calculating the likelihood values. Numerical results obtained through computer simulation show that the proposed scheme improves the performance by from 0.2dB to 3.1dB for power allocation coefficients of 0.2 to 0.4 at a bit error rate (BER) of 10-2 relative to the conventional scheme.

  • An Encryption-then-Compression System for Lossless Image Compression Standards

    Kenta KURIHARA  Shoko IMAIZUMI  Sayaka SHIOTA  Hitoshi KIYA  

     
    LETTER

      Pubricized:
    2016/10/07
      Vol:
    E100-D No:1
      Page(s):
    52-56

    In many multimedia applications, image encryption has to be conducted prior to image compression. This letter proposes an Encryption-then-Compression system using JPEG XR/JPEG-LS friendly perceptual encryption method, which enables to be conducted prior to the JPEG XR/JPEG-LS standard used as an international standard lossless compression method. The proposed encryption scheme can provides approximately the same compression performance as that of the lossless compression without any encryption. It is also shown that the proposed system consists of four block-based encryption steps, and provides a reasonably high level of security. Existing conventional encryption methods have not been designed for international lossless compression standards, but for the first time this letter focuses on applying the standards.

  • Key Recovery Attacks on Multivariate Public Key Cryptosystems Derived from Quadratic Forms over an Extension Field

    Yasufumi HASHIMOTO  

     
    PAPER

      Vol:
    E100-A No:1
      Page(s):
    18-25

    One of major ideas to design a multivariate public key cryptosystem (MPKC) is to generate its quadratic forms by a polynomial map over an extension field. In fact, Matsumoto-Imai's scheme (1988), HFE (Patarin, 1996), MFE (Wang et al., 2006) and multi-HFE (Chen et al., 2008) are constructed in this way and Sflash (Akkar et al., 2003), Quartz (Patarin et al., 2001), Gui (Petzoldt et al, 2015) are variants of these schemes. An advantage of such extension field type MPKCs is to reduce the numbers of variables and equations to be solved in the decryption process. In the present paper, we study the security of MPKCs whose quadratic forms are derived from a “quadratic” map over an extension field and propose a new attack on such MPKCs. Our attack recovers partial information of the secret affine maps in polynomial time when the field is of odd characteristic. Once such partial information is recovered, the attacker can find the plain-text for a given cipher-text by solving a system of quadratic equations over the extension field whose numbers of variables and equations are same to those of the system of quadratic equations used in the decryption process.

  • A Computationally Efficient Schnorr-Euchner Enumeration for Solving Integer Least-Squares Problem in Wireless Communications

    Junil AHN  Jaewon CHANG  Chiho LEE  

     
    LETTER-Communication Theory and Signals

      Vol:
    E100-A No:1
      Page(s):
    327-331

    The integer least-squares (ILS) problem frequently arises in wireless communication systems. Sphere decoding (SD) is a systematic search scheme for solving ILS problem. The enumeration of candidates is a key part of SD for selecting a lattice point, which will be searched by the algorithm. Herein, the authors present a computationally efficient Schnorr-Euchner enumeration (SEE) algorithm to solve the constrained ILS problems, where the solution is limited into the finite integer lattice. To trace only valid lattice points within the underlying finite lattice, the authors devise an adaptive computation of the enumeration step and counting the valid points enumerated. In contrast to previous SEE methods based on a zig-zag manner, the proposed method completely avoids enumerating invalid points outside the finite lattice, and it further reduces real arithmetic and logical operations.

  • Computational Model of Card-Based Cryptographic Protocols and Its Applications

    Takaaki MIZUKI  Hiroki SHIZUYA  

     
    INVITED PAPER

      Vol:
    E100-A No:1
      Page(s):
    3-11

    Card-based protocols enable us to easily perform cryptographic tasks such as secure multiparty computation using a deck of physical cards. Since the first card-based protocol appeared in 1989, many protocols have been designed. A protocol is usually described with a series of somewhat intuitive and verbal descriptions, such as “turn over this card,” “shuffle these two cards,” “apply a random cut to these five cards,” and so on. On the other hand, a formal computational model of card-based protocols via abstract machine was constructed in 2014. By virtue of the formalization, card-based protocols can be treated more rigorously; for example, it enables one to discuss the lower bounds on the number of cards required for secure computations. In this paper, an overview of the computational model with its applications to designing protocols and a survey of the recent progress in card-based protocols are presented.

  • Home Base-Aware Store-Carry-Forward Routing Using Location-Dependent Utilities of Nodes

    Tomotaka KIMURA  Yutsuki KAYAMA  Tetsuya TAKINE  

     
    PAPER-Network

      Vol:
    E100-B No:1
      Page(s):
    17-27

    We propose a home base-aware store-carry-forward routing scheme using location-dependent utilities of nodes, which adopts different message forwarding strategies depending on location where nodes encounter. Our routing scheme first attempts to deliver messages to its home base, the area with the highest potential for the presence of the destination node in the near future. Once a message copy reaches its home base, message dissemination is limited within the home base, and nodes with message copies wait for encountering the destination node. To realize our routing scheme, we use two different utilities of nodes depending on location: Outside the home base of a message, nodes approaching to the home base have high utility values, while within the home base, nodes staying the home base have high utility values. By using these utilities properly, nodes with message copies will catch the destination node “by ambush” in the home base of the destination node. Through simulation experiments, we demonstrate the effectiveness of our routing scheme.

  • Insufficient Vectorization: A New Method to Exploit Superword Level Parallelism

    Wei GAO  Lin HAN  Rongcai ZHAO  Yingying LI  Jian LIU  

     
    PAPER-Software System

      Pubricized:
    2016/09/29
      Vol:
    E100-D No:1
      Page(s):
    91-106

    Single-instruction multiple-data (SIMD) extension provides an energy-efficient platform to scale the performance of media and scientific applications while still retaining post-programmability. However, the major challenge is to translate the parallel resources of the SIMD hardware into real application performance. Currently, all the slots in the vector register are used when compilers exploit SIMD parallelism of programs, which can be called sufficient vectorization. Sufficient vectorization means all the data in the vector register is valid. Because all the slots which vector register provides must be used, the chances of vectorizing programs with low SIMD parallelism are abandoned by sufficient vectorization method. In addition, the speedup obtained by full use of vector register sometimes is not as great as that obtained by partial use. Specifically, the length of vector register provided by SIMD extension becomes longer, sufficient vectorization method cannot exploit the SIMD parallelism of programs completely. Therefore, insufficient vectorization method is proposed, which refer to partial use of vector register. First, the adaptation scene of insufficient vectorization is analyzed. Second, the methods of computing inter-iteration and intra-iteration SIMD parallelism for loops are put forward. Furthermore, according to the relationship between the parallelism and vector factor a method is established to make the choice of vectorization method, in order to vectorize programs as well as possible. Finally, code generation strategy for insufficient vectorization is presented. Benchmark test results show that insufficient vectorization method vectorized more programs than sufficient vectorization method by 107.5% and the performance achieved by insufficient vectorization method is 12.1% higher than that achieved by sufficient vectorization method.

  • Improved Block Truncation Coding Using Multimode Color Conversion to Reduce Frame Memory in LCD Overdrive

    Moonki CHO  Yungsup YOON  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:1
      Page(s):
    251-258

    The overdrive technique is widely used to eliminate motion blur in liquid-crystal displays (LCDs). However, this technique requires a large frame memory to store the previous frame. A reduction in the frame memory requires an image compression algorithm suitable for real-time data processing. In this paper, we present an algorithm based on multimode-color-conversion block truncation coding (MCC-BTC) to obtain a constant output bit rate and high overdrive performance. The MCC-BTC algorithm uses four compression methods, one of which is selected. The four compression modes either use the single-bitmap-generation method or the subsampling method for chrominance. As shown in the simulation results, the proposed algorithm improves the performance of both coding (up to 2.73dB) and overdrive (up to 2.61dB), and the visual quality is improved in comparison to other competing algorithms in literature.

  • Low Complexity Reed-Solomon Decoder Design with Pipelined Recursive Euclidean Algorithm

    Kazuhito ITO  

     
    PAPER

      Vol:
    E99-A No:12
      Page(s):
    2453-2462

    A Reed-Solomon (RS) decoder is designed based on the pipelined recursive Euclidean algorithm in the key equation solution. While the Euclidean algorithm uses less Galois multipliers than the modified Euclidean (ME) and reformulated inversionless Berlekamp-Massey (RiBM) algorithms, division between two elements in Galois field is required. By implementing the division with a multi-cycle Galois inverter and a serial Galois multiplier, the proposed key equation solver architecture achieves lower complexity than the conventional ME and RiBM based architectures. The proposed RS (255,239) decoder reduces the hardware complexity by 25.9% with 6.5% increase in decoding latency.

  • Signal Power Estimation Based on Orthogonal Projection and Oblique Projection

    Norisato SUGA  Toshihiro FURUKAWA  

     
    LETTER-Digital Signal Processing

      Vol:
    E99-A No:12
      Page(s):
    2571-2575

    In this letter, we show the new signal power estimation method base on the subspace projection. This work mainly contributes to the SINR estimation problem because, in this research, the signal power estimation is implicitly or explicitly performed. The difference between our method and the conventional method related to this topic is the exploitation of the subspace character of the signals constructing the observed signal. As tools to perform subspace operation, we apply orthogonal projection and oblique projection which can extracts desired parameters. In the proposed scheme, the statistics of the projected observed signal by these projection are used to estimate the parameters.

  • A Bit-Write-Reducing and Error-Correcting Code Generation Method by Clustering ECC Codewords for Non-Volatile Memories

    Tatsuro KOJO  Masashi TAWADA  Masao YANAGISAWA  Nozomu TOGAWA  

     
    PAPER

      Vol:
    E99-A No:12
      Page(s):
    2398-2411

    Non-volatile memories are paid attention to as a promising alternative to memory design. Data stored in them still may be destructed due to crosstalk and radiation. We can restore the data by using error-correcting codes which require extra bits to correct bit errors. Further, non-volatile memories consume ten to hundred times more energy than normal memories in bit-writing. When we configure them using error-correcting codes, it is quite necessary to reduce writing bits. In this paper, we propose a method to generate a bit-write-reducing code with error-correcting ability. We first pick up an error-correcting code which can correct t-bit errors. We cluster its codeswords and generate a cluster graph satisfying the S-bit flip conditions. We assign a data to be written to each cluster. In other words, we generate one-to-many mapping from each data to the codewords in the cluster. We prove that, if the cluster graph is a complete graph, every data in a memory cell can be re-written into another data by flipping at most S bits keeping error-correcting ability to t bits. We further propose an efficient method to cluster error-correcting codewords. Experimental results show that the bit-write-reducing and error-correcting codes generated by our proposed method efficiently reduce energy consumption. This paper proposes the world-first theoretically near-optimal bit-write-reducing code with error-correcting ability based on the efficient coding theories.

3361-3380hit(18690hit)