The search functionality is under construction.

Keyword Search Result

[Keyword] error correcting code(17hit)

1-17hit
  • Construction of Locally Repairable Codes with Multiple Localities Based on Encoding Polynomial

    Tomoya HAMADA  Hideki YAGI  

     
    PAPER-Coding theory and techniques

      Vol:
    E101-A No:12
      Page(s):
    2047-2054

    Locally repairable codes, which can repair erased symbols from other symbols, have attracted a good deal of attention in recent years because its local repair property is effective on distributed storage systems. (ru, δu)u∈[s]-locally repairable codes with multiple localities, which are an extension of ordinary locally repairable codes, can repair δu-1 erased symbols simultaneously from a set consisting of at most ru symbols. An upper bound on the minimum distance of these codes and a construction method of optimal codes, attaining this bound with equality, were given by Chen, Hao, and Xia. In this paper, we discuss the parameter restrictions of the existing construction, and we propose explicit constructions of optimal codes with multiple localities with relaxed restrictions based on the encoding polynomial introduced by Tamo and Barg. The proposed construction can design a code whose minimum distance is unrealizable by the existing construction.

  • An Analysis of Time Domain Reed Solomon Decoder with FPGA Implementation

    Kentaro KATO  Somsak CHOOMCHUAY  

     
    PAPER-Computer System

      Pubricized:
    2017/08/23
      Vol:
    E100-D No:12
      Page(s):
    2953-2961

    This paper analyzes the time domain Reed Solomon Decoder with FPGA implementation. Data throughput and area is carefully evaluated compared with typical frequency domain Reed Solomon Decoder. In this analysis, three hardware architecture to enhance the data throughput, namely, the pipelined architecture, the parallel architecture, and the truncated arrays, is evaluated, too. The evaluation reveals that the number of the consumed resources of RS(255, 239) is about 20% smaller than those of the frequency domain decoder although data throughput is less than 10% of the frequency domain decoder. The number of the consumed resources of the pipelined architecture is 28% smaller than that of the parallel architecture when data throughput is same. It is because the pipeline architecture requires less extra logics than the parallel architecture. To get higher data throughput, the pipelined architecture is better than the parallel architecture from the viewpoint of consumed resources.

  • DCT-OFDM Watermarking Scheme Based on Communication System Model

    Minoru KURIBAYASHI  Shogo SHIGEMOTO  Nobuo FUNABIKI  

     
    PAPER-Spread Spectrum Technologies and Applications

      Vol:
    E100-A No:4
      Page(s):
    944-952

    In conventional spread spectrum (SS) watermarking schemes, random sequences are used for the modulation of watermark information. However, because of the mutual interference among those sequences, it requires complicated removal operation to improve the performance. In this paper, we propose an efficient spread spectrum watermarking scheme by introducing the orthogonal frequency divisiion multiplexing (OFDM) technique at the modulation of watermark information. The SS sequences in the proposed method are the DCT basic vectors modulated by a pseudo-random number (PN) sequence. We investigate the SS-based method considering the host interference at the blind detection scenario and analyze the noise caused by attacks. Because every operation is invertible, the quantization index modulation (QIM)-based method is applicable for the OFDM modulated signals. We also consider the property of watermark extracting operation in SS-based and QIM-based method and formalize their models of noisy channel in order to employ an error correcting code. The performance of their methods with error correcting code is numerically evaluated under the constraints of same distortion level in watermarked content. The experimental results indicated a criteria for the selection of SS-based and QIM-based methods for given content, which is determined by the amount of host interference. In case that the host interference is 0.8 times smaller than a watermark signal, the SS-based method is suitable. When it is 1.0 times larger, the QIM-based method should be selected.

  • A Comprehensive Model for Write Disturbance in Resistive Memory Composed of Cross-Point Array

    Yoshiaki ASAO  Fumio HORIGUCHI  

     
    PAPER-Integrated Electronics

      Vol:
    E100-C No:3
      Page(s):
    329-339

    A comprehensive model is presented for estimating the bit error rate (BER) of write disturbance in a resistive memory composed of a cross-point array. While writing a datum into the selected address, the non-selected addresses are biased by word-line (WL) and bit-line (BL). The stored datum in the non-selected addresses will be disturbed if the bias is large enough. It is necessary for the current flowing through the non-selected address to be calculated in order to estimate the BER of the write disturbance. Since it takes a long time to calculate the current flowing in a large-scale cross-point array, several simplified circuits have been utilized to decrease the calculating time. However, these simplified circuits are available to the selected address, not to the non-selected one. In this paper, new simplified circuits are proposed for calculating the current flowing through the non-selected address. The proposed and the conventional simplified circuits are used, and on that basis the trade-off between the write disturbance and the write error is discussed. Furthermore, the error correcting code (ECC) is introduced to improve the trade-off and to provide the low-cost memory chip matching current production lines.

  • An Error Correction Method for Neighborhood-Level Errors in NAND Flash Memories

    Shohei KOTAKI  Masato KITAKAMI  

     
    PAPER-Coding Theory

      Vol:
    E100-A No:2
      Page(s):
    653-662

    Rapid process scaling and the introduction of the multilevel cell (MLC) concept have lowered costs of NAND Flash memories, but also degraded reliability. For this reason, the memories are depending on strong error correcting codes (ECCs), and this has enabled the memories to be used in wide range of storage applications, including solid-state drives (SSDs). Meanwhile, too strong error correcting capability requires excessive decoding complexity and check bits. In NAND Flash memories, cell errors to neighborhood voltage levels are more probable than those to distant levels. Several ECCs reflecting this characteristics, including limited-magnitude ECCs which correct only errors with a certain limited magnitude and low-density parity check (LDPC) codes, have been proposed. However, as most of these ECCs need the multiple bits in a cell for encoding, they cannot be used with multipage programing, a high speed programming method currently employed in the memories. Also, binary ECCs with Gray codes are no longer optimal when multilevel voltage shifts (MVSs) occur. In this paper, an error correction method reflecting the error characteristic is presented. This method detects errors by a binary ECC as a conventional manner, but a nonbinary value or whole the bits in a cell, are subjected to error correction, so as to be corrected into the most probable neighborhood value. The amount of bit error rate (BER) improvement is depending on the probability of the each error magnitude. In case of 2bit/cell, if only errors of magnitude 1 and 2 can occur and the latter occupies 5% of cell errors, acceptable BER is improved by 4%. This is corresponding to extending 2.4% of endurance. This method needs about 15% longer average latency, 19% longer maximum latency, and 15% lower throughput. However, with using the conventional method until the memories' lifetime number of program/erase cycling, and the proposed method after that, BER improvement can be utilized for extending endurance without latency and throughput degradation until the switch of the methods.

  • Variation of SCM/NAND Flash Hybrid SSD Performance, Reliability and Cost by Using Different SSD Configurations and Error Correction Strengths

    Hirofumi TAKISHITA  Shuhei TANAKAMARU  Sheyang NING  Ken TAKEUCHI  

     
    PAPER

      Vol:
    E99-C No:4
      Page(s):
    444-451

    Storage-Class Memory (SCM) and NAND flash hybrid Solid-State Drive (SSD) has advantages of high performance and low power consumption compared with NAND flash only SSD. In this paper, first, three SSD configurations are investigated. Three different SCMs are used with 0.1 µs, 1 µs and 10 µs read/write latencies, respectively, and the required SCM/NAND flash capacity ratios are analyzed to maintain the same SSD performance. Next, by using the three SSD configurations, the variation of SSD reliability, performance and cost are analyzed by changing error correction strengths. The SSD reliability of acceptable SCM and NAND flash Bit Error Rates (BERs) is limited by achieving specified SSD performance with error correction, and/or limited by SCM and NAND flash parity size and SSD cost. Lastly, the SSD replacement cost is also analyzed by considering the limitation of NAND flash write/erase cycles. The purpose of this paper is to provide a design guideline for obtaining high performance, highly reliable and cost-effective SCM/NAND hybrid structure SSD with ECC.

  • Protection of On-chip Memory Systems against Multiple Cell Upsets Using Double-adjacent Error Correction Codes

    Hoyoon JUN  Yongsurk LEE  

     
    PAPER-Integrated Electronics

      Vol:
    E98-C No:3
      Page(s):
    267-274

    As semiconductor devices scale into deep sub-micron regime, the reliability issue due to radiation-induced soft errors increases in on-chip memory systems. Neutron-induced soft errors transiently upset adjacent information of multiple cells in these systems. Although single error correction and double error detection (SEC--DED) codes have been employed to protect on-chip memories from soft errors, they are not sufficient against multiple cell upsets (MCUs). SEC--DED and double adjacent error correction (SEC--DED--DAEC) codes have recently been proposed to address this problem. However, these codes do not the resolve mis-correction of double non-adjacent errors because syndromes for double non-adjacent errors are equal to that of double adjacent errors. The occurrence of this mis-correction in region of critical memory section such as operating systems may lead to system malfunction. To eliminate mis-correction, the syndrome spaces for double adjacent and double non-adjacent errors are not shared using the matrix with reversed colexicographic order. The proposed codes are implemented using hardware description language and synthesized using 32 nm technology library. The results show that there is no mis-correction in the proposed codes. In addition, the performance enhancement of the decoder is approximately 51.9% compared to double error correction codes for on-chip memories. The proposed SEC--DED--DAEC codes is suitable for protecting on-chip memory applications from MCUs-type soft errors.

  • Single Error Correcting Quantum Codes for the Amplitude Damping Channel Based on Classical Codes over GF(7)

    Keisuke KODAIRA  Mihoko WADA  Tomoharu SHIBUYA  

     
    PAPER-Coding Theory

      Vol:
    E97-A No:11
      Page(s):
    2247-2253

    The amplitude damping (AD) quantum channel is one of the models describing evolution of quantum states. The construction of quantum error correcting codes for the AD channel based on classical codes has been presented, and Shor et al. proposed a class of classical codes over F3 which are efficiently applicable to this construction. In this study, we expand Shor's construction to that over F7, and succeeded to construct an AD code that has better parameters than AD codes constructed by Shor et al.

  • A New Construction of Permutation Arrays

    Jung Youl PARK  Hong-Yeop SONG  

     
    PAPER-Sequences

      Vol:
    E95-A No:11
      Page(s):
    1855-1861

    Let PA(n, d) be a permutation array (PA) of order n and the minimum distance d. We propose a new construction of the permutation array PA(pm, pm-1k) for a given prime number p, a positive integer k < p and a positive integer m. The resulted array has (|PA(p,k)|p(m-1)(p-k))m rows. Compared to the other constructions, the new construction gives a permutation array of far bigger size with a large minimum distance, for example, when k ≥ 2p/3. Moreover the proposed construction provides an algorithm to find the i-th row of PA (pm, pm-1k) for a given index i very simply.

  • A Fault-Tolerant Architecture with Error Correcting Code for the Instruction-Level Temporal Redundancy

    Chao YAN  Hongjun DAI  Tianzhou CHEN  

     
    PAPER-Trust

      Vol:
    E95-D No:1
      Page(s):
    38-45

    Soft error has become an increasingly significant concern in modern micro-processor design, it is reported that the instruction-level temporal redundancy in out-of-order cores suffers an performance degradation up to 45%. In this work, we propose a fault tolerant architecture with fast error correcting codes (such as the two-dimensional code) based on double execution. Experimental results show that our scheme can gain back IPC loss between 9.1% and 10.2%, with an average around 9.2% compared with the conventional double execution architecture.

  • Reliable Multicast with Local Retransmission and FEC Using Group-Aided Multicast Scheme

    Alex FUNG  Iwao SASASE  

     
    PAPER-Fundamental Theories for Communications

      Vol:
    E92-B No:3
      Page(s):
    811-818

    In reliable multicast, feedback and recovery traffic limit the performance and scalability of the multicast session. In this paper, we present an improvement to the many-to-many reliable multicast protocol, Group-Aided Multicast protocol (GAM), with a local-group based recovery by making use of forward error correction (FEC) locally in addition to NACK/retransmission. In contrast to the original GAM, which only makes use of NACK-based recovery, our scheme produces FEC packets and multicasts the packets within the scope of a local group in order to correct uncorrelated errors of the local members in each group of the multicast session, which reduces the need for NACK/retransmission. By using our scheme, redundancy traffic can be localized in each group within a multicast session, and the overall recovery traffic can be reduced.

  • Modified Algorithm on Maximum Detected Bit Flipping Decoding for High Dimensional Parity-Check Code

    Yuuki FUNAHASHI  Shogo USAMI  Ichi TAKUMI  Masayasu HATA  

     
    LETTER-Coding Theory

      Vol:
    E89-A No:10
      Page(s):
    2670-2675

    We have researched high dimensional parity-check (HDPC) codes that give good performance over a channel that has a very high error rate. HDPC code has a little coding overhead because of its simple structure. It has hard-in, maximum detected bit flipping (MDBF) decoding that has reasonable decoding performance and computational cost. In this paper, we propose a modified algorithm for MDBF decoding and compare the proposed MDBF decoding with conventional hard-in decoding.

  • Naive Mean Field Approximation for Sourlas Error Correcting Code

    Masami TAKATA  Hayaru SHOUNO  Masato OKADA  

     
    PAPER-Biocybernetics, Neurocomputing

      Vol:
    E89-D No:8
      Page(s):
    2439-2447

    Solving the error correcting code is an important goal with regard to communication theory. To reveal the error correcting code characteristics, several researchers have applied a statistical-mechanical approach to this problem. In our research, we have treated the error correcting code as a Bayes inference framework. Carrying out the inference in practice, we have applied the NMF (naive mean field) approximation to the MPM (maximizer of the posterior marginals) inference, which is a kind of Bayes inference. In the field of artificial neural networks, this approximation is used to reduce computational cost through the substitution of stochastic binary units with the deterministic continuous value units. However, few reports have quantitatively described the performance of this approximation. Therefore, we have analyzed the approximation performance from a theoretical viewpoint, and have compared our results with the computer simulation.

  • An Adaptive FEC Scheme for Firm Real-Time Multimedia Communications in Wireless Networks

    Kyong Hoon KIM  Jong KIM  Sung Je HONG  

     
    PAPER

      Vol:
    E88-B No:7
      Page(s):
    2794-2801

    The technological development of wireless environment has made real-time multimedia communications possible in wireless networks. Many studies have been done on real-time communications in wireless networks in order to overcome a higher bit error rate in wireless channels. However, none of work deals with firm real-time communications which can be applied to multimedia communications. In this paper, we propose an adaptive error correcting scheme for firm real-time multimedia communications in wireless networks in order to maximize the expected net profit. The proposed scheme adaptively selects an error correcting code under the current air state and the message state of a message stream. Throughout simulation results, we show that the suggested scheme provides more profit than single error-correcting code schemes.

  • Systematic Binary Deletion/Insertion Error Correcting Codes Capable of Correcting Random Bit Errors

    Kiattichai SAOWAPA  Haruhiko KANEKO  Eiji FUJIWARA  

     
    PAPER-Coding Theory

      Vol:
    E83-A No:12
      Page(s):
    2699-2705

    This paper presents a class of binary block codes capable of correcting single synchronization errors and single reversal errors with fewer check bits than the existing codes by 3 bits. This also shows a decoding circuit and analyzes its complexity.

  • A Code Construction for M-Choose-T Communication over the Multiple-Access Adder Channel

    Kin-ichiroh TOKIWA  Hiroshi MATSUDA  Hatsukazu TANAKA  

     
    PAPER-Information Theory and Coding Theory

      Vol:
    E78-A No:1
      Page(s):
    94-99

    Coding scheme is discussed for M-Choose-T communication in which at most T active users out of M potential users simultaneously transmit their messages over a common channel. The multiple-access channel considered in this paper is assumed to be a time-discrete noiseless adder channel without feedback with T binary inputs and one real-valued output, and is used on the assumption of perfect block and bit synchronization among users. In this paper a new class of uniquely decodable codes is proposed in order to realize error-free M-Choose-T communication over the adder channel described above. These codes are uniquely decodable in the sense that not only the set of active users can be specified but also their transmitted messages can be recovered uniquely as long as T or fewer users are active simultaneously. It is shown that these codes have a simple decoding algorithm and can achieve a very high sum rate arbitrarily close to unity if exactly T users are active.

  • Variable Error Controlling Schemes for Intelligent Error Controlling Systems

    Taroh SASAKI  Ryuji KOHNO  Hideki IMAI  

     
    PAPER

      Vol:
    E77-A No:8
      Page(s):
    1281-1288

    Recently, a lot of research works have been carried out regarding intelligent communication. If the final information sink is assumed as a human being, a communication channel can be used more effectively when encoders/decoders work "intelligently" or take into account of the semantics of information to be sent. We have been studying error-controlling systems based on different importance of segmental information. The system divides the information input into segments to which individual importance can be assigned. The segments are individually encoded by appropriate error-correcting codes (ECCs) which correspond to their importance among codes with different error-correcting capabilities. For the information that difference of the importance is systematically aligned, conventional UEP (unequal error protection) codes can be applied, but we treat the case that alignment of the importance of the information source is not systematically aligned. Since the system uses multiple ECCs with different (n,k,d) parameters, information regarding what length of the next codeword is required for decoding. We propose error controlling schemes using mulriple ECCs; the first scheme and the second scheme use the obvious codelength identifying information. In the second scheme, information bits are sorted so that segments with the same importance can be encoded by an ECC with the same error-correcting capability. The third scheme is a main proposal in this paper and uses Variable Capability Coding scheme (VCC) which uses some ECCs having different error-correcting capabilities and codelengths. A sequence encoded by the VCC is separable into appropriate segments without obvious codelength identifying information when the channel error probability is low. Subsequently, we evaluate these schemes by coderate when (1) error correcting capability (2) codelength identifying capability are the same. One of the feature of VCC is the capability of resuming from propagative errors because errors beyond the codelength identifying capability occur and the proper beginning of the codeword is lost in the decoder. We also evaluate this capability as (3) resynchronizing capability.