The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] error-correcting(40hit)

21-40hit(40hit)

  • Compound-Error-Correcting Codes and Their Augmentation

    Masaya FUJISAWA  Shusuke MAEDA  Shojiro SAKATA  

     
    PAPER-Coding Theory

      Vol:
    E86-A No:7
      Page(s):
    1813-1819

    A compound error is any combination of burst errors with various burst lengths including random errors. The compound weight of any such error is defined as a kind of combinational metric which is a generalization of Gabidulin's metric. First, we present a fast method for calculating the weight of any word. Based on this method, as an extension of Wadayama's augmenting method in the case of Hamming weight, we propose a method of constructing codes having higher coding rate by augmenting any compound-error-correcting codes. Furthermore, we show some examples of good compound-error-correcting codes obtained by using our augmenting method.

  • Existence Condition for Tail-Biting Convolutional Codes

    Young KIM  Pil Joong LEE  

     
    PAPER-Fundamental Theories

      Vol:
    E85-B No:11
      Page(s):
    2362-2368

    We investigated the truncated convolutional code with the characteristics of a block code for block-based communication systems. Three truncation methods (direct truncation, tail-terminating, and tail-biting method) were introduced by other researchers. Each of the three methods has a weakness: the direct truncation method decreases the minimum distance, the tail-terminating method uses tail bits, and the tail-biting method can only be applied by using a complicated decoder. Although the tail-biting method gives a better BER performance than the other two methods, we cannot apply the tail-biting method in all situations. Occasionally, the tail-biting convolutional code does not exist. Wang et al. presented two necessary conditions for the existence of the tail-biting convolutional code of the rate-1/2 recursive systematic convolutional code. In this paper, we analyze the encoder of the convolutional code as a linear time invariant system, and present two theorems and six corollaries on the existence of the tail-biting convolutional code. These existence conditions are adaptable to all convolutional codes. In the communication system using the truncated convolutional code, these results are applicable to determining the truncation method.

  • A Novel Turbo-TCM Scheme Based on Concatenated Tree Codes

    Baoming BAI  Kin Shing HO  Li PING  

     
    LETTER-Fundamental Theories

      Vol:
    E85-B No:9
      Page(s):
    1835-1837

    In this letter, we introduce a two-state turbo-TCM scheme based on the concatenated tree codes. The proposed scheme can achieve near capacity performance yet has considerably lower decoding complexity compared with other existing turbo-TCM codes.

  • Near-Optimality of Subcodes of Hamming Codes on the Two-State Markovian Additive Channel

    Mitsuru HAMADA  

     
    PAPER-Coding Theory

      Vol:
    E84-A No:10
      Page(s):
    2383-2388

    Near-optimality of subcodes of the cyclic Hamming codes is demonstrated on the binary additive channel whose noise process is the two-state homogeneous Markov chain, which is a model of bursty communication channels.

  • A Search Algorithm for Bases of Calderbank-Shor-Steane Type Quantum Error-Correcting Codes

    Kin-ichiroh TOKIWA  Hatsukazu TANAKA  

     
    PAPER-Coding Theory

      Vol:
    E84-A No:3
      Page(s):
    860-865

    Recently, Vatan, Roychowdhury and Anantram have presented two types of revised versions of the Calderbank-Shor-Steane code construction, and have also provided an exhaustive procedure for determining bases of quantum error-correcting codes. In this paper, we investigate the revised versions given by Vatan et al., and point out that there is no essential difference between them. In addition, we propose an efficient algorithm for searching for bases of quantum error-correcting codes. The proposed algorithm is based on some fundamental properties of classical linear codes, and has much lower complexity than Vatan et al.'s procedure.

  • A Sufficient Condition for a Code to Achieve the Minimum Decoding Error Probability--Generalization of Perfect and Quasi-Perfect Codes

    Mitsuru HAMADA  

     
    PAPER-Coding Theory

      Vol:
    E83-A No:10
      Page(s):
    1870-1877

    A sufficient condition for a code to be optimum on discrete channels with finite input and output alphabets is given, where being optimum means achieving the minimum decoding error probability. This condition is derived by generalizing the ideas of binary perfect and quasi-perfect codes, which are known to be optimum on the binary symmetric channel. An application of the sufficient condition shows that the code presented by Hamada and Fujiwara (1997) is optimum on the q-ary channel model proposed by Fuja and Heegard (1990), where q is a prime power with some restriction. The channel model is subject to two types of additive errors of (in general) different probabilities.

  • Almost Sure Convergence of Relative Frequency of Occurrence of Burst Errors on Channels with Memory

    Mitsuru HAMADA  

     
    PAPER-Coding Theory

      Vol:
    E82-A No:10
      Page(s):
    2022-2033

    Motivated by intention to evaluate asymptotically multiple-burst-error-correcting codes on channels with memory, we will derive the following fact. Let {Zi } be a hidden Markov process, i. e. , a functional of a Markov chain with a finite state space, and Wb(Z1Z2Zn) denote the number of burst errors that appear in Z1Z2Zn, where the number of burst errors is counted using Gabidulin's burst metric , 1971. As the main result, we will prove the almost sure convergence of relative burst weight Wb(Z1Z2Zn)/n, i. e. , the relative frequency of occurrence of burst errors, for a broad class of functionals { Zi } of finite Markov chains. Functionals of Markov chains are often adopted as models of the noises on channels, especially on burst-noise channels, the most famous model of which is probably the Gilbert channel proposed in 1960. Those channel models described with Markov chains are called channels with memory (including channels with zero-memory, i. e. , memoryless ones). This work's achievement enables us to extend Gilbert's code performance evaluation in 1952, a landmark that offered the well-known Gilbert bound, discussed its relationship to the (memoryless) binary symmetric channel, and has been serving as a guide for the-Hamming-metric-based design of error-correcting codes, to the case of the-burst-metric-based codes (burst-error-correcting codes) and discrete channels with or without memory.

  • A Study on Performances of Soft-Decision Decoding Algorithm Based on Energy Minimization Principle

    Akira SHIOZAKI  Yasushi NOGAWA  Tomokazu SATO  

     
    LETTER-Coding Theory

      Vol:
    E82-A No:10
      Page(s):
    2194-2198

    We proposed a soft-decision decoding algorithm for cyclic codes based on energy minimization principle. This letter presents the algorithm which improves decoding performance and decoding complexity of the previous method by giving more initial positions and introducing a new criterion for terminating the decoding procedure. Computer simulation results show that both the decoded block error rate and the decoding complexity decrease by this method more than by the previous method.

  • A Mixed Upper Bound on the Maximum Size of Codes for Multiple Burst Error Correction and Detection

    Mitsuru HAMADA  

     
    PAPER-Coding Theory

      Vol:
    E81-A No:10
      Page(s):
    1964-1971

    We derive an upper bound on the size of a block code with prescribed burst-error-correcting capability combining those two ideas underlying the generalized Singleton and sphere-packing bounds. The two ideas are puncturing and sphere-packing. We use the burst metric defined by Gabidulin, which is suitable for burst error correction and detection. It is demonstrated that the proposed bound improves previously known ones for finite code-length, when minimum distance is greater than 3, as well as in the asymptotic forms.

  • On Reducing Complexity of a Soft-Decision Decoding Algorithm for Cyclic Codes Based on Energy Minimization Principle

    Akira SHIOZAKI  Kazutaka AOKI  

     
    PAPER-Coding Theory

      Vol:
    E81-A No:10
      Page(s):
    1998-2004

    We propose a novel soft-decision decoding algorithm for cyclic codes based on energy minimization principle. The well-known soft-decision decoding algorithms for block codes perform algebraic (hard-decision) decoding several times in order to generate candidate codewords using the reliability of received symbols. In contrast, the proposed method defines energy as the Euclidean distance between the received signal and a codeword and alters the values of information symbols so as to decrease the energy in order to seek the codeword of minimum energy, which is the most likely codeword. We let initial positions be the information parts of signals obtained by cyclically shifting a received signal and look for the point, which represents a codeword, of minimum energy by moving each point from several initial positions. This paper presents and investigates reducing complexity of the soft-decision decoding algorithm. We rank initial positions in order of reliability and reduce the number of initial positions in decoding. Computer simulation results show that this method reduces decoding complexity.

  • A New Class of Single Error-Correcting Fixed Block-Length (d, k) Codes

    Hatsukazu TANAKA  

     
    PAPER-Coding Theory

      Vol:
    E80-A No:11
      Page(s):
    2052-2057

    In this paper a new class of single error-correcting fixed block-length (d, k) codes has been proposed. The correctable error types are peak-shift error, insertion or deletion error, symmetric error, etc. The basic technique to construct codes is a systematic construction algorithm of multilevel sequences with a constant Lee weight (TALG algorithm). The coding rate and efficiency are considerably good, and hence the proposed new codes will be very useful for improving the reliability of high density magnetic recording.

  • On a Class of Byte-Error-Correcting Codes from Algebraic Curves and Their Fast Decoding Algorithm

    Masazumi KURIHARA  Shojiro SAKATA  Kingo KOBAYASHI  

     
    PAPER-Coding Theory

      Vol:
    E79-A No:9
      Page(s):
    1298-1304

    In this paper we propose a class of byte-error-correcting codes derived from algebraic curves which is a generalization on the Reed-Solomon codes, and present their fast parallel decoding algorithm. Our algorithm can correct up to (m + b -θ)/2b byte-errors for the byte length b, where m + b -θ + 1dG for the Goppa designed distance dG. This decoding algorithm can be parallelized. In this algorithm, for our code over the finite field GF (q), the total complexity for finding byte-error locations is O (bt(t + q - 1)) with time complexity O (t(t + q - 1)) and space complexity O(b), and the total complexity for finding error values is O (bt(b + q - 1)) with time complexity O (b(b + q - 1)) and space complexity O (t), where t(m + b -θ)/2b. Our byte-error-correcting algorithm is superior to the conventional fast decoding algorithm for randomerrors in regard to the number of correcting byte-errors in several cases.

  • Fault-Tolerant Graphs for Hypercubes and Tori*

    Toshinori YAMADA  Koji YAMAMOTO  Shuichi UENO  

     
    PAPER-Fault Diagnosis/Tolerance

      Vol:
    E79-D No:8
      Page(s):
    1147-1152

    Motivated by the design of fault-tolerant multiprocessor interconnection networks, this paper considers the following problem: Given a positive integer t and a graph H, construct a graph G from H by adding a minimum number Δ(t, H) of edges such that even after deleting any t edges from G the remaining graph contains H as a subgraph. We estimate Δ(t, H) for the hypercube and torus, which are well-known as important interconnection networks for multiprocessor systems. If we denote the hypercube and the square torus on N vertices by QN and DN respectively, we show, among others, that Δ(t, QN) = O(tN log(log N/t + log 2e)) for any t and N (t 2), and Δ(1, DN) = N/2 for N even.

  • Analysis of Aliasing Probability for MISRs by Using Complete Weight Distributions

    Kazuhiko IWASAKI  Sandeep K. GUPTA  Prawat NAGVAJARA  Tadao KASAMI  

     
    PAPER

      Vol:
    E78-A No:12
      Page(s):
    1691-1698

    The aliasing probability was analyzed for MISRs when the error probability for each input was different. A closed form expression was derived by applying the complete weight distributions of linear codes over a Galois field and its dual codes. The aliasing probability for MISRs characterized by non-primitive polynomials was also analyzed. The inner product for binary representation of symbols was used instead of multiplication over a Galois field. The results show the perfect expression for analyzing the aliasing probability of MISRs.

  • An Error-Correcting Version of the Leiss's Parser for Context-Free Languages

    Ken-ichi KURODA  Eiichi TANAKA  

     
    LETTER-Automaton, Language and Theory of Computing

      Vol:
    E76-D No:12
      Page(s):
    1528-1531

    This paper describes an error-correcting parser (ec-parser) for context-free languages that is an extension of the Leiss's parser. Since the ec-parser uses precomputed informations and a pruning technique by lookahead, the ec-parser is always faster than the Lyon's parser. Several examples are shown.

  • Soft-Decision Decoding Algorithm for Binary Linear Block Codes

    Yong Geol SHIM  Choong Woong LEE  

     
    PAPER-Information Theory and Coding Theory

      Vol:
    E76-A No:11
      Page(s):
    2016-2021

    A soft-decision decoding algorithm for binary linear block codes is proposed. This algorithm seeks to minimize the block error probability. With careful examinations of the first hard-decision decoded results, the candidate codewords are efficiently searched for. Thus, we can reduce the decoding complexity (the number of hard-decision decodings) and lower the block error probability. Computer simulation results are presented for the (23, 12) Golay code. They show that the decoding complexity is considerably reduced and the block error probability is close to that of the maximum likelihood decoder.

  • Theory and Techniques for Testing Check Bits of RAMs with On-Chip ECC

    Manoj FRANKLIN  Kewal K. SALUJA  

     
    PAPER-Fault Tolerant Computing

      Vol:
    E76-D No:10
      Page(s):
    1243-1252

    As RAMs become dense, their reliability reduces because of complex interactions between memory cells and soft errors due to alpha particle radiations. In order to rectify this problem, RAM manufacturers have started incorporating on-chip (built-in) ECC. In order to minimize the area overhead of on-chip ECC, the same technology is used for implementing the check bits and the information bits. Thus the check bits are exposed to the same failure modes as the information bits. Furthermore, faults in the check bits will manifest as uncorrectable multiple errors when a soft error occurs. Therefore it is important to test the check bits for all failure modes expected of other cells. In this paper, we formulate the problem of testing RAMs with on-chip ECC capability. We than derive necessary and sufficient conditions for testing the check bits for arbitrary and adjacent neighborhood pattern sensitive faults. We also provide an efficient solution to test a memory array of N bits (including check bits) for 5-cell neighborhood pattern sensitive faults in O (N) reads and writes, with the check bits also tested for the same fault classes as the information bits.

  • Multihopping and Decoding of Error-Correcting Code for MFSK/FH-SSMA Systems

    Tetsuo MABUCHI  Ryuji KOHNO  Hideki IMAI  

     
    PAPER

      Vol:
    E76-B No:8
      Page(s):
    874-885

    This paper investigates a multihopping scheme for MFSK (Multilevel Frequency Shift Keying) /FH-SSMA (Frequency Hopping-Spread Spectrum Multiple Access) system. Moreover, we propose and investigate a modified decoding scheme for the coded MFSK/FH-SSMA system. In this multi-hopped MFSK/FH-SSMA system, several hopping frequencies per chip are assigned and transmitted in parallel in order to improve its frequency diversity capability for a fading channel. We theoretically analyze the performance of the multihopped MFSK/FH-SSMA system in a Rayleigh fading channel. Moreover, in the coded MFSK/FH-SSMA system, we propose a modified scheme of the error and erasure decoding of an error-correcting code. The modified decoding scheme utilizes the information of rows having the largest number of entries in the decoded time-frequency matrix. Their BER (Bit Error Rate) performance is evaluated by theoretical analysis in order to show the improvement in user capacity.

  • An Error-Controlling Scheme Based on Different Importance of Segments of a Natural Language

    Taroh SASAKI  Ryuji KOHNO  Hideki IMAI  

     
    PAPER

      Vol:
    E75-A No:9
      Page(s):
    1076-1086

    Although individual segments of a natural language such as words have different importance on human interpretation of the meaning, every segment has been uniformly protected by an error-correcting code. If the importance of individual segments is defined by considering their meaning in the sentence, we can adaptively control the level of error-protection for each segment according to its importance in order to reduce errors on human interpretation of the meaning. In this paper, we propose an error-control scheme based on the varying importance of each word. We first introduce a method which determines the importance of each word and then propose an error-control scheme in which several error-correcting codes are alternately used to protect each word according to its importance. Probablity of semantic errors, that is, errors on human interpretation of the meaning, is defined and used as a criterion in mapping error-correcting codes to words possessing different importance. We theoretically formalize the problem of obtaining an optimum mapping which minimizes the probability of semantic errors under some constraint. Given a certain probability distribution of the importance of words and set of error-correcting codes, we can derive the optimum mapping. The proposed error-controlling scheme is theoretically evaluated by comparing its probability of semantic errors with that of a conventional scheme in which every word is uniformly protected by a single error-correcting code. Results show that the proposed scheme can considerably raduce the probability of semantic errors while retaining the same average transmission rate or redundancy.

  • Runlength-Limited Codes which Turn Peak-Shift Errors into Unidirectional Byte Errors

    Yuichi SAITOH  Hideki IMAI  

     
    LETTER

      Vol:
    E75-A No:7
      Page(s):
    898-900

    In this letter, we consider a magnetic or optical recording system employing a concatenated code that consists of a runlength-limited (d, k) block code as an inner code and a byte-error-correcting code as an outer code. (d, k) means that any two consecutive ones in the code bit stream are separated by at least d zeros and by at most k zeros. The minimum separation d and the maximum separation k are imposed in order to reduce intersymbol interference and extract clock control from the received bit stream, respectively. This letter recommends to use as the outer code a unidirectional-byte-error-correcting code instead of an ordinary byte-error-correcting code. If we devise the mapping of the code symbols of the outer code onto the codewords of the inner code, we may improve the error performance. Examples of the mappings are described.

21-40hit(40hit)