1-7hit |
The optimal coding strategy for signal detection in the correlated gaussian noise is established for the distributed sensors system with essentially zero transmission rate constraint. Specifically, we are able to obtain the same performance as in the situation of no restriction on rate from each sensor terminal to the fusion center. This simple result contrasts with the previous ad hoc studies containing many unnatural assumptions such as the independence of noises contaminating received signal at each sensor. For the design of optimal coder, we can use the classical Levinson-Wiggins-Robinson fast algorithm for block Toeplitz matrix to evaluate the necessary weight vector for the maximum-likelihood detection.
Kazuhiko YAMAGUCHI Toshiaki WATANABE Kingo KOBAYASHI
In this paper, we study unequal error protection (UEP) capabilities of punctured convolutional codes. For constructing the good UEP convolutional codes, the conditional weight distributions of UEP convolutional codes are defined and evaluated. The conditional weight distributions are computed by using the transfer functions of time-varying trellis structures of punctured convolutional codes. The best UEP convolutional codes from the viewpoint of the weight distributions are listed.
Hyunho KANG Koutarou YAMAGUCHI Brian KURKOSKI Kazuhiko YAMAGUCHI Kingo KOBAYASHI
For the digital watermarking patchwork algorithm originally given by Bender et al., this paper proposes two improvements applicable to audio watermarking. First, the watermark embedding strength is psychoacoustically adapted, using the Bark frequency scale. Second, whereas previous approaches leave the samples that do not correspond to the data untouched, in this paper, these are modified to reduce the probability of misdetection, a method called full index embedding. In simulations, the proposed combination of these two proposed methods has higher resistance to a variety of attacks than prior algorithms.
One of the basic problems in Information Theory, that is, the determination of the reliability function of binary symmetric channel, is studied by establishing the exponent of cardinality of intersection of two Hamming spheres.
Kazuyoshi HARADA Kingo KOBAYASHI
We study some sufficient conditions of codeword lengths for the existence of a fix-free code. Ahlswede et al. proposed the 3/4 conjecture that Σi=1n a-li 3/4 implies the existence of a fix-free code with lengths li when a=2 i. e. the alphabet is binary. We propose a more general conjecture, and prove that the upper bound of our conjecture is not greater than 3/4 for any finite alphabet. Moreover, we show that for any a2 our conjecture is true if codeword lengths l1,l2,. . . consist of only two kinds of lengths.
Masazumi KURIHARA Shojiro SAKATA Kingo KOBAYASHI
In this paper we propose a class of byte-error-correcting codes derived from algebraic curves which is a generalization on the Reed-Solomon codes, and present their fast parallel decoding algorithm. Our algorithm can correct up to (m + b -θ)/2b byte-errors for the byte length b, where m + b -θ + 1dG for the Goppa designed distance dG. This decoding algorithm can be parallelized. In this algorithm, for our code over the finite field GF (q), the total complexity for finding byte-error locations is O (bt(t + q - 1)) with time complexity O (t(t + q - 1)) and space complexity O(b), and the total complexity for finding error values is O (bt(b + q - 1)) with time complexity O (b(b + q - 1)) and space complexity O (t), where t(m + b -θ)/2b. Our byte-error-correcting algorithm is superior to the conventional fast decoding algorithm for randomerrors in regard to the number of correcting byte-errors in several cases.
Hiroyoshi MORITA Kingo KOBAYASHI
A method for the compression of ECG data is presented. The method is based on the edit distance algorithm developed in the file comparison problems. The edit distance between two sequences of symbols is defined as the number of edit operations required to transform a sequence of symbols into the other. We adopt the edit distance algorithm to obtain a list of edit operations, called edit script, which transforms a reference pulse into a pulse selected from ECG data. If the decoder knows the same reference, it can reproduce the original pulse, only from the edit script. The amount of the edit script is expected to be smaller than that of the original pulse when the two pulses look alike and thereby we can reduce the amount of space to store the data. Applying the proposed scheme to the raw data of ECG, we have achieved a high compression about 14: 1 without losing the significant features of signals.