In this paper, a joint blind synchronization and demodulation scheme is developed for ultra-wideband (UWB) impulse radio systems. Based on the prior knowledge of the direct-sequence (DS) spread codes, the proposed approach can achieve frame-level synchronization with the help of frame-rate samples. Taking advantage of the periodicity of the DS spread codes, the frame-level synchronization can be carried out even in one symbol interval. On the other hand, after timing acquisition, these frame-rate samples can be re-utilized also for demodulation. Thus the acquisition time and the implementation complexity are reduced considerably. The performance improvement can be justified by both theoretical analysis and simulation results, in terms of acquisition probability and bit error rate (BER).
Akira SHIOZAKI Masashi KISHIMOTO Genmon MARUOKA
This letter proposes extended single parity check product codes and presents their empirical performances on a Gaussian channel by belief propagation (BP) decoding algorithm. The simulation results show that the codes can achieve close-to-capacity performance in high coding rate. The code of length 9603 and of rate 0.96 is only 0.77 dB away from the Shannon limit for a BER of 10-5.
Hristo KOSTADINOV Hiroyoshi MORITA Noboru IIJIMA A. J. HAN VINCK Nikolai MANEV
Integer codes are very flexible and can be applied in different modulation schemes. A soft decoding algorithm for integer codes will be introduced. Comparison of symbol error probability (SEP) versus signal-to-noise ratio (SNR) between soft and hard decoding using integer coded modulation shows us that we can obtain at least 2 dB coding gain. Also, we shall compare our results with trellis coded modulation (TCM) because of their similar decoding schemes and complexity.
This paper presents a high-speed, low-complexity VLSI architecture based on the modified Euclidean (ME) algorithm for Reed-Solomon decoders. The low-complexity feature of the proposed architecture is obtained by reformulating the error locator and error evaluator polynomials to remove redundant information in the ME algorithm proposed by Truong. This increases the hardware utilization of the processing elements used to solve the key equation and reduces hardware by 30.4%. The proposed architecture retains the high-speed feature of Truong's ME algorithm with a reduced latency, achieved by changing the initial settings of the design. Analytical results show that the proposed architecture has the smallest critical path delay, latency, and area-time complexity in comparison with similar studies. An example RS(255,239) decoder design, implemented using the TSMC 0.18 µm process, can reach a throughput rate of 3 Gbps at an operating frequency of 375 MHz and with a total gate count of 27,271.
Yoshifumi UKITA Tomohiko SAITO Toshiyasu MATSUSHIMA Shigeichi HIRASAWA
In digital signal processing, the sampling theorem states that any real valued function f can be reconstructed from a sequence of values of f that are discretely sampled with a frequency at least twice as high as the maximum frequency of the spectrum of f. This theorem can also be applied to functions over finite domain. Then, the range of frequencies of f can be expressed in more detail by using a bounded set instead of the maximum frequency. A function whose range of frequencies is confined to a bounded set is referred to as bandlimited function. And a sampling theorem for bandlimited functions over Boolean domain has been obtained. Here, it is important to obtain a sampling theorem for bandlimited functions not only over Boolean domain (GF(2)n domain) but also over GF(q)n domain, where q is a prime power and GF(q) is Galois field of order q. For example, in experimental designs, although the model can be expressed as a linear combination of the Fourier basis functions and the levels of each factor can be represented by GF(q), the number of levels often take a value greater than two. However, the sampling theorem for bandlimited functions over GF(q)n domain has not been obtained. On the other hand, the sampling points are closely related to the codewords of a linear code. However, the relation between the parity check matrix of a linear code and any distinct error vectors has not been obtained, although it is necessary for understanding the meaning of the sampling theorem for bandlimited functions. In this paper, we generalize the sampling theorem for bandlimited functions over Boolean domain to a sampling theorem for bandlimited functions over GF(q)n domain. We also present a theorem for the relation between the parity check matrix of a linear code and any distinct error vectors. Lastly, we clarify the relation between the sampling theorem for functions over GF(q)n domain and linear codes.
Hao WANG Shi CHEN Xiaokang LIN
The bit-error-rate (BER) performance predicted by the semi-analytical evolution technique proposed by Li Ping et al. becomes inaccurate for parallel concatenated coded interleave-division multiple-access (PCC-IDMA) systems. To solve this problem, we develop a novel evolution technique of such systems. Numerical results show that the predicted performance agrees well with the simulation results, and that this technique is useful for system optimization.
Sangjoon PARK Sooyong CHOI Seung-Hoon HWANG
A continuous belief propagation (BP) decoding algorithm for a hybrid automatic repeat request (ARQ) system is proposed in this paper. The proposed continuous BP decoding algorithm utilizes the extrinsic information generated in the last iteration of the previous transmission for a continuous progression of the decoding through retransmissions. This allows the continuous BP decoding algorithm to accelerate the decoding convergence for codeword determination, especially when the number of retransmissions is large or a currently combined packet has punctured nodes. Simulation results verify the effectiveness of the proposed continuous BP decoding algorithm.
Haruka SUZUKI Marco HERNANDEZ Ryuji KOHNO
This paper presents hybrid type-II automatic repeat request (H-ARQ) for wireless wearable body area networks (BANs) based on ultra wideband (UWB) technology. The proposed model is based on three schemes, namely, high rate optimized rate compatible punctured convolutional codes (HRO-RCPC), Reed Solomon (RS) invertible codes and their concatenation. Forward error correction (FEC) coding is combined with simple cyclic redundancy check (CRC) error detection. The performance is investigated for two channels: CM3 (on-body to on-body) and CM4 (on-body to a gateway) scenarios of the IEEE802.15.6 BAN channel models for BANs. It is shown that the improvement in performance in terms of throughput and error protection robustness is very significant. Thus, the proposed H-ARQ schemes can be employed and optimized to suit medical and non-medical applications. In particular we propose the use of FEC coding for non-medical applications as those require less stringent quality of service (QoS), while the incremental redundancy and ARQ configuration is utilized only for medical applications. Thus, higher QoS is guaranteed for medical application of BANs while allowing coexistence with non-medical applications.
Xueqin JIANG Moon Ho LEE Tae Chol SHIN
This letter presents an approach to the construction of multiple-rate quasi-cyclic (QC) low-density parity-check (LDPC) codes based on hyperplanes (µ-flats) of two different dimensions in Euclidean geometries. The codes constructed with this method have the same code length, multiple-rate and large stopping sets while maintaining the same basic hardware architecture. The code performance is investigated in terms of the bit error rate (BER) and compared with those of the LDPC codes which are proposed in IEEE 802.16e standard. Simulation results show that our codes perform very well and have low error floors over the AWGN channel.
Shojiro SAKATA Masaya FUJISAWA
It is a well-known fact that the BMS algorithm with majority voting can decode up to half the Feng-Rao designed distance dFR. Since dFR is not smaller than the Goppa designed distance dG, that algorithm can correct up to errors. On the other hand, it has been considered to be evident that the original BMS algorithm (without voting) can correct up to errors similarly to the basic algorithm by Skorobogatov-Vladut. But, is it true? In this short paper, we show that it is true, although we need a few remarks and some additional procedures for determining the Groebner basis of the error locator ideal exactly. In fact, as the basic algorithm gives a set of polynomials whose zero set contains the error locators as a subset, it cannot always give the exact error locators, unless the syndrome equation is solved to find the error values in addition.
A (k,δ,ε)-locally decodable code C:Fqn FqN is an error-correcting code that encodes
In order to reduce the iterative decoding delay of convolutional turbo codes, this paper presents a concurrent decoding algorithm for the hardware implementation of turbo convolutional decoders. Different than a general turbo code, the hardware turbo decoder based on the proposed algorithm can update the priori information of message for each component code in a bit-by-bit manner as soon as it is generated by the other component code. The two component codes in a turbo code can thus be decoded concurrently, by using a single MAP decoder, subsequently reducing the decoding latency by approximately half while maintaining the bit error rate performance and a comparable hardware complexity, as a general turbo decoder.
LeThanh HA Chun-Su PARK Seung-Won JUNG Sung-Jea KO
Context-based Adaptive Binary Arithmetic Coding (CA-BAC) is adopted as an entropy coding tool for main profile of the video coding standard H.264/AVC. CABAC achieves higher degree of redundancy reduction by estimating the conditional probability of each binary symbol which is the input to the arithmetic coder. This paper presents an entropy coding method based on CABAC. In the proposed method, the binary symbol is coded using more precisely estimated conditional probability, thereby leading to performance improvement. We apply our method to the standard and evaluate its performance for different video sources and various quantization parameters (QP). Experiment results show that our method outperforms the original CABAC in term of coding efficiency, and the average bit-rate savings are up to 1.2%.
Manabu HAGIWARA Marc P.C. FOSSORIER Takashi KITAGAWA Hideki IMAI
In this paper, we investigate the smallest value of p for which a (J,L,p)-QC LDPC code with girth 6 exists for J=3 and J=4. For J=3, we determine the smallest value of p for any L. For J=4, we determine the smallest value of p for L ≤ 301. Furthermore we provide examples of specific constructions meeting these smallest values of p.
Srijidtra MAHAPAKULCHAI Chalie CHAROENLARPNOPPARUT
In the modern day, MPEG-4 image compression technique have been commonly applied in various indoor wireless communication systems. The efficient system design mostly relies on the joint source channel coding algorithms, which aim to reduce the complexity of channel coding process, while maintaining the quality of the receiving images. In this paper, we design the MAP source-controlled channel decoders with both random and semirandom interleavers for Rician slow flat block-fading channels. The MAP-Viterbi decoder employs the residual redundancy from zerotree symbol sequences of MPEG-4 HFS packets. The interleaving processes are done after the overall channel coding process to combat the block-fading effects. The computer simulations summarize the system performance in terms of average WER and PSNR (dB). With the interleavers, the significant improvement in PSNR of about 15-17 dB is obtained for both ML and MAP decoding. Also in many cases, we obtain more improvement of about 0.2-0.4 dB for using MAP decoding with the interleavers.
Yutaka MURAKAMI Shutai OKAMURA Shozo OKASAKA Takaaki KISHIGAMI Masayuki ORIHASHI
We newly design time-varying low-density parity-check convolutional codes (LDPC-CCs) based on parity check polynomials of the convolutional codes with a time period of 3, and show that BER (Bit Error Rate) performance in the time-varying LDPC-CCs with a time period of 3 is better than that in the conventional time-varying LDPC-CCs with a time period of 2 in the same coding rate with the nearly equal constraint length.
Yoshihide TONOMURA Takayuki NAKACHI Tatsuya FUJII Hitoshi KIYA
This paper proposes a parallelized DVC framework that treats each bitplane independently to reduce the decoding time. Unfortunately, simple parallelization generates inaccurate bit probabilities because additional side information is not available for the decoding of subsequent bitplanes, which degrades encoding efficiency. Our solution is an effective estimation method that can calculate the bit probability as accurately as possible by index assignment without recourse to side information. Moreover, we improve the coding performance of Rate-Adaptive LDPC (RA-LDPC), which is used in the parallelized DVC framework. This proposal selects a fitting sparse matrix for each bitplane according to the syndrome rate estimation results at the encoder side. Simulations show that our parallelization method reduces the decoding time by up to 35[%] and achieves a bit rate reduction of about 10[%].
Katsumi SAKAKIBARA Jumpei TAKETSUGU
We propose the use of an invertible code in cooperative multi-hop relaying networks. The effect of the code on the probability that an information block is undelivered to the destination is analyzed at the link level with a simple network topology. Numerical results indicate that significant improvement is feasible by an incorporation of an invertible code, since an information block can be reproduced by correcting channel errors in the received blocks at a relaying node.
Kentaro KOBAYASHI Takaya YAMAZATO Masaaki KATAYAMA
We propose an iterative channel decoding scheme for two or more multiple correlated sources. The correlated sources are separately turbo encoded without knowledge of the correlation and transmitted over noisy channels. The proposed decoder exploits the correlation of the multiple sources in an iterative soft decision decoding manner for joint detection of each of the transmitted data. Simulation results show that achieved performance for the more than two sources is also close to the Shannon and Slepian-Wolf limit and large additional SNR gain is obtained in comparison with the case of two sources. We also verify through simulation that no significant penalty results from the estimation of the source correlation in the decoding process and the code with a low error floor achieves good performance for a large number of the correlated sources.
Hironori UCHIKAWA Kohsuke HARADA
We propose a complexity-reducing algorithm for serial scheduled min-sum decoding that reduces the number of check nodes to process during an iteration. The check nodes to skip are chosen based on the reliability, a syndrome and a log-likelihood-ratio (LLR) value, of the incoming messages. The proposed algorithm is evaluated by computer simulations and shown to reduce the decoding complexity about 20% compared with a conventional serial scheduled min-sum decoding with small fractional decibel degradation in error correction performance.