Nat PAVASANT Takashi MORITA Masayuki NUMAO Ken-ichi FUKUI
We proposed a procedure to pre-process data used in a vector autoregressive (VAR) modeling of a temporal point process by using kernel density estimation. Vector autoregressive modeling of point-process data, for example, is being used for causality inference. The VAR model discretizes the timeline into small windows, and creates a time series by the presence of events in each window, and then models the presence of an event at the next time step by its history. The problem is that to get a longer history with high temporal resolution required a large number of windows, and thus, model parameters. We proposed the local density estimation procedure, which, instead of using the binary presence as the input to the model, performed kernel density estimation of the event history, and discretized the estimation to be used as the input. This allowed us to reduce the number of model parameters, especially in sparse data. Our experiment on a sparse Poisson process showed that this procedure vastly increases model prediction performance.
Yun JIANG Huiyang LIU Xiaopeng JIAO Ji WANG Qiaoqiao XIA
In this letter, a novel projection algorithm is proposed in which projection onto a triangle consisting of the three even-vertices closest to the vector to be projected replaces check polytope projection, achieving the same FER performance as exact projection algorithm in both high-iteration and low-iteration regime. Simulation results show that compared with the sparse affine projection algorithm (SAPA), it can improve the FER performance by 0.2 dB as well as save average number of iterations by 4.3%.
Minami SATO Sosuke MINAMOTO Ryuichi SAKAI Yasuyuki MURAKAMI
It is proven that many public-key cryptosystems would be broken by the quantum computer. The knapsack cryptosystem which is based on the subset sum problem has the potential to be a quantum-resistant cryptosystem. Murakami and Kasahara proposed a SOSI trapdoor sequence which is made by combining shifted-odd (SO) and super-increasing (SI) sequence in the modular knapsack cryptosystem. This paper firstly show that the key generation method could not achieve a secure density against the low-density attack. Second, we propose a high-density key generation method and confirmed that the proposed scheme is secure against the low-density attack.
Takumi HAYASHI Takeru ANDO Shouhei KIDERA
In this study, we propose an accurate range-Doppler analysis algorithm for moving multiple objects in a short range using microwave (including millimeter wave) radars. As a promising Doppler analysis for the above model, we previously proposed a weighted kernel density (WKD) estimator algorithm, which overcomes several disadvantages in coherent integration based methods, such as a trade-off between temporal and frequency resolutions. However, in handling multiple objects like human body, it is difficult to maintain the accuracy of the Doppler velocity estimation, because there are multiple responses from multiple parts of object, like human body, incurring inaccuracies in range or Doppler velocity estimation. To address this issue, we propose an iterative algorithm by exploiting an output of the WKD algorithm. Three-dimensional numerical analysis, assuming a human body model in motion, and experimental tests demonstrate that the proposed algorithm provides more accurate, high-resolution range-Doppler velocity profiles than the original WKD algorithm, without increasing computational complexity. Particularly, the simulation results show that the cumulative probabilities of range errors within 10mm, and Doppler velocity error within 0.1m/s are enhanced from 34% (by the former method) to 63% (by the proposed method).
When confirming the ACLR (adjacent channel leakage power ratio), which are representative indicators of distortion in the design of PA (power amplifier), it is well known how to calculate the AM-AM/PM characteristics of PA, input time series data of modulated signals, and analyze the output by Fourier analysis. In 5G (5th generation) mobile phones, not only QPSK (quadrature phase shift keying) modulation but also 16QAM (quadrature modulation), 64QAM, and 256QAM are becoming more multivalued as modulation signals. In addition, the modulation band may exceed 100MHz, and the amount of time series data increases, and the increase in calculation time becomes a problem. In order to shorten the calculation time, calculating the total amount of distortion generated by PA from the probability density of the modulation signal and the AM (amplitude modulation)-AM/PM (phase modulation) characteristics of PA is considered. For the AM-AM characteristics of PA, in this paper, IMD3 (inter modulation distortion 3) obtained from probability density and IMD3 by Fourier analysis, which are often used so long, are compared. As a result, it was confirmed that the result of probability density analysis is similar to that of Fourier analysis, when the nonlinearity is somewhat small. In addition, the agreement between the proposed method and the conventional method was confirmed with an error of about 2.0dB of ACLR using the modulation waves with a bandwidth of 5MHz, RB (resource block) being 25, and QPSK modulation.
Yujin ZHENG Junwei ZHANG Yan LIN Qinglin ZHANG Qiaoqiao XIA
The Euclidean projection operation is the most complex and time-consuming of the alternating direction method of multipliers (ADMM) decoding algorithms, resulting in a large number of resources when deployed on hardware platforms. We propose a simplified line segment projection algorithm (SLSA) and present the hardware design and the quantization scheme of the SLSA. In simulation results, the proposed SLSA module has a better performance than the original algorithm with the same fixed bitwidths due to the centrosymmetric structure of SLSA. Furthermore, the proposed SLSA module with a simpler structure without hypercube projection can reduce time consuming by up to 72.2% and reduce hardware resource usage by more than 87% compared to other Euclidean projection modules in the experiments.
Takayuki NOZAKI Motohiko ISAKA
Low-density parity-check (LDPC) codes are widely used in communication systems for their high error-correcting performance. This survey introduces the elements of LDPC codes: decoding algorithms, code construction, encoding algorithms, and several classes of LDPC codes.
Satoshi DENNO Kazuma HOTTA Yafei HOU
This paper proposes a novel maximum Doppler frequency detection technique for user moving velocity estimation. The maximum Doppler frequency is estimated in the proposed detection technique by making use of the fact that user moving velocity is not distributed continuously. The fluctuation of the channel state information during a packet is applied for the proposed detection, in which likelihood estimation is performed by comparing the fluctuation with the thresholds. The thresholds are theoretically derived on the assumption that the fluctuation is distributed with an exponential function. An approximated detection technique is proposed to simplify the theoretical threshold derivation. The performance of the proposed detection is evaluated by computer simulation. The proposed detection accomplishes better detection performance as the fluctuation values are summed over more packets. The proposed detection achieves about 90% correct detection performance in a fading channel with the Eb/N0 = 35dB, when the fluctuation values are summed over only three packets. Furthermore, the approximated detection also achieves the same detection performance.
In this letter, we study low-density parity-check (LDPC) codes for noisy channels with insertion and deletion (ID) errors. We first propose a design method of irregular LDPC codes for such channels, which can be used to simultaneously obtain degree distributions for different noise levels. We then show the asymptotic/finite-length decoding performances of designed codes and compare them with the symmetric information rates of cascaded ID-noisy channels. Moreover, we examine the relationship between decoding performance and a code structure of irregular LDPC codes.
Hiroki KAWAHARA Kyo INOUE Koji IGARASHI
This paper provides on a theoretical and numerical study of the probability density function (PDF) of the on-off keying (OOK) signals in ASE-limited systems. We present simple closed formulas of PDFs for the optical intensity and the received baseband signal. To confirm the validity of our model, the calculation results yielded by the proposed formulas are compared with those of numerical simulations and the conventional Gaussian model. Our theoretical and numerical results confirm that the signal distribution differs from a Gaussian profile. It is also demonstrated that our model can properly evaluate the signal distribution and the resultant BER performance, especially for systems with an optical bandwidth close to the receiver baseband width.
Yujin ZHENG Yan LIN Zhuo ZHANG Qinglin ZHANG Qiaoqiao XIA
Linear programming (LP) decoding based on the alternating direction method of multipliers (ADMM) has proved to be effective for low-density parity-check (LDPC) codes. However, for high-density parity-check (HDPC) codes, the ADMM-LP decoder encounters two problems, namely a high-density check matrix in HDPC codes and a great number of pseudocodewords in HDPC codes' fundamental polytope. The former problem makes the check polytope projection extremely complex, and the latter one leads to poor frame error rates (FER) performance. To address these issues, we introduce the even vertex algorithm (EVA) into the ADMM-LP decoding algorithm for HDPC codes, named as HDPC-EVA. HDPC-EVA can reduce the complexity of the projection process and improve the FER performance. We further enhance the proposed decoder by the automorphism groups of codes, creating diversity in the parity-check matrix. The simulation results show that the proposed decoder is capable of cutting down the average decoding time for each iteration by 30%-60%, as well as achieving near maximum likelihood (ML) performance on some BCH codes.
Mutsuo HIDAKA Shuichi NAGASAWA
This review provides a current overview of the fabrication processes for superconducting digital circuits at CRAVITY (clean room for analog and digital superconductivity) at the National Institute of Advanced Industrial Science and Technology (AIST), Japan. CRAVITY routinely fabricates superconducting digital circuits using three types of fabrication processes and supplies several thousand chips to its collaborators each year. Researchers at CRAVITY have focused on improving the controllability and uniformity of device parameters and the reliability, which means reducing defects. These three aspects are important for the correct operation of large-scale digital circuits. The current technologies used at CRAVITY permit ±10% controllability over the critical current density (Jc) of Josephson junctions (JJs) with respect to the design values, while the critical current (Ic) uniformity is within 1σ=2% for JJs with areas exceeding 1.0 µm2 and the defect density is on the order of one defect for every 100,000 JJs.
In this paper, we propose a robust parameters estimation algorithm for channel coded systems based on the low-density parity-check (LDPC) code over fading channels with impulse noise. The estimated parameters are then used to generate bit log-likelihood ratios (LLRs) for a soft-inputLDPC decoder. The expectation-maximization (EM) algorithm is used to estimate the parameters, including the channel gain and the parameters of the Bernoulli-Gaussian (B-G) impulse noise model. The parameters can be estimated accurately and the average number of iterations of the proposed algorithm is acceptable. Simulation results show that over a wide range of impulse noise power, the proposed algorithm approaches the optimal performance under different Rician channel factors and even under Middleton class-A (M-CA) impulse noise models.
Shunsuke YAMAKI Kazuhiro FUKUI Masahide ABE Masayuki KAWAMATA
This paper proposes statistical analysis of phase-only correlation (POC) functions under the phase fluctuation of signals due to additive Gaussian noise. We derive probability density function of phase-spectrum differences between original signal and its noise-corrupted signal with additive Gaussian noise. Furthermore, we evaluate the expectation and variance of the POC functions between these two signals. As the variance of Gaussian noise increases, the expectation of the peak of the POC function monotonically decreases and variance of the POC function monotonically increases. These results mathematically guarantee the validity of the POC functions used for similarity measure in matching techniques.
Chenyu WANG Kengo IOKIBE Yoshitaka TOYOTA
The plain bend in a pair of differential transmission lines causes a path difference, which leads to differential-to-common mode conversion due to the phase difference. This conversion can cause serious common-mode noise issues. We previously proposed a tightly coupled asymmetrically tapered bend to suppress forward differential-to-common mode conversion and derived the constraint conditions for high-density wiring. To provide sufficient suppression of mode conversion, however, the additional correction was required to make the effective path difference vanish. This paper proposes a practical and straightforward design methodology by using a very tightly coupled bend (decreasing the line width and the line separation of the tightly coupled bend). Full-wave simulations below 20GHz demonstrated that sufficient suppression of the forward differential-to-common mode conversion is successfully achieved as designed. Measurements showed that our design methodology is effective.
Kazunori IWATA Hiroki YAMAMOTO Kazushi MIMURA
Shape matching with local descriptors is an underlying scheme in shape analysis. We can visually confirm the matching results and also assess them for shape classification. Generally, shape matching is implemented by determining the correspondence between shapes that are represented by their respective sets of sampled points. Some matching methods have already been proposed; the main difference between them lies in their choice of matching cost function. This function measures the dissimilarity between the local distribution of sampled points around a focusing point of one shape and the local distribution of sampled points around a referring point of another shape. A local descriptor is used to describe the distribution of sampled points around the point of the shape. In this paper, we propose an extended scheme for shape matching that can compensate for errors in existing local descriptors. It is convenient for local descriptors to adopt our scheme because it does not require the local descriptors to be modified. The main idea of our scheme is to consider the correspondence of neighboring sampled points to a focusing point when determining the correspondence of the focusing point. This is useful because it increases the chance of finding a suitable correspondence. However, considering the correspondence of neighboring points causes a problem regarding computational feasibility, because there is a substantial increase in the number of possible correspondences that need to be considered in shape matching. We solve this problem using a branch-and-bound algorithm, for efficient approximation. Using several shape datasets, we demonstrate that our scheme yields a more suitable matching than the conventional scheme that does not consider the correspondence of neighboring sampled points, even though our scheme requires only a small increase in execution time.
Ryo SHIBATA Gou HOSOYA Hiroyuki YASHIMA
We propose a coding/decoding strategy that surpasses the symmetric information rate of a binary insertion/deletion (ID) channel and approaches the Markov capacity of the channel. The proposed codes comprise inner trellis codes and outer irregular low-density parity-check (LDPC) codes. The trellis codes are designed to mimic the transition probabilities of a Markov input process that achieves a high information rate, whereas the LDPC codes are designed to maximize an iterative decoding threshold in the superchannel (concatenation of the ID channels and trellis codes).
For low-density parity-check (LDPC) codes, the penalized decoding method based on the alternating direction method of multipliers (ADMM) can improve the decoding performance at low signal-to-noise ratios and also has low decoding complexity. There are three effective methods that could increase the ADMM penalized decoding speed, which are reducing the number of Euclidean projections in ADMM penalized decoding, designing an effective penalty function and selecting an appropriate layered scheduling strategy for message transmission. In order to further increase the ADMM penalized decoding speed, through reducing the number of Euclidean projections and using the vertical layered scheduling strategy, this paper designs a fast converging ADMM penalized decoding method based on the improved penalty function. Simulation results show that the proposed method not only improves the decoding performance but also reduces the average number of iterations and the average decoding time.
Masamune NOMURA Yuki NAKAMURA Hiroo TARAO Amane TAKEI
This paper describes the effectiveness of the geometric multi-grid method in a current density analysis using a numerical human body model. The scalar potential finite difference (SPFD) method is used as a numerical method for analyzing the current density inside a human body due to contact with charged objects in a low-frequency band, and research related to methods to solve faster large-scale simultaneous equations based on the SPFD method has been conducted. In previous research, the block incomplete Cholesky conjugate gradients (ICCG) method is proposed as an effective method to solve the simultaneous equations faster. However, even though the block ICCG method is used, many iterations are still needed. Therefore, in this study, we focus on the geometric multi-grid method as a method to solve the problem. We develop the geometric-multi-grid method and evaluate performances by comparing it with the block ICCG method in terms of computation time and the number of iterations. The results show that the number of iterations needed for the geometric multi-grid method is much less than that for the block ICCG method. In addition, the computation time is much shorter, depending on the number of threads and the number of coarse grids. Also, by using multi-color ordering, the parallel performance of the geometric multi-grid method can be greatly improved.
Haiyang LIU Hao ZHANG Lianrong MA Lingjun KONG
In this letter, the structural analysis of nonbinary cyclic and quasi-cyclic (QC) low-density parity-check (LDPC) codes with α-multiplied parity-check matrices (PCMs) is concerned. Using analytical methods, several structural parameters of nonbinary cyclic and QC LDPC codes with α-multiplied PCMs are determined. In particular, some classes of nonbinary LDPC codes constructed from finite fields and finite geometries are shown to have good minimum and stopping distances properties, which may explain to some extent their wonderful decoding performances.