This paper shows that sum-of-product expression (SOP) minimization produces the generalization ability. We show this in three steps. First, various classes of SOPs are generated. Second, minterms of SOP are randomly selected to generate partially defined functions. And, third, from the partially defined functions, original functions are reconstructed by SOP minimization. We consider Achilles heel functions, majority functions, monotone increasing cascade functions, functions generated from random SOPs, monotone increasing random SOPs, circle functions, and globe functions. As for the generalization ability, the presented method is compared with Naive Bayes, multi-level perceptron, support vector machine, JRIP, J48, and random forest. For these functions, in many cases, only 10% of the input combinations are sufficient to reconstruct more than 90% of the truth tables of the original functions.
Mengmeng ZHANG Zeliang ZHANG Yuan LI Ran CHENG Hongyuan JING Zhi LIU
Point cloud video contains not only color information but also spatial position information and usually has large volume of data. Typical rate distortion optimization algorithms based on Human Visual System only consider the color information, which limit the coding performance. In this paper, a Coding Tree Unit (CTU) level quantization parameter (QP) adjustment algorithm based on JND and spatial complexity is proposed to improve the subjective and objective quality of Video-Based Point Cloud Compression (V-PCC). Firstly, it is found that the JND model is degraded at CTU level for attribute video due to the pixel filling strategy of V-PCC, and an improved JND model is designed using the occupancy map. Secondly, a spatial complexity detection metric is designed to measure the visual importance of each CTU. Finally, a CTU-level QP adjustment scheme based on both JND levels and visual importance is proposed for geometry and attribute video. The experimental results show that, compared with the latest V-PCC (TMC2-18.0) anchors, the BD-rate is reduced by -2.8% and -3.2% for D1 and D2 metrics, respectively, and the subjective quality is improved significantly.
Daisuke HIBINO Tomoharu SHIBUYA
Distributed computing is one of the powerful solutions for computational tasks that need the massive size of dataset. Lagrange coded computing (LCC), proposed by Yu et al. [15], realizes private and secure distributed computing under the existence of stragglers, malicious workers, and colluding workers by using an encoding polynomial. Since the encoding polynomial depends on a dataset, it must be updated every arrival of new dataset. Therefore, it is necessary to employ efficient algorithm to construct the encoding polynomial. In this paper, we propose Newton coded computing (NCC) which is based on Newton interpolation to construct the encoding polynomial. Let K, L, and T be the number of data, the length of each data, and the number of colluding workers, respectively. Then, the computational complexity for construction of an encoding polynomial is improved from O(L(K+T)log 2(K+T)log log (K+T)) for LCC to O(L(K+T)log (K+T)) for the proposed method. Furthermore, by applying the proposed method, the computational complexity for updating the encoding polynomial is improved from O(L(K+T)log 2(K+T)log log (K+T)) for LCC to O(L) for the proposed method.
Satoshi DENNO Shuhei MAKABE Yafei HOU
This paper proposes a non-linear overloaded MIMO detector that outperforms the conventional soft-input maximum likelihood detector (MLD) with less computational complexity. We propose iterative log-likelihood ratio (LLR) estimation and multi stage LLR estimation for the proposed detector to achieve such superior performance. While the iterative LLR estimation achieves better BER performance, the multi stage LLR estimation makes the detector less complex than the conventional soft-input maximum likelihood detector (MLD). The computer simulation reveals that the proposed detector achieves about 0.6dB better BER performance than the soft-input MLD with about half of the soft-input MLD's complexity in a 6×3 overloaded MIMO OFDM system.
Jinguang HAO Gang WANG Honggang WANG Lili WANG Xuefeng LIU
In software defined radio systems, a channelizer plays an important role in extracting the desired signals from a wideband signal. Compared to the conventional methods, the proposed scheme provides a solution to design a digital channelizer extracting the multiple subband signals at different center frequencies with low complexity. To do this, this paper formulates the problem as an optimization problem, which minimizes the required multiplications number subject to the constraints of the ripple in the passbands and the stopbands for single channel and combined multiple channels. In addition, a solution to solve the optimization problem is also presented and the corresponding structure is demonstrated. Simulation results show that the proposed scheme requires smaller number of the multiplications than other conventional methods. Moreover, unlike other methods, this structure can process signals with different bandwidths at different center frequencies simultaneously only by changing the status of the corresponding multiplexers without hardware reimplementation.
In this survey we summarize properties of pseudorandomness and non-randomness of some number-theoretic sequences and present results on their behaviour under the following measures of pseudorandomness: balance, linear complexity, correlation measure of order k, expansion complexity and 2-adic complexity. The number-theoretic sequences are the Legendre sequence and the two-prime generator, the Thue-Morse sequence and its sub-sequence along squares, and the prime omega sequences for integers and polynomials.
Satoshi DENNO Taichi YAMAGAMI Yafei HOU
This paper proposes low complexity resource allocation in frequency domain non-orthogonal multiple access where many devices access with a base station. The number of the devices is assumed to be more than that of the resource for network capacity enhancement, which is demanded in massive machine type communications (mMTC). This paper proposes two types of resource allocation techniques, all of which are based on the MIN-MAX approach. One of them seeks for nicer resource allocation with only channel gains. The other technique applies the message passing algorithm (MPA) for better resource allocation. The proposed resource allocation techniques are evaluated by computer simulation in frequency domain non-orthogonal multiple access. The proposed technique with the MPA achieves the best bit error rate (BER) performance in the proposed techniques. However, the computational complexity of the proposed techniques with channel gains is much smaller than that of the proposed technique with the MPA, whereas the BER performance of the proposed techniques with channel gains is only about 0.1dB inferior to that with the MPA in the multiple access with the overloading ratio of 1.5 at the BER of 10-4. They attain the gain of about 10dB at the BER of 10-4 in the multiple access with the overloading ration of 2.0. Their complexity is 10-16 as small as the conventional technique.
Jiang MA Jun ZHANG Yanguo JIA Xiumin SHEN
Pseudorandom sequences with large linear complexity can resist the linear attack. The trace representation plays an important role in analysis and design of pseudorandom sequences. In this letter, we present the construction of a family of new binary sequences derived from Euler quotients modulo pq, where pq is a product of two primes and p divides q-1. Firstly, the linear complexity of the sequences are investigated. It is proved that the sequences have larger linear complexity and can resist the attack of Berlekamp-Massey algorithm. Then, we give the trace representation of the proposed sequences by determining the corresponding defining pair. Moreover, we generalize the result to the Euler quotients modulo pmqn with m≤n. Results indicate that the generalized sequences still have high linear complexity. We also give the trace representation of the generalized sequences by determining the corresponding defining pair. The result will be helpful for the implementation and the pseudorandom properties analysis of the sequences.
Automatic modulation recognition (AMR) plays a critical role in modern communication systems. Owing to the recent advancements of deep learning (DL) techniques, the application of DL has been widely studied in AMR, and a large number of DL-AMR algorithms with high recognition rates have been developed. Most DL-AMR algorithm models have high recognition accuracy but have numerous parameters and are huge, complex models, which make them hard to deploy on resource-constrained platforms, such as satellite platforms. Some lightweight and low-complexity DL-AMR algorithm models also struggle to meet the accuracy requirements. Based on this, this paper proposes a lightweight and high-recognition-rate DL-AMR algorithm model called Lightweight Densely Connected Convolutional Network (DenseNet) Long Short-Term Memory network (LDLSTM). The model cascade of DenseNet and LSTM can achieve the same recognition accuracy as other advanced DL-AMR algorithms, but the parameter volume is only 1/12 that of these algorithms. Thus, it is advantageous to deploy LDLSTM in resource-constrained systems.
Tomoko K. MATSUSHIMA Shoichiro YAMASAKI Hirokazu TANAKA
Recently, complex orthogonal variable spreading factor (OVSF) codes based on polyphase orthogonal codes have been proposed to support multi-user/multi-rate data transmission services in synchronous direct-sequence code-division multiple access (DS-CDMA) systems. This study investigates the low signal-envelope fluctuation property of the complex OVSF codes in terms of transmission signal trajectories. In addition, a new method is proposed to suppress the envelope fluctuation more strongly at the expense of reducing the number of spreading sequences of the codes.
Shoji KASAHARA Jun KAWAHARA Shin-ichi MINATO Jumpei MORI
This paper analyzes a blockchain network forming a directed acyclic graph (DAG), called a DAG-type blockchain, from the viewpoint of graph algorithm theory. To use a DAG-type blockchain, NP-hard graph optimization problems on the DAG are required to be solved. Although various problems for undirected and directed graphs can be efficiently solved by using the notions of graph parameters, these currently known parameters are meaningless for DAGs, which implies that it is hopeless to design efficient algorithms based on the parameters for such problems. In this work, we propose a novel graph parameter for directed graphs called a DAG-pathwidth, which represents the closeness to a directed path. This is an extension of the pathwidth, a well-known graph parameter for undirected graphs. We analyze the features of the DAG-pathwidth and prove that computing the DAG-pathwidth of a DAG (directed graph in general) is NP-complete. Finally, we propose an efficient algorithm for a variant of the maximum k-independent set problem for the DAG-type blockchain when the DAG-pathwidth of the input graph is small.
Satoshi DENNO Koki KASHIHARA Yafei HOU
This paper proposes a novel approach to low complexity soft input decoding for lattice reduction-aided MIMO receivers. The proposed approach feeds a soft input decoder with soft signals made from hard decision signals generated by using a lattice reduction-aided linear detector. The soft signal is a weighted-sum of some candidate vectors that are near by the hard decision signal coming out from the lattice reduction-aided linear detector. This paper proposes a technique to adjust the weight adapt to the channel for the higher transmission performance. Furthermore, we propose to introduce a coefficient that is used for the weights in order to enhance the transmission performance. The transmission performance is evaluated in a 4×4 MIMO channel. When a linear MMSE filter or a serial interference canceller is used as the linear detector, the proposed technique achieves about 1.0dB better transmission performance at the BER of 10-5 than the decoder fed with the hard decision signals. In addition, the low computational complexity of the proposed technique is quantitatively evaluated.
Seiya KISHIMOTO Naoya ISHIKAWA Shinichiro OHNUKI
In this study, a computational method is proposed for acoustic field analysis tasks that require lengthy observation times. The acoustic fields at a given observation time are obtained using a fast inverse Laplace transform with a finite-difference complex-frequency-domain. The transient acoustic field can be evaluated at arbitrary sampling intervals by obtaining the instantaneous acoustic field at the desired observation time using the proposed method.
Lu ZHAO Bo XU Tianqing CAO Jiao DU
A unified construction for yielding optimal and balanced quaternary sequences from ideal/optimal balanced binary sequences was proposed by Zeng et al. In this paper, the linear complexity over finite field 𝔽2, 𝔽4 and Galois ring ℤ4 of the quaternary sequences are discussed, respectively. The exact values of linear complexity of sequences obtained by Legendre sequence pair, twin-prime sequence pair and Hall's sextic sequence pair are derived.
Duc A. HOANG Akira SUZUKI Tsuyoshi YAGITA
A vertex subset I of a graph G is called a k-path vertex cover if every path on k vertices in G contains at least one vertex from I. The K-PATH VERTEX COVER RECONFIGURATION (K-PVCR) problem asks if one can transform one k-path vertex cover into another via a sequence of k-path vertex covers where each intermediate member is obtained from its predecessor by applying a given reconfiguration rule exactly once. We investigate the computational complexity of K-PVCR from the viewpoint of graph classes under the well-known reconfiguration rules: TS, TJ, and TAR. The problem for k=2, known as the VERTEX COVER RECONFIGURATION (VCR) problem, has been well-studied in the literature. We show that certain known hardness results for VCR on different graph classes can be extended for K-PVCR. In particular, we prove a complexity dichotomy for K-PVCR on general graphs: on those whose maximum degree is three (and even planar), the problem is PSPACE-complete, while on those whose maximum degree is two (i.e., paths and cycles), the problem can be solved in polynomial time. Additionally, we also design polynomial-time algorithms for K-PVCR on trees under each of TJ and TAR. Moreover, on paths, cycles, and trees, we describe how one can construct a reconfiguration sequence between two given k-path vertex covers in a yes-instance. In particular, on paths, our constructed reconfiguration sequence is shortest.
At present, the application of different types of memristors in electronics is being deeply studied. Given the nonlinearity characterizing memristors, a circuit with memristors cannot be treated by classical circuit analysis. In this paper, memristor is equivalent to a nonlinear dynamic system composed of linear dynamic system and nonlinear static system by Volterra series. The nonlinear transfer function of memristor is derived. In the complex frequency domain, the n-order complex frequency response of memristor is established by multiple Laplace transform, and the response of MLC parallel circuit is taken as an example to verify. Theoretical analysis shows that the complex frequency domain analysis method of memristor transforms the problem of solving nonlinear circuit in time domain into n times complex frequency domain analysis of linear circuit, which provides an idea for nonlinear dynamic system analysis.
Chun-e ZHAO Yuhua SUN Tongjiang YAN Xubo ZHAO
Binary sequences with high linear complexity and high 2-adic complexity have important applications in communication and cryptography. In this paper, the 2-adic complexity of a class of balanced Whiteman generalized cyclotomic sequences which have high linear complexity is considered. Through calculating the determinant of the circulant matrix constructed by one of these sequences, the result shows that the 2-adic complexity of this class of sequences is large enough to resist the attack of the rational approximation algorithm (RAA) for feedback with carry shift registers (FCSRs).
Tianfeng FENG Leonie RYVKIN Jérôme URHAUSEN Giovanni VIGLIETTA
We study the computational complexity of the puzzle game Critter Crunch, where the player has to rearrange Critters on a board in order to eliminate them all. Smaller Critters can be fed to larger Critters, and Critters will explode if they eat too much. Critters come in several different types, sizes, and colors. We prove the NP-hardness of levels that contain Blocker Critters, as well as levels where the player must clear the board in a given number of moves (i.e., “puzzle mode”). We also characterize the complexity of the game, as a function of the number of columns on the board, in two settings: (i) the setting where Critters may have several different colors, but only two possible sizes, and (ii) the setting where Critters come in all three sizes, but with no color variations. In both settings, the game is NP-hard for levels with exactly two columns, and solvable in linear time for levels with only one column or more than two columns.
Aye Mon HTUN Maung SANN MAW Iwao SASASE P. Takis MATHIOPOULOS
In this paper, we propose a novel user selection scheme based on jointly combining channel gain (CG) and signal to interference plus noise ratio (SINR) to improve the sum-rate as well as to reduce the computation complexity of multi-user massive multi-input multi-output (MU-massive MIMO) downlink transmission through a block diagonalization (BD) precoding technique. By jointly considering CG and SINR based user sets, sum-rate performance improvement can be achieved by selecting higher gain users with better SINR conditions as well as by eliminating the users who cause low sum-rate in the system. Through this approach, the number of possible outcomes for the user selection scheme can be reduced by counting the common users for every pair of user combinations in the selection process since the common users of CG-based and SINR-based sets possess both higher channel gains and better SINR conditions. The common users set offers not only sum-rate performance improvements but also computation complexity reduction in the proposed scheme. It is shown by means of computer simulation experiments that the proposed scheme can increase the sum-rate with lower computation complexity for various numbers of users as compared to conventional schemes requiring the same or less computational complexity.
Sakae NAGAOKA Mark BROWN Daniel DELAHAYE
Air traffic management (ATM) systems around the world are being modernized to accommodate shifts towards performance- and trajectory-based operations. These shifts will require new indices for safety, efficiency and complexity. The authors have been developing an index for evaluating air traffic control (ATC) difficulty that utilizes the relative positions and velocity vectors of aircraft pairs as input data. Prior to practical application of the index, it is necessary to understand the effects of input data error, i.e. errors in the positions and velocities of a pair of aircraft, on the estimated difficulty value. Two sensitivity analyses were therefore performed for a pair of aircraft cruising at constant speeds on intersecting linear tracks at the same altitude. Sensitivity analysis examines how uncertainty in inputs relates to uncertainty in outputs. Firstly, an analysis of propagation error was carried out. The formula of the propagation error at a certain point was derived based on the assumed input error, and the distribution of propagation error was investigated for all possible situations and compared with the distribution of difficulty values to clarify its characteristics. Secondly, a sensitivity analysis based on variance was carried out that evaluated the effect of each input parameter using a conditional variance value called the Sobol indices. Using a Monte Carlo method, we investigated the effect of each input parameter on the calculated difficulty value for all possible situations of aircraft pairs on intersecting trajectories. As a result, it was found that the parameter that most affects the difficulty value is the intersection angle of the trajectories.