Taiki ARAKAWA Kazuhiro YAMAGUCHI Kazunori KAMEDA Shinichi FURUKAWA
We study the device length and/or band characteristics examined by two coupling analysis methods for our proposed fiber-type polarization splitter (FPS) composed of single mode fiber and polarization maintaining fiber. The first method is based on the power transition characteristics of the coupled-mode theory (CMT), and the second, a more accurate analysis method, is based on improved fundamental mode excitation (IFME). The CMT and IFME were evaluated and investigated with respect to the device length and bandwidth characteristics of the FPS. In addition, the influence of the excitation point shift of the fundamental mode, which has not been almost researched so far, is also analysed by using IFME.
Hiroshi SUENOBU Shin-ichi YAMAMOTO Michio TAKIKAWA Naofumi YONEDA
A method for bandwidth enhancement of radar cross section (RCS) reduction by metasurfaces was studied. Scattering cancellation is one of common methods for reducing RCS of target scatterers. It occurs when the wave scattered by the target scatterer and the wave scattered by the canceling scatterer are the same amplitude and opposite phase. Since bandwidth of scattering cancellation is usually narrow, we proposed the bandwidth enhancement method using metasurfaces, which can control the frequency dependence of the scattering phase. We designed and fabricated a metasurface composed of a patch array on a grounded dielectric substrate. Numerical and experimental evaluations confirmed that the metasurface enhances the bandwidth of 10dB RCS reduction by 52% bandwidth ratio of the metasurface from 34% bandwidth ratio of metallic cancelling scatterers.
Displacement current is the last piece of the puzzle of electromagnetic theory. Its existence implies that electromagnetic disturbance can propagate at the speed of light and finally it led to the discovery of Hertzian waves. On the other hand, since magnetic fields can be calculated only with conduction currents using Biot-Savart's law, a popular belief that displacement current does not produce magnetic fields has started to circulate. But some people think if this is correct, what is the displacement current introduced for. The controversy over the meaning of displacement currents has been going on for more than hundred years. Such confusion is caused by forgetting the fact that in the case of non-stationary currents, neither magnetic fields created by conduction currents nor those created by displacement currents can be defined. It is also forgotten that the effect of displacement current is automatically incorporated in the magnetic field calculated by Biot-Savart's law. In this paper, mainly with the help of Helmholtz decomposition, we would like to clarify the confusion surrounding displacement currents and provide an opportunity to end the long standing controversy.
Jichen BIAN Min ZHENG Hong LIU Jiahui MAO Hui LI Chong TAN
Wi-Fi-based person identification (PI) tasks are performed by analyzing the fluctuating characteristics of the Channel State Information (CSI) data to determine whether the person's identity is legitimate. This technology can be used for intrusion detection and keyless access to restricted areas. However, the related research rarely considers the restricted computing resources and the complexity of real-world environments, resulting in lacking practicality in some scenarios, such as intrusion detection tasks in remote substations without public network coverage. In this paper, we propose a novel neural network model named SimpleViTFi, a lightweight classification model based on Vision Transformer (ViT), which adds a downsampling mechanism, a distinctive patch embedding method and learnable positional embedding to the cropped ViT architecture. We employ the latest IEEE 802.11ac 80MHz CSI dataset provided by [1]. The CSI matrix is abstracted into a special “image” after pre-processing and fed into the trained SimpleViTFi for classification. The experimental results demonstrate that the proposed SimpleViTFi has lower computational resource overhead and better accuracy than traditional classification models, reflecting the robustness on LOS or NLOS CSI data generated by different Tx-Rx devices and acquired by different monitors.
Zikang CHEN Wenping GE Henghai FEI Haipeng ZHAO Bowen LI
The combination of multiple-input multiple-output (MIMO) technology and sparse code multiple access (SCMA) can significantly enhance the spectral efficiency of future wireless communication networks. However, the receiver design for downlink MIMO-SCMA systems faces challenges in developing multi-user detection (MUD) schemes that achieve both low latency and low bit error rate (BER). The separated detection scheme in the MIMO-SCMA system involves performing MIMO detection first to obtain estimated signals, followed by SCMA decoding. We propose an enhanced separated detection scheme based on lightweight graph neural networks (GNNs). In this scheme, we raise the concept of coordinate point relay and full-category training, which allow for the substitution of the conventional message passing algorithm (MPA) in SCMA decoding with image classification techniques based on deep learning (DL). The features of the images used for training encompass crucial information such as the amplitude and phase of estimated signals, as well as channel characteristics they have encountered. Furthermore, various types of images demonstrate distinct directional trends, contributing additional features that enhance the precision of classification by GNNs. Simulation results demonstrate that the enhanced separated detection scheme outperforms existing separated and joint detection schemes in terms of computational complexity, while having a better BER performance than the joint detection schemes at high Eb/N0 (energy per bit to noise power spectral density ratio) values.
Varuliantor DEAR Annis SIRADJ MARDIANI Nandang DEDI Prayitno ABADI Baud HARYO PRANANTO ISKANDAR
Low capacity and reliability are the challenges in the development of ionosphere communication channel systems. To overcome this problem, one promising and state-of-the-art method is applying a multi-carrier modulation technique. Currently, the use of multi-carrier modulation technique is using a single transmission frequency with a bandwidth is no more than 24 kHz in real-world implementation. However, based on the range of the minimum and maximum ionospheric plasma frequency values, which could be in the MHz range, the use of these values as the main bandwidth in multi-carrier modulation techniques can optimize the use of available channel capacity. In this paper, we propose a multi-carrier modulation technique in combination with a model variation of Lowest Usable Frequency (LUF) and Maximum Usable Frequency (MUF) values as the main bandwidth to optimize the use of available channel capacity while also maintaining its reliability by following the variation of the ionosphere plasma frequency. To analyze its capacity and reliability, we performed a numeric simulation using a LUF-MUF model based on Long Short Term-Memory (LSTM) and Advanced Stand Alone Prediction System (ASAPS) in Near Vertical Incidence Skywave (NVIS) propagation mode with the assumption of perfect synchronization between transmitter and receiver with no Doppler and no time offsets. The results show the achievement of the ergodic channel capacity varies for every hour of the day, with values in the range of 10 Mbps and 100 Mbps with 0 to 20 dB SNR. Meanwhile, the reliability of the system is in the range of 8% to 100% for every hour of one day based on two different Mode Reliability calculation scenarios. The results also show that channel capacity and system reliability optimization are determined by the accuracy of the LUF-MUF model.
Kai IKUTA Jinya NAKAMURA Moriya NAKAMURA
In this paper, we investigated the overfitting characteristics of nonlinear equalizers based on an artificial neural network (ANN) and the Volterra series transfer function (VSTF), which were designed to compensate for optical nonlinear waveform distortion in optical fiber communication systems. Linear waveform distortion caused by, e.g., chromatic dispersion (CD) is commonly compensated by linear equalizers using digital signal processing (DSP) in digital coherent receivers. However, mitigation of nonlinear waveform distortion is considered to be one of the next important issues. An ANN-based nonlinear equalizer is one possible candidate for solving this problem. However, the risk of overfitting of ANNs is one obstacle in using the technology in practical applications. We evaluated and compared the overfitting of ANN- and conventional VSTF-based nonlinear equalizers used to compensate for optical nonlinear distortion. The equalizers were trained on repeated random bit sequences (RRBSs), while varying the length of the bit sequences. When the number of hidden-layer units of the ANN was as large as 100 or 1000, the overfitting characteristics were comparable to those of the VSTF. However, when the number of hidden-layer units was 10, which is usually enough to compensate for optical nonlinear distortion, the overfitting was weaker than that of the VSTF. Furthermore, we confirmed that even commonly used finite impulse response (FIR) filters showed overfitting to the RRBS when the length of the RRBS was equal to or shorter than the length of the tapped delay line of the filters. Conversely, when the RRBS used for the training was sufficiently longer than the tapped delay line, the overfitting could be suppressed, even when using an ANN-based nonlinear equalizer with 10 hidden-layer units.
This letter deals with the joint direction of arrival and direction of departure estimation problem for overloaded target in bistatic multiple-input multiple-output radar system. In order to achieve the purpose of effective estimation, the presented Khatri-Rao (KR) MUSIC estimator with the ability to handle overloaded targets mainly combines the subspace characteristics of the target reflected wave signal and the KR product based on the array response. This letter also presents a computationally efficient KR noise subspace projection matrix estimation technique to reduce the computational load due to perform high-dimensional singular value decomposition. Finally, the effectiveness of the proposed method is verified by computer simulation.
Given an odd prime q and an integer m ≤ q, a binary mq × q2 quasi-cyclic parity-check matrix H(m, q) can be constructed for an array low-density parity-check (LDPC) code C (m, q). In this letter, we investigate the first separating redundancy of C (m, q). We prove that H (m, q) is 1-separating for any pair of (m, q), from which we conclude that the first separating redundancy of C (m, q) is upper bounded by mq. Then we show that our upper bound on the first separating redundancy of C (m, q) is tighter than the general deterministic and constructive upper bounds in the literature. For m=2, we further prove that the first separating redundancy of C (2, q) is 2q for any odd prime q. For m ≥ 3, we conjecture that the first separating redundancy of C (m, q) is mq for any fixed m and sufficiently large q.
Katsuya KOSUKEGAWA Kazuhiko KAWAMOTO
We considered the problem of forecasting the degradation recovery process of civil structures for prognosis and health management. In this process, structural health degrades over time but recovers when a maintenance intervention is performed. Maintenance interventions are typically recorded in terms of date and type. Such records can be represented as binary time series. Using binary maintenance intervention records, we forecast the process by using Long Short-Term Memory (LSTM). In this study, we experimentally examined how to feed binary time series data into LSTM. To this end, we compared the concatenation and reinitialization methods. The former is used to concatenate maintenance intervention records and health data and feed them into LSTM. The latter is used to reinitialize the LSTM internal memory when maintenance intervention is performed. The experimental results with the synthetic data revealed that the concatenation method outperformed the reinitialization method.
Lin CHEN Xueyuan YIN Dandan ZHAO Hongwei LU Lu LI Yixiang CHEN
ARM chips with low energy consumption and low-cost investment have been rapidly applied to smart office and smart entertainment including cloud mobile phones and cloud games. This paper first summarizes key technologies and development status of the above scenarios including CPU, memory, IO hardware virtualization characteristics, ARM hypervisor and container, GPU virtualization, network virtualization, resource management and remote transmission technologies. Then, in view of the current lack of publicly referenced ARM cloud constructing solutions, this paper proposes and constructs an implementation framework for building an ARM cloud, and successively focuses on the formal definition of virtualization framework, Android container system and resource quota management methods, GPU virtualization based on API remoting and GPU pass-through, and the remote transmission technology. Finally, the experimental results show that the proposed model and corresponding component implementation methods are effective, especially, the pass-through mode for virtualizing GPU resources has higher performance and higher parallelism.
Kazuya SHIMEI Kentaro KOBAYASHI Wataru CHUJO
We study a visible light communication (VLC) system that modulates data signals by changing the color components of image contents on a digital signage display, captures them with an image sensor, and demodulates them using image processing. This system requires that the modulated data signals should not be perceived by the human eye. Previous studies have proposed modulation methods with a chromaticity component that is difficult for the human eye to perceive, and we have also proposed a modulation method with perceptually uniform color space based on human perception characteristics. However, which chromaticity component performs better depends on the image contents, and the evaluation only for some specific image contents was not sufficient. In this paper, we evaluate the communication and visual quality of the modulation methods with chromaticity components for various standard images to clarify the superiority of the method with perceptually uniform color space. In addition, we propose a novel modulation and demodulation method using diversity combining to eliminate the dependency of performance on the image contents. Experimental results show that the proposed method can improve the communication and visual quality for almost all the standard images.
Noboru HAYASAKA Riku KASAI Takuya FUTAGAMI
In this paper, we propose a noise-robust scream detection method with the aim of expanding the scream detection system, a sound-based security system. The proposed method uses enhanced screams using Wave-U-Net, which was effective as a noise reduction method for noisy screams. However, the enhanced screams showed different frequency components from clean screams and erroneously emphasized frequency components similar to scream in noise. Therefore, Wave-U-Net was applied even in the process of training Gaussian mixture models, which are discriminators. We conducted detection experiments using the proposed method in various noise environments and determined that the false acceptance rate was reduced by an average of 2.1% or more compared with the conventional method.
Sumiko MIYATA Ryoichi SHINKUMA
Streaming systems that can maintain Quality of Experience (QoE) for users have attracted much attention because they can be applied in various fields, such as emergency response training and medical surgery. Dynamic Adaptive Streaming over HTTP (DASH) is a typical protocol for streaming system. In order to improve QoE in DASH, a multi-server system has been presented by pseudo-increasing bandwidth through multiple servers. This multi-server system is designed to share streaming content efficiently in addition to having redundant server resources for each streaming content, which is excellent for fault tolerance. Assigning DASH server to users in these multi-servers environment is important to maintain QoE, thus a method of server assignment of users (user allocation method) for multi-servers is presented by using cooperative game theory. However, this conventional user allocation method does not take into account the size of the server bandwidth, thus users are concentrated on a particular server at the start of playback. Although the average required bit rate of video usually fluctuates, bit rate fluctuations are not taken into account. These phenomena may decrease QoE. In this paper, we propose a novel user allocation method using coalition structure generation in cooperative game theory to improve the QoE of all users in an immediate and stable manner in DASH environment. Our proposed method can avoid user concentration, since the bandwidth used by the overall system is taken into account. Moreover, our proposed method can be performed every time the average required bit rate changes. We demonstrate the effectiveness of our method through simulations using Network Simulator 3 (NS3).
Adiabatic logic circuits are regarded as one of the most attractive solutions for low-power circuit design. This study is dedicated to optimizing the design of the Two-Level Adiabatic Logic (2LAL) circuit, which boasts a relatively simple structure and superior low-power performance among many asymptotically adiabatic or quasi-adiabatic logic families, but suffers from a large number of timing buffers for “decompute”. Our focus is on the “early decompute” technique for fully pipelined 2LAL, and we propose two ILP approaches for minimizing hardware cost through optimization of early decompute. In the first approach, the problem is formulated as a kind of scheduling problem, while it is reformulated as node selection problem (stable set problem). The performance of the proposed methods are evaluated using several benchmark circuits from ISCAS-85, and the maximum 70% hardware reduction is observed compared with an existing method.
Sohei SHIMOMAI Kei UEDA Shinji KIMURA
Recently, Quantum Annealing (QA) has attracted attention as an efficient algorithm for combinatorial optimization problems. In QA, the input data size becomes large and its reduction is important for accelerating by the hardware emulation since the usable memory size and its bandwidth are limited. The paper proposes the compression method of input sparse matrices for QA emulator. The proposed method uses the sparseness of the coefficient matrix and the reappearance of the same values. An independent table is introduced and data are compressed by the search and registration method of two consecutive data in the value table. The proposed method is applied to Traveling Salesman Problem (TSP) with 32, 64 and 96 cities and Nurse Scheduling Problem (NSP). The proposed method could reduce the amount of data by 1/40 for 96 city TSP and could manage 96 city TSP on the hardware emulator. When applied to NSP, we confirmed the effectiveness of the proposed method by the compression ratio ranging from 1/4 to 1/11.8. The data reduction is also useful for the simulation/emulation performance when using the compressed data directly and 1.9 times faster speed can be found on 96 city TSP than the CSR-based method.
We have realized a design automation platform of hardware accelerator for pairing operation over multiple elliptic curve parameters. Pairing operation is one of the fundamental operations to realize functional encryption. However, known as a computational complexity-heavy algorithm. Also because there have been not yet identified standard parameters, we need to choose curve parameters based on the required security level and affordable hardware resources. To explore this design optimization for each curve parameter is essential. In this research, we have realized an automated design platform for pairing hardware for such purposes. Optimization results show almost equivalent to those prior-art designs by hand.
Jiaxuan LU Yutaka MASUDA Tohru ISHIHARA
Approximate computing (AC) saves energy and improves performance by introducing approximation into computation in error-torrent applications. This work focuses on an AC strategy that accurately performs important computations and approximates others. In order to make AC circuits practical, we need to determine which computation is how important carefully, and thus need to appropriately approximate the redundant computation for maintaining the required computational quality. In this paper, we focus on the importance of computations at the flip-flop (FF) level and propose a novel importance evaluation methodology. The key idea of the proposed methodology is a two-step fault injection algorithm to extract the near-optimal set of redundant FFs in the circuit. In the first step, the proposed methodology performs the FI simulation for each FF and extracts the candidates of redundant FFs. Then, in the second step, the proposed methodology extracts the set of redundant FFs in a binary search manner. Thanks to the two-step strategy, the proposed algorithm reduces the complexity of architecture exploration from an exponential order to a linear order without understanding the functionality and behavior of the target application program. Experimental results show that the proposed methodology identifies the candidates of redundant FFs depending on the given constraints. In a case study of an image processing accelerator, the truncation for identified redundant FFs reduces the circuit area by 29.6% and saves power dissipation by 44.8% under the ASIC implementation while satisfying the PSNR constraint. Similarly, the dynamic power dissipation is saved by 47.2% under the FPGA implementation.
Shogo CHIWAKI Ryutaroh MATSUMOTO
Stabilizer-based quantum secret sharing has two methods to reconstruct a quantum secret: The erasure correcting procedure and the unitary procedure. It is known that the unitary procedure has a smaller circuit width. On the other hand, it is unknown which method has smaller depth and fewer circuit gates. In this letter, it is shown that the unitary procedure has smaller depth and fewer circuit gates than the erasure correcting procedure which follows a standard framework performing measurements and unitary operators according to the measurements outcomes, when the circuits are designed for quantum secret sharing using the [[5, 1, 3]] binary stabilizer code. The evaluation can be reversed if one discovers a better circuit for the erasure correcting procedure which does not follow the standard framework.
Yuta NAKAHARA Toshiyasu MATSUSHIMA
Previously, we proposed a probabilistic data generation model represented by an unobservable tree and a sequential updating method to calculate a posterior distribution over a set of trees. The set is called a meta-tree. In this paper, we propose a more efficient batch updating method.