The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] MPO(945hit)

41-60hit(945hit)

  • An Improved Adaptive Algorithm for Locating Faulty Interactions in Combinatorial Testing Open Access

    Qianqian YANG  Xiao-Nan LU  

     
    PAPER-Algorithms and Data Structures

      Pubricized:
    2021/11/29
      Vol:
    E105-A No:6
      Page(s):
    930-942

    Combinatorial testing is an effective testing technique for detecting faults in a software or hardware system with multiple factors using combinatorial methods. By performing a test, which is an assignment of possible values to all the factors, and verifying whether the system functions as expected (pass) or not (fail), the presence of faults can be detected. The failures of the tests are possibly caused by combinations of multiple factors assigned with specific values, called faulty interactions. Martínez et al. [1] proposed the first deterministic adaptive algorithm for discovering faulty interactions involving at most two factors where each factor has two values, for which graph representations are adopted. In this paper, we improve Martínez et al.'s algorithm by an adaptive algorithmic approach for discovering faulty interactions in the so-called “non-2-locatable” graphs. We show that, for any system where each “non-2-locatable factor-component” involves two faulty interactions (for example, a system having at most two faulty interactions), our improved algorithm efficiently discovers all the faulty interactions with an extremely low mistaken probability caused by the random selection process in Martínez et al.'s algorithm. The effectiveness of our improved algorithm are revealed by both theoretical discussions and experimental evaluations.

  • Number of Failed Components in Consecutive-k-out-of-n:G Systems and Their Applications in Optimization Problems

    Lei ZHOU  Hisashi YAMAMOTO  

     
    PAPER-Reliability, Maintainability and Safety Analysis

      Pubricized:
    2021/12/16
      Vol:
    E105-A No:6
      Page(s):
    943-951

    In this paper, we study the number of failed components in a consecutive-k-out-of-n:G system. The distributions and expected values of the number of failed components when system is failed or working at a particular time t are evaluated. We also apply them to the optimization problems concerned with the optimal number of components and the optimal replacement time. Finally, we present the illustrative examples for the expected number of failed components and give the numerical results for the optimization problems.

  • A Cost-Sensitive Golden Chip-Free Hardware Trojan Detection Using Principal Component Analysis and Naïve Bayes Classification Algorithm

    Yanjiang LIU  Xianzhao XIA  Jingxin ZHONG  Pengfei GUO  Chunsheng ZHU  Zibin DAI  

     
    PAPER-Cryptography and Information Security

      Pubricized:
    2021/12/03
      Vol:
    E105-A No:6
      Page(s):
    965-974

    Side-channel analysis is one of the most investigated hardware Trojan detection approaches. However, nearly all the side-channel analysis approaches require golden chips for reference, which are hard to obtain actually. Besides, majority of existing Trojan detection algorithms focus on the data similarity and ignore the Trojan misclassification during the detection. In this paper, we propose a cost-sensitive golden chip-free hardware Trojan detection framework, which aims to minimize the probability of Trojan misclassification during the detection. The post-layout simulation data of voltage variations at different process corners is utilized as a golden reference. Further, a classification algorithm based on the combination of principal component analysis and Naïve bayes is exploited to identify the existence of hardware Trojan with a minimum misclassification risk. Experimental results on ASIC demonstrate that the proposed approach improves the detection accuracy ratio compared with the three detection algorithms and distinguishes the Trojan with only 0.27% area occupies even under ±15% process variations.

  • Performance Analysis on the Uplink of Massive MIMO Systems with Superimposed Pilots and Arbitrary-Bit ADCs

    Chen CHEN  Wence ZHANG  Xu BAO  Jing XIA  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2021/10/28
      Vol:
    E105-B No:5
      Page(s):
    629-637

    This paper studies the performance of quantized massive multiple-input multiple-output (MIMO) systems with superimposed pilots (SP), using linear minimum mean-square-error (LMMSE) channel estimation and maximum ratio combining (MRC) detection. In contrast to previous works, arbitrary-bit analog-to-digital converters (ADCs) are considered. We derive an accurate approximation of the uplink achievable rate considering the removal of estimated pilots. Based on the analytical expression, the optimal pilot power factor that maximizes the achievable rate is deduced and an expression for energy efficiency (EE) is given. In addition, the achievable rate and the optimal power allocation policy under some asymptotic limits are analyzed. Analysis shows that the systems with higher-resolution ADCs or larger number of base station (BS) antennas need to allocate more power to pilots. In contrast, more power needs to be allocated to data when the channel is slowly varying. Numerical results show that in the low signal-to-noise ratio (SNR) region, for 1-bit quantizers, SP outperforms time-multiplexed pilots (TP) in most cases, while for systems with higher-resolution ADCs, the SP scheme is suitable for the scenarios with comparatively small number of BS antennas and relatively long channel coherence time.

  • Anomaly Detection Using Spatio-Temporal Context Learned by Video Clip Sorting

    Wen SHAO  Rei KAWAKAMI  Takeshi NAEMURA  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2022/02/08
      Vol:
    E105-D No:5
      Page(s):
    1094-1102

    Previous studies on anomaly detection in videos have trained detectors in which reconstruction and prediction tasks are performed on normal data so that frames on which their task performance is low will be detected as anomalies during testing. This paper proposes a new approach that involves sorting video clips, by using a generative network structure. Our approach learns spatial contexts from appearances and temporal contexts from the order relationship of the frames. Experiments were conducted on four datasets, and we categorized the anomalous sequences by appearance and motion. Evaluations were conducted not only on each total dataset but also on each of the categories. Our method improved detection performance on both anomalies with different appearance and different motion from normality. Moreover, combining our approach with a prediction method produced improvements in precision at a high recall.

  • SVM Based Intrusion Detection Method with Nonlinear Scaling and Feature Selection

    Fei ZHANG  Peining ZHEN  Dishan JING  Xiaotang TANG  Hai-Bao CHEN  Jie YAN  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2022/02/14
      Vol:
    E105-D No:5
      Page(s):
    1024-1038

    Intrusion is one of major security issues of internet with the rapid growth in smart and Internet of Thing (IoT) devices, and it becomes important to detect attacks and set out alarm in IoT systems. In this paper, the support vector machine (SVM) and principal component analysis (PCA) based method is used to detect attacks in smart IoT systems. SVM with nonlinear scheme is used for intrusion classification and PCA is adopted for feature selection on the training and testing datasets. Experiments on the NSL-KDD dataset show that the test accuracy of the proposed method can reach 82.2% with 16 features selected from PCA for binary-classification which is almost the same as the result obtained with all the 41 features; and the test accuracy can achieve 78.3% with 29 features selected from PCA for multi-classification while 79.6% without feature selection. The Denial of Service (DoS) attack detection accuracy of the proposed method can achieve 8.8% improvement compared with existing artificial neural network based method.

  • Low-Complexity VBI-Based Channel Estimation for Massive MIMO Systems

    Chen JI  Shun WANG  Haijun FU  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2021/11/11
      Vol:
    E105-B No:5
      Page(s):
    600-607

    This paper proposes a low-complexity variational Bayesian inference (VBI)-based method for massive multiple-input multiple-output (MIMO) downlink channel estimation. The temporal correlation at the mobile user side is jointly exploited to enhance the channel estimation performance. The key to the success of the proposed method is the column-independent factorization imposed in the VBI framework. Since we separate the Bayesian inference for each column vector of signal-of-interest, the computational complexity of the proposed method is significantly reduced. Moreover, the temporal correlation is automatically uncoupled to facilitate the updating rule derivation for the temporal correlation itself. Simulation results illustrate the substantial performance improvement achieved by the proposed method.

  • Optimal Control of Timed Petri Nets Under Temporal Logic Constraints with Generalized Mutual Exclusion

    Kohei FUJITA  Toshimitsu USHIO  

     
    PAPER

      Pubricized:
    2021/10/13
      Vol:
    E105-A No:5
      Page(s):
    808-815

    This paper presents a novel method for optimal control of timed Petri nets, introducing a novel temporal logic based constraint called a generalized mutual exclusion temporal constraint (GMETC). The GMETC is described by a metric temporal logic (MTL) formula where each atomic proposition represents a generalized mutual exclusion constraint (GMEC). We formulate an optimal control problem of the timed Petri nets under a given GMETC and solve the problem by transforming it into an integer linear programming problem where the MTL formula is encoded by linear inequalities. We show the effectiveness of the proposed approach by a numerical simulation.

  • Efficient Computation of Betweenness Centrality by Graph Decompositions and Their Applications to Real-World Networks

    Tatsuya INOHA  Kunihiko SADAKANE  Yushi UNO  Yuma YONEBAYASHI  

     
    PAPER

      Pubricized:
    2021/11/08
      Vol:
    E105-D No:3
      Page(s):
    451-458

    Betweenness centrality is one of the most significant and commonly used centralities, where centrality is a notion of measuring the importance of nodes in networks. In 2001, Brandes proposed an algorithm for computing betweenness centrality efficiently, and it can compute those values for all nodes in O(nm) time for unweighted networks, where n and m denote the number of nodes and links in networks, respectively. However, even Brandes' algorithm is not fast enough for recent large-scale real-world networks, and therefore, much faster algorithms are expected. The objective of this research is to theoretically improve the efficiency of Brandes' algorithm by introducing graph decompositions, and to verify the practical effectiveness of our approaches by implementing them as computer programs and by applying them to various kinds of real-world networks. A series of computational experiments shows that our proposed algorithms run several times faster than the original Brandes' algorithm, which are guaranteed by theoretical analyses.

  • Improving Practical UC-Secure Commitments based on the DDH Assumption

    Eiichiro FUJISAKI  

     
    PAPER

      Pubricized:
    2021/10/05
      Vol:
    E105-A No:3
      Page(s):
    182-194

    At Eurocrypt 2011, Lindell presented practical static and adaptively UC-secure commitment schemes based on the DDH assumption. Later, Blazy et al. (at ACNS 2013) improved the efficiency of the Lindell's commitment schemes. In this paper, we present static and adaptively UC-secure commitment schemes based on the same assumption and further improve the communication and computational complexity, as well as the size of the common reference string.

  • A Novel Method for Adaptive Beamforming under the Strong Interference Condition

    Zongli RUAN  Hongshu LIAO  Guobing QIAN  

     
    LETTER-Digital Signal Processing

      Pubricized:
    2021/08/02
      Vol:
    E105-A No:2
      Page(s):
    109-113

    In this letter, firstly, a novel adaptive beamformer using independent component analysis (ICA) algorithm is proposed. By this algorithm, the ambiguity of amplitude and phase resulted from blind source separation is removed utilizing the special structure of array manifolds matrix. However, there might exist great calibration error when the powers of interferences are far larger than that of desired signal at many applications such as sonar, radio astronomy, biomedical engineering and earthquake detection. As a result, this will lead to a significant reduction in separation performance. Then, a new method based on the combination of ICA and primary component analysis (PCA) is proposed to recover the desired signal's amplitude under strong interference. Finally, computer simulation is carried out to indicate the effectiveness of our methods. The simulation results show that the proposed methods can obtain higher SNR and more accurate power estimation of desired signal than diagonal loading sample matrix inversion (LSMI) and worst-case performance optimization (WCPO) method.

  • SimpleZSL: Extremely Simple and Fast Zero-Shot Learning with Nearest Neighbor Classifiers

    Masayuki HIROMOTO  Hisanao AKIMA  Teruo ISHIHARA  Takuji YAMAMOTO  

     
    PAPER-Pattern Recognition

      Pubricized:
    2021/10/29
      Vol:
    E105-D No:2
      Page(s):
    396-405

    Zero-shot learning (ZSL) aims to classify images of unseen classes by learning relationship between visual and semantic features. Existing works have been improving recognition accuracy from various approaches, but they employ computationally intensive algorithms that require iterative optimization. In this work, we revisit the primary approach of the pattern recognition, ı.e., nearest neighbor classifiers, to solve the ZSL task by an extremely simple and fast way, called SimpleZSL. Our algorithm consists of the following three simple techniques: (1) just averaging feature vectors to obtain visual prototypes of seen classes, (2) calculating a pseudo-inverse matrix via singular value decomposition to generate visual features of unseen classes, and (3) inferring unseen classes by a nearest neighbor classifier in which cosine similarity is used to measure distance between feature vectors. Through the experiments on common datasets, the proposed method achieves good recognition accuracy with drastically small computational costs. The execution time of the proposed method on a single CPU is more than 100 times faster than those of the GPU implementations of the existing methods with comparable accuracies.

  • A Robust Canonical Polyadic Tensor Decomposition via Structured Low-Rank Matrix Approximation

    Riku AKEMA  Masao YAMAGISHI  Isao YAMADA  

     
    PAPER-Digital Signal Processing

      Pubricized:
    2021/06/23
      Vol:
    E105-A No:1
      Page(s):
    11-24

    The Canonical Polyadic Decomposition (CPD) is the tensor analog of the Singular Value Decomposition (SVD) for a matrix and has many data science applications including signal processing and machine learning. For the CPD, the Alternating Least Squares (ALS) algorithm has been used extensively. Although the ALS algorithm is simple, it is sensitive to a noise of a data tensor in the applications. In this paper, we propose a novel strategy to realize the noise suppression for the CPD. The proposed strategy is decomposed into two steps: (Step 1) denoising the given tensor and (Step 2) solving the exact CPD of the denoised tensor. Step 1 can be realized by solving a structured low-rank approximation with the Douglas-Rachford splitting algorithm and then Step 2 can be realized by solving the simultaneous diagonalization of a matrix tuple constructed by the denoised tensor with the DODO method. Numerical experiments show that the proposed algorithm works well even in typical cases where the ALS algorithm suffers from the so-called bottleneck/swamp effect.

  • Weighted PCA-LDA Based Color Quantization Method Suppressing Saturation Decrease

    Seiichi KOJIMA  Momoka HARADA  Yoshiaki UEDA  Noriaki SUETAKE  

     
    LETTER-Image

      Pubricized:
    2021/06/02
      Vol:
    E104-A No:12
      Page(s):
    1728-1732

    In this letter, we propose a new color quantization method suppressing saturation decrease. In the proposed method, saturation-based weight and intensity-based weight are used so that vivid colors are selected as the representative colors preferentially. Experiments show that the proposed method tends to select vivid colors even if they occupy only a small area in the image.

  • Monocone Antenna with Short Elements on Wideband Choke Structure Using Composite Right/Left-Handed Coaxial Line

    Kazuya MATSUBAYASHI  Naobumi MICHISHITA  Hisashi MORISHITA  

     
    PAPER-Antennas and Propagation

      Pubricized:
    2021/06/01
      Vol:
    E104-B No:11
      Page(s):
    1408-1418

    The composite right/left-handed (CRLH) coaxial line (CL) with wideband electromagnetic band gap (EBG) is applied to the wideband choke structure for a monocone antenna with short elements, and the resulting characteristics are considered. In the proposed antenna, impedance matching and leakage current suppression can be achieved across a wideband off. The lowest frequency (|S11| ≤ -10dB) of the proposed antenna is about the same as that of the monocone antenna on an infinite ground plane. In addition, the radiation patterns of the proposed antenna are close to the figure of eight in wideband. The proposed antenna is prototyped, and the validity of the simulation is verified through measurement.

  • Influence of Access to Reading Material during Concept Map Recomposition in Reading Comprehension and Retention

    Pedro GABRIEL FONTELES FURTADO  Tsukasa HIRASHIMA  Nawras KHUDHUR  Aryo PINANDITO  Yusuke HAYASHI  

     
    PAPER-Educational Technology

      Pubricized:
    2021/08/02
      Vol:
    E104-D No:11
      Page(s):
    1941-1950

    This study investigated the influence of reading time while building a closed concept map on reading comprehension and retention. It also investigated the effect of having access to the text during closed concept map creation on reading comprehension and retention. Participants from Amazon Mechanical Turk (N =101) read a text, took an after-text test, and took part in one of three conditions, “Map & Text”, “Map only”, and “Double Text”, took an after-activity test, followed by a two-week retention period and then one final delayed test. Analysis revealed that higher reading times were associated with better reading comprehension and better retention. Furthermore, when comparing “Map & Text” to the “Map only” condition, short-term reading comprehension was improved, but long-term retention was not improved. This suggests that having access to the text while building closed concept maps can improve reading comprehension, but long term learning can only be improved if students invest time accessing both the map and the text.

  • Health Indicator Estimation by Video-Based Gait Analysis

    Ruochen LIAO  Kousuke MORIWAKI  Yasushi MAKIHARA  Daigo MURAMATSU  Noriko TAKEMURA  Yasushi YAGI  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2021/07/09
      Vol:
    E104-D No:10
      Page(s):
    1678-1690

    In this study, we propose a method to estimate body composition-related health indicators (e.g., ratio of body fat, body water, and muscle, etc.) using video-based gait analysis. This method is more efficient than individual measurement using a conventional body composition meter. Specifically, we designed a deep-learning framework with a convolutional neural network (CNN), where the input is a gait energy image (GEI) and the output consists of the health indicators. Although a vast amount of training data is typically required to train network parameters, it is unfeasible to collect sufficient ground-truth data, i.e., pairs consisting of the gait video and the health indicators measured using a body composition meter for each subject. We therefore use a two-step approach to exploit an auxiliary gait dataset that contains a large number of subjects but lacks the ground-truth health indicators. At the first step, we pre-train a backbone network using the auxiliary dataset to output gait primitives such as arm swing, stride, the degree of stoop, and the body width — considered to be relevant to the health indicators. At the second step, we add some layers to the backbone network and fine-tune the entire network to output the health indicators even with a limited number of ground-truth data points of the health indicators. Experimental results show that the proposed method outperforms the other methods when training from scratch as well as when using an auto-encoder-based pre-training and fine-tuning approach; it achieves relatively high estimation accuracy for the body composition-related health indicators except for body fat-relevant ones.

  • Global Optimization Algorithm for Cloud Service Composition

    Hongwei YANG  Fucheng XUE  Dan LIU  Li LI  Jiahui FENG  

     
    PAPER-Computer System

      Pubricized:
    2021/06/30
      Vol:
    E104-D No:10
      Page(s):
    1580-1591

    Service composition optimization is a classic NP-hard problem. How to quickly select high-quality services that meet user needs from a large number of candidate services is a hot topic in cloud service composition research. An efficient second-order beetle swarm optimization is proposed with a global search ability to solve the problem of cloud service composition optimization in this study. First, the beetle antennae search algorithm is introduced into the modified particle swarm optimization algorithm, initialize the population bying using a chaotic sequence, and the modified nonlinear dynamic trigonometric learning factors are adopted to control the expanding capacity of particles and global convergence capability. Second, modified secondary oscillation factors are incorporated, increasing the search precision of the algorithm and global searching ability. An adaptive step adjustment is utilized to improve the stability of the algorithm. Experimental results founded on a real data set indicated that the proposed global optimization algorithm can solve web service composition optimization problems in a cloud environment. It exhibits excellent global searching ability, has comparatively fast convergence speed, favorable stability, and requires less time cost.

  • The Explicit Dual of Leander's Monomial Bent Function

    Yanjun LI  Haibin KAN  Jie PENG  Chik How TAN  Baixiang LIU  

     
    LETTER-Cryptography and Information Security

      Pubricized:
    2021/03/08
      Vol:
    E104-A No:9
      Page(s):
    1357-1360

    Permutation polynomials and their compositional inverses are crucial for construction of Maiorana-McFarland bent functions and their dual functions, which have the optimal nonlinearity for resisting against the linear attack on block ciphers and on stream ciphers. In this letter, we give the explicit compositional inverse of the permutation binomial $f(z)=z^{2^{r}+2}+alpha zinmathbb{F}_{2^{2r}}[z]$. Based on that, we obtain the dual of monomial bent function $f(x)={ m Tr}_1^{4r}(x^{2^{2r}+2^{r+1}+1})$. Our result suggests that the dual of f is not a monomial any more, and it is not always EA-equivalent to f.

  • Impossibility on the Schnorr Signature from the One-More DL Assumption in the Non-Programmable Random Oracle Model Open Access

    Masayuki FUKUMITSU  Shingo HASEGAWA  

     
    PAPER-Cryptography and Information Security

      Pubricized:
    2021/03/08
      Vol:
    E104-A No:9
      Page(s):
    1163-1174

    The Schnorr signature is one of the representative signature schemes and its security was widely discussed. In the random oracle model (ROM), it is provable from the DL assumption, whereas there is negative circumstantial evidence in the standard model. Fleischhacker, Jager, and Schröder showed that the tight security of the Schnorr signature is unprovable from a strong cryptographic assumption, such as the One-More DL (OM-DL) assumption and the computational and decisional Diffie-Hellman assumption, in the ROM via a generic reduction as long as the underlying cryptographic assumption holds. However, it remains open whether or not the impossibility of the provable security of the Schnorr signature from a strong assumption via a non-tight and reasonable reduction. In this paper, we show that the security of the Schnorr signature is unprovable from the OM-DL assumption in the non-programmable ROM as long as the OM-DL assumption holds. Our impossibility result is proven via a non-tight Turing reduction.

41-60hit(945hit)