The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] MPO(945hit)

161-180hit(945hit)

  • A Balanced Decision Tree Based Heuristic for Linear Decomposition of Index Generation Functions

    Shinobu NAGAYAMA  Tsutomu SASAO  Jon T. BUTLER  

     
    PAPER-Logic Design

      Pubricized:
    2017/05/19
      Vol:
    E100-D No:8
      Page(s):
    1583-1591

    Index generation functions model content-addressable memory, and are useful in virus detectors and routers. Linear decompositions yield simpler circuits that realize index generation functions. This paper proposes a balanced decision tree based heuristic to efficiently design linear decompositions for index generation functions. The proposed heuristic finds a good linear decomposition of an index generation function by using appropriate cost functions and a constraint to construct a balanced tree. Since the proposed heuristic is fast and requires a small amount of memory, it is applicable even to large index generation functions that cannot be solved in a reasonable time by existing heuristics. This paper shows time and space complexities of the proposed heuristic, and experimental results using some large examples to show its efficiency.

  • Improved Multiple Impossible Differential Cryptanalysis of Midori128

    Mohamed TOLBA  Ahmed ABDELKHALEK  Amr M. YOUSSEF  

     
    LETTER-Cryptography and Information Security

      Vol:
    E100-A No:8
      Page(s):
    1733-1737

    Midori128 is a lightweight block cipher proposed at ASIACRYPT 2015 to achieve low energy consumption per bit. Currently, the best published impossible differential attack on Midori128 covers 10 rounds without the pre-whitening key. By exploiting the special structure of the S-boxes and the binary linear transformation layer in Midori128, we present impossible differential distinguishers that cover 7 full rounds including the mix column operations. Then, we exploit four of these distinguishers to launch multiple impossible differential attack against 11 rounds of the cipher with the pre-whitening and post-whitening keys.

  • Semi-Supervised Speech Enhancement Combining Nonnegative Matrix Factorization and Robust Principal Component Analysis

    Yonggang HU  Xiongwei ZHANG  Xia ZOU  Meng SUN  Yunfei ZHENG  Gang MIN  

     
    LETTER-Speech and Hearing

      Vol:
    E100-A No:8
      Page(s):
    1714-1719

    Nonnegative matrix factorization (NMF) is one of the most popular machine learning tools for speech enhancement. The supervised NMF-based speech enhancement is accomplished by updating iteratively with the prior knowledge of the clean speech and noise spectra bases. However, in many real-world scenarios, it is not always possible for conducting any prior training. The traditional semi-supervised NMF (SNMF) version overcomes this shortcoming while the performance degrades. In this letter, without any prior knowledge of the speech and noise, we present an improved semi-supervised NMF-based speech enhancement algorithm combining techniques of NMF and robust principal component analysis (RPCA). In this approach, fixed speech bases are obtained from the training samples chosen from public dateset offline. The noise samples used for noise bases training, instead of characterizing a priori as usual, can be obtained via RPCA algorithm on the fly. This letter also conducts a study on the assumption whether the time length of the estimated noise samples may have an effect on the performance of the algorithm. Three metrics, including PESQ, SDR and SNR are applied to evaluate the performance of the algorithms by making experiments on TIMIT with 20 noise types at various signal-to-noise ratio levels. Extensive experimental results demonstrate the superiority of the proposed algorithm over the competing speech enhancement algorithm.

  • A Fast Updatable Implementation of Index Generation Functions Using Multiple IGUs

    Tsutomu SASAO  

     
    PAPER-Logic Design

      Pubricized:
    2017/05/19
      Vol:
    E100-D No:8
      Page(s):
    1574-1582

    This paper presents a method to realize index generation functions using multiple Index Generation Units (IGUs). The architecture implements index generation functions more efficiently than a single IGU when the number of registered vectors is very large. This paper proves that independent linear transformations are necessary in IGUs for efficient realization. Experimental results confirm this statement. Finally, it shows a fast update method to IGUs.

  • A New Bayesian Network Structure Learning Algorithm Mechanism Based on the Decomposability of Scoring Functions

    Guoliang LI  Lining XING  Zhongshan ZHANG  Yingwu CHEN  

     
    PAPER-Graphs and Networks

      Vol:
    E100-A No:7
      Page(s):
    1541-1551

    Bayesian networks are a powerful approach for representation and reasoning under conditions of uncertainty. Of the many good algorithms for learning Bayesian networks from data, the bio-inspired search algorithm is one of the most effective. In this paper, we propose a hybrid mutual information-modified binary particle swarm optimization (MI-MBPSO) algorithm. This technique first constructs a network based on MI to improve the quality of the initial population, and then uses the decomposability of the scoring function to modify the BPSO algorithm. Experimental results show that, the proposed hybrid algorithm outperforms various other state-of-the-art structure learning algorithms.

  • A Spatiotemporal Statistical Model for Eyeballs of Human Embryos

    Masashi KISHIMOTO  Atsushi SAITO  Tetsuya TAKAKUWA  Shigehito YAMADA  Hiroshi MATSUZOE  Hidekata HONTANI  Akinobu SHIMIZU  

     
    PAPER-Biological Engineering

      Pubricized:
    2017/04/17
      Vol:
    E100-D No:7
      Page(s):
    1505-1515

    During the development of a human embryo, the position of eyes moves medially and caudally in the viscerocranium. A statistical model of this process can play an important role in embryology by facilitating qualitative analyses of change. This paper proposes an algorithm to construct a spatiotemporal statistical model for the eyeballs of a human embryo. The proposed modeling algorithm builds a statistical model of the spatial coordinates of the eyeballs independently for each Carnegie stage (CS) by using principal component analysis (PCA). In the process, a q-Gaussian distribution with a model selection scheme based on the Aaike information criterion is used to handle a non-Gaussian distribution with a small sample size. Subsequently, it seamlessly interpolates the statistical models of neighboring CSs, and we present 10 interpolation methods. We also propose an estimation algorithm for the CS using our spatiotemporal statistical model. A set of images of eyeballs in human embryos from the Kyoto Collection was used to train the model and assess its performance. The modeling results suggested that information geometry-based interpolation under the assumption of a q-Gaussian distribution is the best modeling method. The average error in CS estimation was 0.409. We proposed an algorithm to construct a spatiotemporal statistical model of the eyeballs of a human embryo and tested its performance using the Kyoto Collection.

  • Particle Filter Target Tracking Algorithm Based on Dynamic Niche Genetic Algorithm

    Weicheng XIE  Junxu WEI  Zhichao CHEN  Tianqian LI  

     
    PAPER-Vision

      Vol:
    E100-A No:6
      Page(s):
    1325-1332

    Particle filter algorithm is an important algorithm in the field of target tracking. however, this algorithm faces the problem of sample impoverishment which is caused by the introduction of re-sampling and easily affected by illumination variation. This problem seriously affects the tracking performance of a particle filter algorithm. To solve this problem, we introduce a particle filter target tracking algorithm based on a dynamic niche genetic algorithm. The application of this dynamic niche genetic algorithm to re-sampling ensures particle diversity and dynamically fuses the color and profile features of the target in order to increase the algorithm accuracy under the illumination variation. According to the test results, the proposed algorithm accurately tracks the target, significantly increases the number of particles, enhances the particle diversity, and exhibits better robustness and better accuracy.

  • A Novel 3D Gradient LBP Descriptor for Action Recognition

    Zhaoyang GUO  Xin'an WANG  Bo WANG  Zheng XIE  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2017/03/02
      Vol:
    E100-D No:6
      Page(s):
    1388-1392

    In the field of action recognition, Spatio-Temporal Interest Points (STIPs)-based features have shown high efficiency and robustness. However, most of state-of-the-art work to describe STIPs, they typically focus on 2-dimensions (2D) images, which ignore information in 3D spatio-temporal space. Besides, the compact representation of descriptors should be considered due to the costs of storage and computational time. In this paper, a novel local descriptor named 3D Gradient LBP is proposed, which extends the traditional descriptor Local Binary Patterns (LBP) into 3D spatio-temporal space. The proposed descriptor takes advantage of the neighbourhood information of cuboids in three dimensions, which accounts for its excellent descriptive power for the distribution of grey-level space. Experiments on three challenging datasets (KTH, Weizmann and UT Interaction) validate the effectiveness of our approach in the recognition of human actions.

  • A Noise Inference Method Based on Fast Context-Aware Tensor Decomposition

    Qingfu FAN  Lei ZHANG  Wen LI  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2017/03/08
      Vol:
    E100-D No:6
      Page(s):
    1360-1363

    Existing noise inference algorithms neglected the smooth characteristics of noise data, which results in executing slowly of noise inference. In order to address this problem, we present a noise inference algorithm based on fast context-aware tensor decomposition (F-CATD). F-CATD improves the noise inference algorithm based on context-aware tensor decomposition algorithm. It combines the smoothness constraint with context-aware tensor decomposition to speed up the process of decomposition. Experiments with New York City 311 noise data show that the proposed method accelerates the noise inference. Compared with the existing method, F-CATD reduces 4-5 times in terms of time consumption while keeping the effectiveness of the results.

  • Traffic Anomaly Detection Based on Robust Principal Component Analysis Using Periodic Traffic Behavior

    Takahiro MATSUDA  Tatsuya MORITA  Takanori KUDO  Tetsuya TAKINE  

     
    PAPER-Network

      Pubricized:
    2016/11/21
      Vol:
    E100-B No:5
      Page(s):
    749-761

    In this paper, we study robust Principal Component Analysis (PCA)-based anomaly detection techniques in network traffic, which can detect traffic anomalies by projecting measured traffic data onto a normal subspace and an anomalous subspace. In a PCA-based anomaly detection, outliers, anomalies with excessively large traffic volume, may contaminate the subspaces and degrade the performance of the detector. To solve this problem, robust PCA methods have been studied. In a robust PCA-based anomaly detection scheme, outliers can be removed from the measured traffic data before constructing the subspaces. Although the robust PCA methods are promising, they incure high computational cost to obtain the optimal location vector and scatter matrix for the subspace. We propose a novel anomaly detection scheme by extending the minimum covariance determinant (MCD) estimator, a robust PCA method. The proposed scheme utilizes the daily periodicity in traffic volume and attempts to detect anomalies for every period of measured traffic. In each period, before constructing the subspace, outliers are removed from the measured traffic data by using a location vector and a scatter matrix obtained in the preceding period. We validate the proposed scheme by applying it to measured traffic data in the Abiline network. Numerical results show that the proposed scheme provides robust anomaly detection with less computational cost.

  • XY-Separable Scale-Space Filtering by Polynomial Representations and Its Applications Open Access

    Gou KOUTAKI  Keiichi UCHIMURA  

     
    INVITED PAPER

      Pubricized:
    2017/01/11
      Vol:
    E100-D No:4
      Page(s):
    645-654

    In this paper, we propose the application of principal component analysis (PCA) to scale-spaces. PCA is a standard method used in computer vision. Because the translation of an input image into scale-space is a continuous operation, it requires the extension of conventional finite matrix-based PCA to an infinite number of dimensions. Here, we use spectral theory to resolve this infinite eigenvalue problem through the use of integration, and we propose an approximate solution based on polynomial equations. In order to clarify its eigensolutions, we apply spectral decomposition to Gaussian scale-space and scale-normalized Laplacian of Gaussian (sLoG) space. As an application of this proposed method, we introduce a method for generating Gaussian blur images and sLoG images, demonstrating that the accuracy of such an image can be made very high by using an arbitrary scale calculated through simple linear combination. Furthermore, to make the scale-space filtering efficient, we approximate the basis filter set using Gaussian lobes approximation and we can obtain XY-Separable filters. As a more practical example, we propose a new Scale Invariant Feature Transform (SIFT) detector.

  • Microblog Retrieval Using Ensemble of Feature Sets through Supervised Feature Selection

    Abu Nowshed CHY  Md Zia ULLAH  Masaki AONO  

     
    PAPER

      Pubricized:
    2017/01/17
      Vol:
    E100-D No:4
      Page(s):
    793-806

    Microblog, especially twitter, has become an integral part of our daily life for searching latest news and events information. Due to the short length characteristics of tweets and frequent use of unconventional abbreviations, content-relevance based search cannot satisfy user's information need. Recent research has shown that considering temporal and contextual aspects in this regard has improved the retrieval performance significantly. In this paper, we focus on microblog retrieval, emphasizing the alleviation of the vocabulary mismatch, and the leverage of the temporal (e.g., recency and burst nature) and contextual characteristics of tweets. To address the temporal and contextual aspect of tweets, we propose new features based on query-tweet time, word embedding, and query-tweet sentiment correlation. We also introduce some popularity features to estimate the importance of a tweet. A three-stage query expansion technique is applied to improve the relevancy of tweets. Moreover, to determine the temporal and sentiment sensitivity of a query, we introduce query type determination techniques. After supervised feature selection, we apply random forest as a feature ranking method to estimate the importance of selected features. Then, we make use of ensemble of learning to rank (L2R) framework to estimate the relevance of query-tweet pair. We conducted experiments on TREC Microblog 2011 and 2012 test collections over the TREC Tweets2011 corpus. Experimental results demonstrate the effectiveness of our method over the baseline and known related works in terms of precision at 30 (P@30), mean average precision (MAP), normalized discounted cumulative gain at 30 (NDCG@30), and R-precision (R-Prec) metrics.

  • An Improved Multivariate Wavelet Denoising Method Using Subspace Projection

    Huan HAO  Huali WANG  Naveed ur REHMAN  Liang CHEN  Hui TIAN  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:3
      Page(s):
    769-775

    An improved multivariate wavelet denoising algorithm combined with subspace and principal component analysis is presented in this paper. The key element is deriving an optimal orthogonal matrix that can project the multivariate observation signal to a signal subspace from observation space. Univariate wavelet shrinkage operator is then applied to the projected signals channel-wise resulting in the improvement of the output SNR. Finally, principal component analysis is performed on the denoised signal in the observation space to further improve the denoising performance. Experimental results based on synthesized and real world ECG data verify the effectiveness of the proposed algorithm.

  • Human Wearable Attribute Recognition Using Probability-Map-Based Decomposition of Thermal Infrared Images

    Brahmastro KRESNARAMAN  Yasutomo KAWANISHI  Daisuke DEGUCHI  Tomokazu TAKAHASHI  Yoshito MEKADA  Ichiro IDE  Hiroshi MURASE  

     
    PAPER-Image

      Vol:
    E100-A No:3
      Page(s):
    854-864

    This paper addresses the attribute recognition problem, a field of research that is dominated by studies in the visible spectrum. Only a few works are available in the thermal spectrum, which is fundamentally different from the visible one. This research performs recognition specifically on wearable attributes, such as glasses and masks. Usually these attributes are relatively small in size when compared with the human body, on top of a large intra-class variation of the human body itself, therefore recognizing them is not an easy task. Our method utilizes a decomposition framework based on Robust Principal Component Analysis (RPCA) to extract the attribute information for recognition. However, because it is difficult to separate the body and the attributes without any prior knowledge, noise is also extracted along with attributes, hampering the recognition capability. We made use of prior knowledge; namely the location where the attribute is likely to be present. The knowledge is referred to as the Probability Map, incorporated as a weight in the decomposition by RPCA. Using the Probability Map, we achieve an attribute-wise decomposition. The results show a significant improvement with this approach compared to the baseline, and the proposed method achieved the highest performance in average with a 0.83 F-score.

  • Signal Reconstruction Algorithm of Finite Rate of Innovation with Matrix Pencil and Principal Component Analysis

    Yujie SHI  Li ZENG  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:3
      Page(s):
    761-768

    In this paper, we study the problem of noise with regard to the perfect reconstruction of non-bandlimited signals, the class of signals having a finite number of degrees of freedom per unit time. The finite rate of innovation (FRI) method provides a means of recovering a non-bandlimited signal through using of appropriate kernels. In the presence of noise, however, the reconstruction function of this scheme may become ill-conditioned. Further, the reduced sampling rates afforded by this scheme can be accompanied by increased error sensitivity. In this paper, to obtain improved noise robustness, we propose the matrix pencil (MP) method for sample signal reconstruction, which is based on principal component analysis (PCA). Through the selection of an adaptive eigenvalue, a non-bandlimited signal can be perfectly reconstructed via a stable solution of the Yule-Walker equation. The proposed method can obtain a high signal-to-noise-ratio (SNR) for the reconstruction results. Herein, the method is applied to certain non-bandlimited signals, such as a stream of Diracs and nonuniform splines. The simulation results demonstrate that the MP and PCA are more effective than the FRI method in suppressing noise. The FRI method can be used in many applications, including those related to bioimaging, radar, and ultrasound imaging.

  • Dry Etching Technologies of Optical Device and III-V Compound Semiconductors Open Access

    Ryuichiro KAMIMURA  Kanji FURUTA  

     
    INVITED PAPER

      Vol:
    E100-C No:2
      Page(s):
    150-155

    Dry etching is one of the elemental technologies for the fabrication of optical devices. In order to obtain the desired shape using the dry etching process, it is necessary to understand the reactivity of the materials being used to plasma. In particular, III-V compound semiconductors have a multi-layered structure comprising a plurality of elements and thus it is important to first have a full understanding of the basic trends of plasma dry etching, the plasma type and the characteristics of etching plasma sources. In this paper, III-V compound semiconductor etching for use in light sources such as LDs and LEDs, will be described. Glass, LN and LT used in the formation of waveguides and MLA will be introduced as well. And finally, the future prospects of dry etching will be described briefly.

  • A Video Salient Region Detection Framework Using Spatiotemporal Consistency Optimization

    Yunfei ZHENG  Xiongwei ZHANG  Lei BAO  Tieyong CAO  Yonggang HU  Meng SUN  

     
    PAPER-Image

      Vol:
    E100-A No:2
      Page(s):
    688-701

    Labeling a salient region accurately in video with cluttered background and complex motion condition is still a challenging work. Most existing video salient region detection models mainly extract the stimulus-driven saliency features to detect the salient region in video. They are easily influenced by the cluttered background and complex motion conditions. It may lead to incomplete or wrong detection results. In this paper, we propose a video salient region detection framework by fusing the stimulus-driven saliency features and spatiotemporal consistency cue to improve the performance of detection under these complex conditions. On one hand, stimulus-driven spatial saliency features and temporal saliency features are extracted effectively to derive the initial spatial and temporal salient region map. On the other hand, in order to make use of the spatiotemporal consistency cue, an effective spatiotemporal consistency optimization model is presented. We use this model optimize the initial spatial and temporal salient region map. Then the superpixel-level spatiotemporal salient region map is derived by optimizing the initial spatiotemporal salient region map. Finally, the pixel-level spatiotemporal salient region map is derived by solving a self-defined energy model. Experimental results on the challenging video datasets demonstrate that the proposed video salient region detection framework outperforms state-of-the-art methods.

  • Reciprocity Theorems and Their Application to Numerical Analysis in Grating Theory

    Junichi NAKAYAMA  Yasuhiko TAMURA  

     
    PAPER

      Vol:
    E100-C No:1
      Page(s):
    3-10

    This paper deals with the diffraction of a monochromatic plane wave by a periodic grating. We discuss a problem how to obtain a numerical diffraction efficiency (NDE) satisfying the reciprocity theorem for diffraction efficiencies, because diffraction efficiencies are the subject of the diffraction theories. First, this paper introduces a new formula that decomposes an NDE into two components: the even component and the odd one. The former satisfies the reciprocity theorem for diffraction efficiencies, but the latter does not. Therefore, the even component of an NDE becomes an answer to our problem. On the other hand, the odd component of an NDE represents an unwanted error. Using such the decomposition formula, we then obtain another new formula that decomposes the conventional energy error into two components. One is the energy error made by even components of NDE's. The other is the energy error constructed by unwanted odd ones and it may be used as a reciprocity criterion of a numerical solution. This decomposition formula shows a drawback of the conventional energy balance. The total energy error is newly introduced as a more strict condition for a desirable solution. We point out theoretically that the reciprocal wave solution, an approximate solution satisfying the reciprocity for wave fields, gives another solution to our problem. Numerical examples are given for the diffraction of a TM plane wave by a very rough periodic surface with perfect conductivity. In the case of a numerical solution by the image integral equation of the second kind, we found that the energy error is much reduced by use of the even component of an NDE as an approximate diffraction efficiency or by use of a reciprocal wave solution.

  • Initial Value Problem Formulation TDBEM with 4-D Domain Decomposition Method and Application to Wake Fields Analysis

    Hideki KAWAGUCHI  Thomas WEILAND  

     
    PAPER

      Vol:
    E100-C No:1
      Page(s):
    37-44

    The Time Domain Boundary Element Method (TDBEM) has its advantages in the analysis of transient electromagnetic fields (wake fields) induced by a charged particle beam with curved trajectory in a particle accelerator. On the other hand, the TDBEM has disadvantages of huge required memory and computation time compared with those of the Finite Difference Time Domain (FDTD) method or the Finite Integration Technique (FIT). This paper presents a comparison of the FDTD method and 4-D domain decomposition method of the TDBEM based on an initial value problem formulation for the curved trajectory electron beam, and application to a full model simulation of the bunch compressor section of the high-energy particle accelerators.

  • Prefiltering and Postfiltering Based on Global Motion Compensation for Improving Coding Efficiency in H.264 and HEVC Codecs

    Ho Hyeong RYU  Kwang Yeon CHOI  Byung Cheol SONG  

     
    PAPER-Image Processing and Video Processing

      Pubricized:
    2016/10/07
      Vol:
    E100-D No:1
      Page(s):
    160-165

    In this paper, we propose a filtering approach based on global motion estimation (GME) and global motion compensation (GMC) for pre- and postprocessing of video codecs. For preprocessing a video codec, group of pictures (GOP), which is a basic unit for GMC, and reference frames are first defined for an input video sequence. Next, GME and GMC are sequentially performed for every frame in each GOP. Finally, a block-based adaptive temporal filter is applied between the GMC frames before video encoding. For postprocessing a video codec at the decoder end, every decoded frame is inversely motion-compensated using the transmitted global motion information. The holes generated during inverse motion compensation can be filled with the reference frames. The experimental results show that the proposed algorithm provides higher Bjontegaard-delta peak signal-to-noise ratios (BD-PSNRs) of 0.63 and 0.57 dB on an average compared with conventional H.264 and HEVC platforms, respectively.

161-180hit(945hit)