The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] Z(5900hit)

941-960hit(5900hit)

  • Maximum Volume Constrained Graph Nonnegative Matrix Factorization for Facial Expression Recognition

    Viet-Hang DUONG  Manh-Quan BUI  Jian-Jiun DING  Bach-Tung PHAM  Pham The BAO  Jia-Ching WANG  

     
    LETTER-Image

      Vol:
    E100-A No:12
      Page(s):
    3081-3085

    In this work, two new proposed NMF models are developed for facial expression recognition. They are called maximum volume constrained nonnegative matrix factorization (MV_NMF) and maximum volume constrained graph nonnegative matrix factorization (MV_GNMF). They achieve sparseness from a larger simplicial cone constraint and the extracted features preserve the topological structure of the original images.

  • CyclicSRP - A Multivariate Encryption Scheme with a Partially Cyclic Public Key

    Dung Hoang DUONG  Albrecht PETZOLDT  Tsuyoshi TAKAGI  

     
    PAPER-Cryptography and Information Security

      Vol:
    E100-A No:12
      Page(s):
    2691-2698

    Multivariate Public Key Cryptography (MPKC) is one of the main candidates for secure communication in a post-quantum era. Recently, Yasuda and Sakurai proposed at ICICS 2015 a new multivariate encryption scheme called SRP, which offers efficient decryption, a small blow up factor between plaintext and ciphertext and resists all known attacks against multivariate schemes. However, similar to other MPKC schemes, the key sizes of SRP are quite large. In this paper we propose a technique to reduce the key size of the SRP scheme, which enables us to reduce the size of the public key by up to 54%. Furthermore, we can use the additional structure in the public key polynomials to speed up the encryption process of the scheme by up to 50%. We show by experiments that our modifications do not weaken the security of the scheme.

  • Study on LVRT of DFIG Based on Fuzzy-Neural D-STATCOM

    Xueqin ZHENG  Xiaoxiong CHEN  Tung-Chin PAN  

     
    PAPER-Systems and Control

      Vol:
    E100-A No:12
      Page(s):
    2948-2955

    This paper aims to improve the ability of low voltage ride through (LVRT) of doubly-fed induction generation (DFIG) under the asymmetric grid fault. The traditional rotor of the Crowbar device requires a large reactive support during the period of protection, which causes large fluctuations to the reactive power of the output grid while cut in and off for Crowbar. This case would influence the quality and efficiency of entire power system. In order to solve the fluctuation of reactive power and the stability of the wind power system, this paper proposes the coordinated control of the fuzzy-neural D-STATCOM and the rotor of the Crowbar. The simulation results show that the system has the performance of the rotor current with faster decay and faster dynamic response, high steady-state characteristic during the grid fault, which improve the ability of LVRT of DFIG.

  • Design of New Spatial Modulation Scheme Based on Quaternary Quasi-Orthogonal Sequences

    Hojun KIM  Yulong SHANG  Taejin JUNG  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2017/06/02
      Vol:
    E100-B No:12
      Page(s):
    2129-2138

    In this paper, we propose a new spatial modulation (SM) scheme based on quaternary quasi-orthogonal sequences (Q-QOSs), referred to as Q-QOS-SM. First, the conventional SM and generalized-SM (GSM) schemes are reinterpreted as a new transmission scheme based on a spatial modulation matrix (SMM), whose column indices are used for the mapping of spatial-information bits unlike the conventional ones. Next, by adopting the SMM comprising the Q-QOSs, the proposed Q-QOS-SM that guarantees twice the number of spatial bits at a transmitter compared with the SM with a constraint of transmit antennas, is designed. From the computer-simulation results, the Q-QOS-SM is shown to achieve a greatly improved throughput compared with the conventional SM and GSM schemes, especially, for a large number of the receive antennas. This finding is mainly because the new scheme offers a much higher minimum Euclidean distance than the other schemes.

  • Distributed Pareto Local Search for Multi-Objective DCOPs

    Maxime CLEMENT  Tenda OKIMOTO  Katsumi INOUE  

     
    PAPER-Information Network

      Pubricized:
    2017/09/15
      Vol:
    E100-D No:12
      Page(s):
    2897-2905

    Many real world optimization problems involving sets of agents can be modeled as Distributed Constraint Optimization Problems (DCOPs). A DCOP is defined as a set of variables taking values from finite domains, and a set of constraints that yield costs based on the variables' values. Agents are in charge of the variables and must communicate to find a solution minimizing the sum of costs over all constraints. Many applications of DCOPs include multiple criteria. For example, mobile sensor networks must optimize the quality of the measurements and the quality of communication between the agents. This introduces trade-offs between solutions that are compared using the concept of Pareto dominance. Multi-Objective Distributed Constraint Optimization Problems (MO-DCOPs) are used to model such problems where the goal is to find the set of Pareto optimal solutions. This set being exponential in the number of variables, it is important to consider fast approximation algorithms for MO-DCOPs. The bounded multi-objective max-sum (B-MOMS) algorithm is the first and only existing approximation algorithm for MO-DCOPs and is suited for solving a less-constrained problem. In this paper, we propose a novel approximation MO-DCOP algorithm called Distributed Pareto Local Search (DPLS) that uses a local search approach to find an approximation of the set of Pareto optimal solutions. DPLS provides a distributed version of an existing centralized algorithm by complying with the communication limitations and the privacy concerns of multi-agent systems. Experiments on a multi-objective extension of the graph-coloring problem show that DPLS finds significantly better solutions than B-MOMS for problems with medium to high constraint density while requiring a similar runtime.

  • Gauss-Seidel HALS Algorithm for Nonnegative Matrix Factorization with Sparseness and Smoothness Constraints

    Takumi KIMURA  Norikazu TAKAHASHI  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:12
      Page(s):
    2925-2935

    Nonnegative Matrix Factorization (NMF) with sparseness and smoothness constraints has attracted increasing attention. When these properties are considered, NMF is usually formulated as an optimization problem in which a linear combination of an approximation error term and some regularization terms must be minimized under the constraint that the factor matrices are nonnegative. In this paper, we focus our attention on the error measure based on the Euclidean distance and propose a new iterative method for solving those optimization problems. The proposed method is based on the Hierarchical Alternating Least Squares (HALS) algorithm developed by Cichocki et al. We first present an example to show that the original HALS algorithm can increase the objective value. We then propose a new algorithm called the Gauss-Seidel HALS algorithm that decreases the objective value monotonically. We also prove that it has the global convergence property in the sense of Zangwill. We finally verify the effectiveness of the proposed algorithm through numerical experiments using synthetic and real data.

  • Single Image Dehazing Using Invariance Principle

    Mingye JU  Zhenfei GU  Dengyin ZHANG  Jian LIU  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2017/09/01
      Vol:
    E100-D No:12
      Page(s):
    3068-3072

    In this letter, we propose a novel technique to increase the visibility of the hazy image. Benefiting from the atmospheric scattering model and the invariance principle for scene structure, we formulate structure constraint equations that derive from two simulated inputs by performing gamma correction on the input image. Relying on the inherent boundary constraint of the scattering function, the expected scene albedo can be well restored via these constraint equations. Extensive experimental results verify the power of the proposed dehazing technique.

  • Resample-Based Hybrid Multi-Hypothesis Scheme for Distributed Compressive Video Sensing

    Can CHEN  Dengyin ZHANG  Jian LIU  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2017/09/08
      Vol:
    E100-D No:12
      Page(s):
    3073-3076

    Multi-hypothesis prediction technique, which exploits inter-frame correlation efficiently, is widely used in block-based distributed compressive video sensing. To solve the problem of inaccurate prediction in multi-hypothesis prediction technique at a low sampling rate and enhance the reconstruction quality of non-key frames, we present a resample-based hybrid multi-hypothesis scheme for block-based distributed compressive video sensing. The innovations in this paper include: (1) multi-hypothesis reconstruction based on measurements reorganization (MR-MH) which integrates side information into the original measurements; (2) hybrid multi-hypothesis (H-MH) reconstruction which mixes multiple multi-hypothesis reconstructions adaptively by resampling each reconstruction. Experimental results show that the proposed scheme outperforms the state-of-the-art technique at the same low sampling rate.

  • New Generalized Sidelobe Canceller with Denoising Auto-Encoder for Improved Speech Enhancement

    Minkyu SHIN  Seongkyu MUN  David K. HAN  Hanseok KO  

     
    LETTER-Speech and Hearing

      Vol:
    E100-A No:12
      Page(s):
    3038-3040

    In this paper, a multichannel speech enhancement system which adopts a denoising auto-encoder as part of the beamformer is proposed. The proposed structure of the generalized sidelobe canceller generates enhanced multi-channel signals, instead of merely one channel, to which the following denoising auto-encoder can be applied. Because the beamformer exploits spatial information and compensates for differences in the transfer functions of each channel, the proposed system is expected to resolve the difficulty of modelling relative transfer functions consisting of complex numbers which are hard to model with a denoising auto-encoder. As a result, the modelling capability of the denoising auto-encoder can concentrate on removing the artefacts caused by the beamformer. Unlike conventional beamformers, which combine these artefacts into one channel, they remain separated for each channel in the proposed method. As a result, the denoising auto-encoder can remove the artefacts by referring to other channels. Experimental results prove that the proposed structure is effective for the six-channel data in CHiME, as indicated by improvements in terms of speech enhancement and word error rate in automatic speech recognition.

  • A Study on the Error Performance of Soft-Decision Decodings for Binary Linear Codes on a 4-Level Quantization over an AWGN Channel

    Takuya KUSAKA  

     
    PAPER-Coding Theory

      Vol:
    E100-A No:12
      Page(s):
    3016-3022

    In this paper, a study on the design and implementation of uniform 4-level quantizers for soft-decision decodings for binary linear codes is shown. Simulation results on quantized Viterbi decoding with a 4-level quantizer for the (64,42,8) Reed-Muller code show that the optimum stepsize, which is derived from the cutoff rate, gives an almost optimum error performance. In addition, the simulation results show that the case where the number of optimum codewords is larger than the one for a received sequence causes non-negligible degradation on error performance at high SN ratios of Eb/N0.

  • New Constructions of Multiple Binary ZCZ Sequence Sets with Inter-Set Zero Cross-Correlation Zone

    Tao LIU  Chengqian XU  Yubo LI  Xiaoyu CHEN  

     
    PAPER-Information Theory

      Vol:
    E100-A No:12
      Page(s):
    3007-3015

    In this correspondence, two types of multiple binary zero correlation zone (ZCZ) sequence sets with inter-set zero cross-correlation zone (ZCCZ) are constructed. Based on orthogonal matrices with order N×N, multiple binary ZCZ sequence sets with inter-set even and odd ZCCZ lengthes are constructed, each set is an optimal ZCZ sequence set with parameters (2N2, N, N+1)-ZCZ, among these ZCZ sequence sets, sequences possess ideal cross-correlation property within a zone of length 2Z or 2Z+1. These resultant multiple ZCZ sequence sets can be used in quasi-synchronous CDMA systems to remove the inter-cell interference (ICI).

  • Power-Effective File Layout Based on Large Scale Data-Intensive Application in Virtualized Environment

    Shunsuke YAGAI  Masato OGUCHI  Miyuki NAKANO  Saneyasu YAMAGUCHI  

     
    PAPER-Database system

      Pubricized:
    2017/07/14
      Vol:
    E100-D No:12
      Page(s):
    2761-2770

    In data centers, large numbers of computers are run simultaneously. These computers consume an enormous amount of energy. Several challenges related to this issue have been published. An energy-efficient storage management method that cooperates with applications was one effective approach. In this method, data and storage devices are managed using application support and the power consumption of storage devices is significantly decreased. However, existing studies do not take the virtualized environment into account. Recently, many data-intensive applications have been run in a virtualized environment, such as the cloud computing environment. In this paper, we focus on a virtualized environment wherein multiple virtual machines run on a physical computer and a data intensive application runs on each virtual machine. We discuss a method for reducing storage device power consumption using application support. First, we propose two storage management methods using application information. One method optimizes the inter-HDD file layout. This method removes frequently-accessed files from a certain HDD and switches the HDD to power-off mode. To balance loads and reduce seek distances, this method separates a heavily accessed file and consolidates files in a virtual machine with low access frequency. The other method optimizes the intra-HDD file layout, in addition to performing inter-HDD optimization. This method places frequently accessed files near each other. Second, we present our experimental results and demonstrate that the proposed methods can create sufficiently long HDD access intervals that power-off mode can be used, and thereby, reduce the power consumption of storage devices.

  • A New Approach of Matrix Factorization on Complex Domain for Data Representation

    Viet-Hang DUONG  Manh-Quan BUI  Jian-Jiun DING  Yuan-Shan LEE  Bach-Tung PHAM  Pham The BAO  Jia-Ching WANG  

     
    LETTER-Pattern Recognition

      Pubricized:
    2017/09/15
      Vol:
    E100-D No:12
      Page(s):
    3059-3063

    This work presents a new approach which derives a learned data representation method through matrix factorization on the complex domain. In particular, we introduce an encoding matrix-a new representation of data-that satisfies the simplicial constraint of the projective basis matrix on the field of complex numbers. A complex optimization framework is provided. It employs the gradient descent method and computes the derivative of the cost function based on Wirtinger's calculus.

  • Towards 5G Network Slicing over Multiple-Domains Open Access

    Ibrahim AFOLABI  Adlen KSENTINI  Miloud BAGAA  Tarik TALEB  Marius CORICI  Akihiro NAKAO  

     
    INVITED PAPER

      Pubricized:
    2017/05/16
      Vol:
    E100-B No:11
      Page(s):
    1992-2006

    One of the key objectives of 5G is to evolve the current mobile network architecture from “one-fit-all” design model to a more customized and dynamically scaling one that enables the deployment of parallel systems, tailored to the service requirements on top of a shared infrastructure. Indeed, the envisioned 5G services may require different needs in terms of capacity, latency, bandwidth, reliability and security, which cannot be efficiently sustained by the same network infrastructure. Coming to address these customization challenges, network softwarization expressed through Software Defined Networking (SDN) programmable network infrastructures, Network Function Virtualization (NFV) running network functions as software and cloud computing flexibility paradigms, is seen as a possible panacea to addressing the variations in the network requirements posed by the 5G use cases. This will enable network flexibility and programmability, allow the creation and lifecycle management of virtual network slices tailored to the needs of 5G verticals expressed in the form of Mobile Virtual Network Operators (MVNOs) for automotive, eHealth, massive IoT, massive multimedia broadband. In this vein, this paper introduces a potential 5G architecture that enables the orchestration, instantiation and management of end-to-end network slices over multiple administrative and technological domains. The architecture is described from both the management and the service perspective, underlining the common functionality as well as how the response to the diversified service requirements can be achieved through proper software network components development.

  • Network Function Virtualization: A Survey Open Access

    Malathi VEERARAGHAVAN  Takehiro SATO  Molly BUCHANAN  Reza RAHIMI  Satoru OKAMOTO  Naoaki YAMANAKA  

     
    INVITED PAPER

      Pubricized:
    2017/05/16
      Vol:
    E100-B No:11
      Page(s):
    1978-1991

    The objectives of this survey are to provide an in-depth coverage of a few selected research papers that have made significant contributions to the development of Network Function Virtualization (NFV), and to provide readers insights into the key advantages and disadvantages of NFV and Software Defined Networks (SDN) when compared to traditional networks. The research papers covered are classified into four categories: NFV Infrastructure (NFVI), Network Functions (NFs), Management And Network Orchestration (MANO), and service chaining. The NFVI papers describe “framework” software that implement common functions, such as dynamic scaling and load balancing, required by NF developers. Papers on NFs are classified as offering solutions for software switches or middleboxes. MANO papers covered in this survey are primarily on resource allocation (virtual network embedding), which is an orchestrator function. Finally, service chaining papers that offer examples and extensions are reviewed. Our conclusions are that with the current level of investment in NFV from cloud and Internet service providers, the promised cost savings are likely to be realized, though many challenges remain.

  • Energy-Efficient Resource Allocation Strategy for Low Probability of Intercept and Anti-Jamming Systems

    Yu Min HWANG  Jun Hee JUNG  Kwang Yul KIM  Yong Sin KIM  Jae Seang LEE  Yoan SHIN  Jin Young KIM  

     
    LETTER-Digital Signal Processing

      Vol:
    E100-A No:11
      Page(s):
    2498-2502

    The aim of this letter is to guarantee the ability of low probability of intercept (LPI) and anti-jamming (AJ) by maximizing the energy efficiency (EE) to improve wireless communication survivability and sustain wireless communication in jamming environments. We studied a scenario based on one transceiver pair with a partial-band noise jammer in a Rician fading channel and proposed an EE optimization algorithm to solve the optimization problem. With the proposed EE optimization algorithm, the LPI and AJ can be simultaneously guaranteed while satisfying the constraint of the maximum signal-to-jamming-and-noise ratio and combinatorial subchannel allocation condition, respectively. The results of the simulation indicate that the proposed algorithm is more energy-efficient than those of the baseline schemes and guarantees the LPI and AJ performance in a jamming environment.

  • Single Image Haze Removal Using Structure-Aware Atmospheric Veil

    Yun LIU  Rui CHEN  Jinxia SHANG  Minghui WANG  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2017/08/04
      Vol:
    E100-D No:11
      Page(s):
    2729-2733

    In this letter, we propose a novel and effective haze removal method by using the structure-aware atmospheric veil. More specifically, the initial atmospheric veil is first estimated based on dark channel prior and morphological operator. Furthermore, an energy optimization function considering the structure feature of the input image is constructed to refine the initial atmospheric veil. At last, the haze-free image can be restored by inverting the atmospheric scattering model. Additionally, brightness adjustment is also performed for preventing the dehazing result too dark. Experimental results on hazy images reveal that the proposed method can effectively remove the haze and yield dehazing results with vivid color and high scene visibility.

  • Precise Indoor Localization Method Using Dual-Facing Cameras on a Smart Device via Visible Light Communication

    Yohei NAKAZAWA  Hideo MAKINO  Kentaro NISHIMORI  Daisuke WAKATSUKI  Makoto KOBAYASHI  Hideki KOMAGATA  

     
    PAPER-Vision

      Vol:
    E100-A No:11
      Page(s):
    2295-2303

    In this paper, we propose a precise indoor localization method using visible light communication (VLC) with dual-facing cameras on a smart device (mobile phone, smartphone, or tablet device). This approach can assist the visually impaired with navigation, or provide mobile-robot control. The proposed method is different from conventional techniques in that dual-facing cameras are used to expand the localization area. The smart device is used as the receiver, and light-emitting diodes on the ceiling are used as localization landmarks. These are identified by VLC using a rolling shutter effect of complementary metal-oxide semiconductor image sensors. The front-facing camera captures the direct incident light of the landmarks, while the rear-facing camera captures mirror images of landmarks reflected from the floor face. We formulated the relationship between the poses (position and attitude) of the two cameras and the arrangement of landmarks using tilt detection by the smart device accelerometer. The equations can be analytically solved with a constant processing time, unlike conventional numerical methods, such as least-squares. We conducted a simulation and confirmed that the localization area was 75.6% using the dual-facing cameras, which was 3.8 times larger than that using only the front-facing camera. As a result of the experiment using two landmarks and a tablet device, the localization error in the horizontal direction was less than 98 mm at 90% of the measurement points. Moreover, the error estimation index can be used for appropriate route selection for pedestrians.

  • An Extreme Learning Machine Architecture Based on Volterra Filtering and PCA

    Li CHEN  Ling YANG  Juan DU  Chao SUN  Shenglei DU  Haipeng XI  

     
    PAPER-Information Network

      Pubricized:
    2017/08/02
      Vol:
    E100-D No:11
      Page(s):
    2690-2701

    Extreme learning machine (ELM) has recently attracted many researchers' interest due to its very fast learning speed, good generalization ability, and ease of implementation. However, it has a linear output layer which may limit the capability of exploring the available information, since higher-order statistics of the signals are not taken into account. To address this, we propose a novel ELM architecture in which the linear output layer is replaced by a Volterra filter structure. Additionally, the principal component analysis (PCA) technique is used to reduce the number of effective signals transmitted to the output layer. This idea not only improves the processing capability of the network, but also preserves the simplicity of the training process. Then we carry out performance evaluation and application analysis for the proposed architecture in the context of supervised classification and unsupervised equalization respectively, and the obtained results either on publicly available datasets or various channels, when compared to those produced by already proposed ELM versions and a state-of-the-art algorithm: support vector machine (SVM), highlight the adequacy and the advantages of the proposed architecture and characterize it as a promising tool to deal with signal processing tasks.

  • Weighted Voting of Discriminative Regions for Face Recognition

    Wenming YANG  Riqiang GAO  Qingmin LIAO  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2017/08/04
      Vol:
    E100-D No:11
      Page(s):
    2734-2737

    This paper presents a strategy, Weighted Voting of Discriminative Regions (WVDR), to improve the face recognition performance, especially in Small Sample Size (SSS) and occlusion situations. In WVDR, we extract the discriminative regions according to facial key points and abandon the rest parts. Considering different regions of face make different contributions to recognition, we assign weights to regions for weighted voting. We construct a decision dictionary according to the recognition results of selected regions in the training phase, and this dictionary is used in a self-defined loss function to obtain weights. The final identity of test sample is the weighted voting of selected regions. In this paper, we combine the WVDR strategy with CRC and SRC separately, and extensive experiments show that our method outperforms the baseline and some representative algorithms.

941-960hit(5900hit)