The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] form(3161hit)

441-460hit(3161hit)

  • Forecasting Service Performance on the Basis of Temporal Information by the Conditional Restricted Boltzmann Machine

    Jiali YOU  Hanxing XUE  Yu ZHUO  Xin ZHANG  Jinlin WANG  

     
    PAPER-Network

      Pubricized:
    2017/11/10
      Vol:
    E101-B No:5
      Page(s):
    1210-1221

    Predicting the service performance of Internet applications is important in service selection, especially for video services. In order to design a predictor for forecasting video service performance in third-party application, two famous service providers in China, Iqiyi and Letv, are monitored and analyzed. The study highlights that the measured performance in the observation period is time-series data, and it has strong autocorrelation, which means it is predictable. In order to combine the temporal information and map the measured data to a proper feature space, the authors propose a predictor based on a Conditional Restricted Boltzmann Machine (CRBM), which can capture the potential temporal relationship of the historical information. Meanwhile, the measured data of different sources are combined to enhance the training process, which can enlarge the training size and avoid the over-fit problem. Experiments show that combining the measured results from different resolutions for a video can raise prediction performance, and the CRBM algorithm shows better prediction ability and more stable performance than the baseline algorithms.

  • Tree-Based Feature Transformation for Purchase Behavior Prediction

    Chunyan HOU  Chen CHEN  Jinsong WANG  

     
    LETTER-Artificial Intelligence, Data Mining

      Pubricized:
    2018/02/02
      Vol:
    E101-D No:5
      Page(s):
    1441-1444

    In the era of e-commerce, purchase behavior prediction is one of the most important issues to promote both online companies' sales and the consumers' experience. The previous researches usually use the feature engineering and ensemble machine learning algorithms for the prediction. The performance really depends on designed features and the scalability of algorithms because the large-scale data and a lot of categorical features lead to huge samples and the high-dimensional feature. In this study, we explore an alternative to use tree-based Feature Transformation (FT) and simple machine learning algorithms (e.g. Logistic Regression). Random Forest (RF) and Gradient Boosting decision tree (GB) are used for FT. Then, the simple algorithm, rather than ensemble algorithms, is used to predict purchase behavior based on transformed features. Tree-based FT regards the leaves of trees as transformed features, and can learn high-order interactions among original features. Compared with RF, if GB is used for FT, simple algorithms are enough to achieve better performance.

  • Towards Ultra-High-Speed Cryogenic Single-Flux-Quantum Computing Open Access

    Koki ISHIDA  Masamitsu TANAKA  Takatsugu ONO  Koji INOUE  

     
    INVITED PAPER

      Vol:
    E101-C No:5
      Page(s):
    359-369

    CMOS microprocessors are limited in their capacity for clock speed improvement because of increasing computing power, i.e., they face a power-wall problem. Single-flux-quantum (SFQ) circuits offer a solution with their ultra-fast-speed and ultra-low-power natures. This paper introduces our contributions towards ultra-high-speed cryogenic SFQ computing. The first step is to design SFQ microprocessors. From qualitatively and quantitatively evaluating past-designed SFQ microprocessors, we have found that revisiting the architecture of SFQ microprocessors and on-chip caches is the first critical challenge. On the basis of cross-layer discussions and analysis, we came to the conclusion that a bit-parallel gate-level pipeline architecture is the best solution for SFQ designs. This paper summarizes our current research results targeting SFQ microprocessors and on-chip cache architectures.

  • Mitigating Pilot Contamination in Massive MIMO Using Cell Size Reduction

    Parfait I. TEBE  Yujun KUANG  Affum E. AMPOMA  Kwasi A. OPARE  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2017/10/24
      Vol:
    E101-B No:5
      Page(s):
    1280-1290

    In this paper, we provide a novel solution to mitigate pilot contamination in massive MIMO technology. In the proposed approach, we consider seven copilot cells of the first layer of interfering cells of a cellular network. We derive and formulate the worst-case signal-to-interference power ratio (SIR) of a typical user in both downlink and uplink of a pilot contaminated cell. Based on the formulated SIR and other considerations of the system, the total pilot sequence length, the reliability of channel estimation within the cell, the spectral and energy efficiencies are derived and formulated in downlink. The user's transmit power and the achievable sum rate are also derived and formulated in uplink. Our results show that when the cell size is reduced the pilot contamination is significantly mitigated and hence the system performance is improved.

  • G-HBase: A High Performance Geographical Database Based on HBase

    Hong Van LE  Atsuhiro TAKASU  

     
    PAPER

      Pubricized:
    2018/01/18
      Vol:
    E101-D No:4
      Page(s):
    1053-1065

    With the recent explosion of geographic data generated by smartphones, sensors, and satellites, a data storage that can handle the massive volume of data and support high-computational spatial queries is becoming essential. Although key-value stores efficiently handle large-scale data, they are not equipped with effective functions for supporting geographic data. To solve this problem, in this paper, we present G-HBase, a high-performance geographical database based on HBase, a standard key-value store. To index geographic data, we first use Geohash as the rowkey in HBase. Then, we present a novel partitioning method, namely binary Geohash rectangle partitioning, to support spatial queries. Our extensive experiments on real datasets have demonstrated an improved performance with k nearest neighbors and range query in G-HBase when compared with SpatialHadoop, a state-of-the-art framework with native support for spatial data. We also observed that performance of spatial join in G-HBase is on par with SpatialHadoop and outperforms SJMR algorithm in HBase.

  • Name Binding is Easy with Hypergraphs

    Alimujiang YASEN  Kazunori UEDA  

     
    PAPER-Software System

      Pubricized:
    2018/01/12
      Vol:
    E101-D No:4
      Page(s):
    1126-1140

    We develop a technique for representing variable names and name binding which is a mechanism of associating a name with an entity in many formal systems including logic, programming languages and mathematics. The idea is to use a general form of graph links (or edges) called hyperlinks to represent variables, graph nodes as constructors of the formal systems, and a graph type called hlground to define substitutions. Our technique is based on simple notions of graph theory in which graph types ensure correct substitutions and keep bound variables distinct. We encode strong reduction of the untyped λ-calculus to introduce our technique. Then we encode a more complex formal system called System F<:, a polymorphic λ-calculus with subtyping that has been one of important theoretical foundations of functional programming languages. The advantage of our technique is that the representation of terms, definition of substitutions, and implementation of formal systems are all straightforward. We formalized the graph type hlground, proved that it ensures correct substitutions in the λ-calculus, and implemented hlground in HyperLMNtal, a modeling language based on hypergraph rewriting. Experiments were conducted to test this technique. By this technique, one can implement formal systems simply by following the steps of their definitions as described in papers.

  • An Ontology-Based Approach to Supporting Knowledge Management in Government Agencies: A Case Study of the Thai Excise Department

    Marut BURANARACH  Chutiporn ANUTARIYA  Nopachat KALAYANAPAN  Taneth RUANGRAJITPAKORN  Vilas WUWONGSE  Thepchai SUPNITHI  

     
    PAPER-Knowledge Representation

      Pubricized:
    2018/01/19
      Vol:
    E101-D No:4
      Page(s):
    884-891

    Knowledge management is important for government agencies in improving service delivery to their customers and data inter-operation within and across organizations. Building organizational knowledge repository for government agency has unique challenges. In this paper, we propose that enterprise ontology can provide support for government agencies in capturing organizational taxonomy, best practices and global data schema. A case study of a large-scale adoption for the Thailand's Excise Department is elaborated. A modular design approach of the enterprise ontology for the excise tax domain is discussed. Two forms of organizational knowledge: global schema and standard practices were captured in form of ontology and rule-based knowledge. The organizational knowledge was deployed to support two KM systems: excise recommender service and linked open data. Finally, we discuss some lessons learned in adopting the framework in the government agency.

  • Static Representation Exposing Spatial Changes in Spatio-Temporal Dependent Data

    Hiroki CHIBA  Yuki HYOGO  Kazuo MISUE  

     
    PAPER-Elemental Technologies for human behavior analysis

      Pubricized:
    2018/01/19
      Vol:
    E101-D No:4
      Page(s):
    933-943

    Spatio-temporal dependent data, such as weather observation data, are data of which the attribute values depend on both time and space. Typical methods for the visualization of such data include plotting the attribute values at each point in time on a map and displaying series of the maps in chronological order with animation, or displaying them by juxtaposing horizontally or vertically. However, these methods are problematic in that they compel readers interested in grasping the spatial changes of the attribute values to memorize the representations on the maps. The problem is exacerbated by considering that the longer the time-period covered by the data, the higher the cognitive load. In order to solve these problems, the authors propose a visualization method capable of overlaying the representations of multiple instantaneous values on a single static map. This paper explains the design of the proposed method and reports two experiments conducted by the authors to investigate the usefulness of the method. The experimental results show that the proposed method is useful in terms of the speed and accuracy with which it reads the spatial changes and its ability to present data with long time series efficiently.

  • Improving DOA Estimation and Preventing Target Split Using Automotive Radar Sensor Arrays

    Heemang SONG  Seunghoon CHO  Kyung-Jin YOU  Hyun-Chool SHIN  

     
    LETTER-Digital Signal Processing

      Vol:
    E101-A No:3
      Page(s):
    590-594

    In this paper, we propose an automotive radar sensor compensation method improving direction of arrival (DOA) and preventing target split tracking. Amplitude and phase mismatching and mutual coupling between radar sensor arrays cause an inaccuracy problem in DOA estimation. By quantifying amplitude and phase distortion levels for each angle, we compensate the sensor distortion. Applying the proposed method to Bartlett, Capon and multiple signal classification (MUSIC) algorithms, we experimentally demonstrate the performance improvement using both experimental data from the chamber and real data obtained in actual road.

  • Full-Automatic Optic Disc Boundary Extraction Based on Active Contour Model with Multiple Energies

    Yuan GAO  Chengdong WU  Xiaosheng YU  Wei ZHOU  Jiahui WU  

     
    LETTER-Vision

      Vol:
    E101-A No:3
      Page(s):
    658-661

    Efficient optic disc (OD) segmentation plays a significant role in retinal image analysis and retinal disease screening. In this paper, we present a full-automatic segmentation approach called double boundary extraction for the OD segmentation. The proposed approach consists of the following two stages: first, we utilize an unsupervised learning technology and statistical method based on OD boundary information to obtain the initial contour adaptively. Second, the final optic disc boundary is extracted using the proposed LSO model. The performance of the proposed method is tested on the public DIARETDB1 database and the experimental results demonstrate the effectiveness and advantage of the proposed method.

  • Adaptive Extrinsic Information Scaling for Concatenated Zigzag Codes Based on Max-Log-APP

    Hao ZHENG  Xingan XU  Changwei LV  Yuanfang SHANG  Guodong WANG  Chunlin JI  

     
    LETTER-Coding Theory

      Vol:
    E101-A No:3
      Page(s):
    627-631

    Concatenated zigzag (CZ) codes are classified as one kind of parallel-concatenated codes with powerful performance and low complexity. This kind of codes has flexible implementation methods and a good application prospect. We propose a modified turbo-type decoder and adaptive extrinsic information scaling method based on the Max-Log-APP (MLA) algorithm, which can provide a performance improvement also under the relatively low decoding complexity. Simulation results show that the proposed method can effectively help the sub-optimal MLA algorithm to approach the optimal performance. Some contrasts with low-density parity-check (LDPC) codes are also presented in this paper.

  • Polynomial-Space Exact Algorithms for the Bipartite Traveling Salesman Problem

    Mohd SHAHRIZAN OTHMAN  Aleksandar SHURBEVSKI  Hiroshi NAGAMOCHI  

     
    LETTER

      Pubricized:
    2017/12/19
      Vol:
    E101-D No:3
      Page(s):
    611-612

    Given an edge-weighted bipartite digraph G=(A,B;E), the Bipartite Traveling Salesman Problem (BTSP) asks to find the minimum cost of a Hamiltonian cycle of G, or determine that none exists. When |A|=|B|=n, the BTSP can be solved using polynomial space in O*(42nnlog n) time by using the divide-and-conquer algorithm of Gurevich and Shelah (SIAM Journal of Computation, 16(3), pp.486-502, 1987). We adapt their algorithm for the bipartite case, and show an improved time bound of O*(42n), saving the nlog n factor.

  • An Efficient Parallel Coding Scheme in Erasure-Coded Storage Systems

    Wenrui DONG  Guangming LIU  

     
    PAPER-Computer System

      Pubricized:
    2017/12/12
      Vol:
    E101-D No:3
      Page(s):
    627-643

    Erasure codes have been considered as one of the most promising techniques for data reliability enhancement and storage efficiency in modern distributed storage systems. However, erasure codes often suffer from a time-consuming coding process which makes them nearly impractical. The opportunity to solve this problem probably rely on the parallelization of erasure-code-based application on the modern multi-/many-core processors to fully take advantage of the adequate hardware resources on those platforms. However, the complicated data allocation and limited I/O throughput pose a great challenge on the parallelization. To address this challenge, we propose a general multi-threaded parallel coding approach in this work. The approach consists of a general multi-threaded parallel coding model named as MTPerasure, and two detailed parallel coding algorithms, named as sdaParallel and ddaParallel, respectively, adapting to different I/O circumstances. MTPerasure is a general parallel coding model focusing on the high level data allocation, and it is applicable for all erasure codes and can be implemented without any modifications of the low level coding algorithms. The sdaParallel divides the data into several parts and the data parts are allocated to different threads statically in order to eliminate synchronization latency among multiple threads, which improves the parallel coding performance under the dummy I/O mode. The ddaParallel employs two threads to execute the I/O reading and writing on the basis of small pieces independently, which increases the I/O throughput. Furthermore, the data pieces are assigned to the coding thread dynamically. A special thread scheduling algorithm is also proposed to reduce thread migration latency. To evaluate our proposal, we parallelize the popular open source library jerasure based on our approach. And a detailed performance comparison with the original sequential coding program indicates that the proposed parallel approach outperforms the original sequential program by an extraordinary speedups from 1.4x up to 7x, and achieves better utilization of the computation and I/O resources.

  • Research on Analytical Solution Tensor Voting

    Hongbin LIN  Zheng WU  Dong LEI  Wei WANG  Xiuping PENG  

     
    LETTER-Pattern Recognition

      Pubricized:
    2017/12/01
      Vol:
    E101-D No:3
      Page(s):
    817-820

    This letter presents a novel tensor voting mechanism — analytic tensor voting (ATV), to get rid of the difficulties in original tensor voting, especially the efficiency. One of the main advantages is its explicit voting formulations, which benefit the completion of tensor voting theory and computational efficiency. Firstly, new decaying function was designed following the basic spirit of decaying function in original tensor voting (OTV). Secondly, analytic stick tensor voting (ASTV) was formulated using the new decaying function. Thirdly, analytic plate and ball tensor voting (APTV, ABTV) were formulated through controllable stick tensor construction and tensorial integration. These make the each voting of tensor can be computed by several non-iterative matrix operations, improving the efficiency of tensor voting remarkably. Experimental results validate the effectiveness of proposed method.

  • A Bayesian Game to Estimate the Optimal Initial Resource Demand for Entrant Virtual Network Operators

    Abu Hena Al MUKTADIR  Ved P. KAFLE  Pedro MARTINEZ-JULIA  Hiroaki HARAI  

     
    PAPER

      Pubricized:
    2017/09/19
      Vol:
    E101-B No:3
      Page(s):
    667-678

    Network virtualization and slicing technologies create opportunity for infrastructure-less virtual network operators (VNOs) to enter the market anytime and provide diverse services. Multiple VNOs compete to provide the same kinds of services to end users (EUs). VNOs lease virtual resources from the infrastructure provider (InP) and sell services to the EUs by using the leased resources. The difference between the selling and leasing is the gross profit for the VNOs. A VNO that leases resources without precise knowledge of future demand, may not consume all the leased resources through service offers to EUs. Consequently, the VNO experiences loss and resources remain unused. In order to improve resource utilization and ensure that new entrant VNOs do not face losses, proper estimation of initial resource demand is important. In this paper, we propose a Bayesian game with Cournot oligopoly model to properly estimate the optimal initial resource demands for multiple entrant competing VNOs (players) with the objective of maximizing the expected profit for each VNO. The VNOs offer the same kinds of services to EUs with different qualities (player's type), which are public information. The exact service quality with which a VNO competes in the market is private information. Therefore, a VNO assumes the type of its opponent VNOs with certain probability. We derive the Bayesian Nash equilibrium (BNE) of the presented game and evaluate numerically the effect of service qualities and prices on the expected profit and market share of the VNOs.

  • Clutter Rank Estimation for Diving Platform Radar

    Fengde JIA  Zishu HE  

     
    LETTER-Analog Signal Processing

      Vol:
    E101-A No:3
      Page(s):
    600-603

    A convenient formula for the estimation of the clutter rank of the diving platform radar is derived. Brennan's rule provides a general formula to estimate the clutter rank for the side looking radar with a linear array, which is normally called one-dimensional (1D) estimation problem. With the help of the clutter wavenumber spectrum, the traditional estimation of the clutter rank is extended to the diving scenario and the estimation problem is two-dimensional (2D). The proposed rule is verified by the numerical simulations.

  • Development of an Evaluation Platform and Performance Experimentation of Flex Power FPGA Device

    Toshihiro KATASHITA  Masakazu HIOKI  Yohei HORI  Hanpei KOIKE  

     
    PAPER-Device and Architecture

      Pubricized:
    2017/11/17
      Vol:
    E101-D No:2
      Page(s):
    303-313

    Field-programmable gate array (FPGA) devices are applied for accelerating specific calculations and reducing power consumption in a wide range of areas. One of the challenges associated with FPGAs is reducing static power for enforcing their power effectiveness. We propose a method involving fine-grained reconfiguration of body biases of logic and net resources to reduce the static power of FPGA devices. In addition, we develop an FPGA device called Flex Power FPGA with SOTB technology and demonstrate its power reduction function with a 32-bit counter circuit. In this paper, we describe the construction of an experimental platform to precisely evaluate power consumption and the maximum operating frequency of the device under various operating voltages and body biases with various practical circuits. Using the abovementioned platform, we evaluate the Flex Power FPGA chip at operating voltages of 0.5-1.0 V and at body biases of 0.0-0.5 V. In the evaluation, we use a 32-bit adder, 16-bit multiplier, and an SBOX circuit for AES cryptography. We operate the chip virtually with uniformed body bias voltage to drive all of the logic resources with the same threshold voltage. We demonstrate the advantage of the Flex Power FPGA by comparing its performance with non-reconfigurable biasing.

  • Early Detection of Performance Degradation from Basic Aggregated Link Utilization Statistics

    David FERNÁNDEZ HERMIDA  Miguel RODELGO LACRUZ  Cristina LÓPEZ BRAVO  Francisco Javier GONZÁLEZ-CASTAO  

     
    PAPER-Network

      Pubricized:
    2017/07/26
      Vol:
    E101-B No:2
      Page(s):
    508-519

    The growth of Internet traffic and the variety of traffic classes make network performance extremely difficult to evaluate. Even though most current methods rely on complex or costly hardware, recent research on bandwidth sharing has suggested the possibility of defining evaluation methods that simply require basic statistics on aggregated link utilization, such as mean and variance. This would greatly simplify monitoring systems as these statistics are easily calculable from Simple Network Management Protocol (SNMP) calls. However, existing methods require knowledge of certain fixed information about the network being monitored (e.g. link capacities). This is usually unavailable when the operator's view is limited to its share of leased links or when shared links carry traffic with different priorities. In this paper, departing from the analysis of aggregated link utilization statistics obtainable from SNMP requests, we propose a method that detects traffic degradation based on link utilization samples. It does not require knowledge of the capacity of the aggregated link or any other network parameters, giving network operators the possibility to control network performance in a more reliable and cost-effective way.

  • Robust Secure Transmit Design for SWIPT System with Many Types of Wireless Users and Passive Eavesdropper

    Pham-Viet TUAN  Insoo KOO  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2017/08/22
      Vol:
    E101-B No:2
      Page(s):
    441-450

    This paper studies a simultaneous wireless information and power transfer (SWIPT) system in which the transmitter not only sends data and energy to many types of wireless users, such as multiple information decoding users, multiple hybrid power-splitting users (i.e., users with a power-splitting structure to receive both information and energy), and multiple energy harvesting users, but also prevents information from being intercepted by a passive eavesdropper. The transmitter is equipped with multiple antennas, whereas all users and the eavesdropper are assumed to be equipped with a single antenna. Since the transmitter does not have any channel state information (CSI) about the eavesdropper, artificial noise (AN) power is maximized to mask information as well as to interfere with the eavesdropper as much as possible. The non-convex optimization problem is formulated to minimize the transmit power satisfying all signal-to-interference-plus-noise (SINR) and harvested energy requirements for all users so that the remaining power for generating AN is maximized. With perfect CSI, a semidefinite relaxation (SDR) technique is applied, and the optimal solution is proven to be tight. With imperfect CSI, SDR and a Gaussian randomization algorithm are proposed to find the suboptimal solution. Finally, numerical performance with respect to the maximum SINR at the eavesdropper is determined by a Monte-Carlo simulation to compare the proposed AN scenario with a no-AN scenario, as well as to compare perfect CSI with imperfect CSI.

  • A RGB-Guided Low-Rank Method for Compressive Hyperspectral Image Reconstruction

    Limin CHEN  Jing XU  Peter Xiaoping LIU  Hui YU  

     
    PAPER-Image

      Vol:
    E101-A No:2
      Page(s):
    481-487

    Compressive spectral imaging (CSI) systems capture the 3D spatiospectral data by measuring the 2D compressed focal plane array (FPA) coded projection with the help of reconstruction algorithms exploiting the sparsity of signals. However, the contradiction between the multi-dimension of the scenes and the limited dimension of the sensors has limited improvement of recovery performance. In order to solve the problem, a novel CSI system based on a coded aperture snapshot spectral imager, RGB-CASSI, is proposed, which has two branches, one for CASSI, another for RGB images. In addition, considering that conventional reconstruction algorithms lead to oversmoothing, a RGB-guided low-rank (RGBLR) method for compressive hyperspectral image reconstruction based on compressed sensing and coded aperture spectral imaging system is presented, in which the available additional RGB information is used to guide the reconstruction and a low-rank regularization for compressive sensing and a non-convex surrogate of the rank is also used instead of nuclear norm for seeking a preferable solution. Experiments show that the proposed algorithm performs better in both PSNR and subjective effects compared with other state-of-art methods.

441-460hit(3161hit)