Jiali YOU Hanxing XUE Yu ZHUO Xin ZHANG Jinlin WANG
Predicting the service performance of Internet applications is important in service selection, especially for video services. In order to design a predictor for forecasting video service performance in third-party application, two famous service providers in China, Iqiyi and Letv, are monitored and analyzed. The study highlights that the measured performance in the observation period is time-series data, and it has strong autocorrelation, which means it is predictable. In order to combine the temporal information and map the measured data to a proper feature space, the authors propose a predictor based on a Conditional Restricted Boltzmann Machine (CRBM), which can capture the potential temporal relationship of the historical information. Meanwhile, the measured data of different sources are combined to enhance the training process, which can enlarge the training size and avoid the over-fit problem. Experiments show that combining the measured results from different resolutions for a video can raise prediction performance, and the CRBM algorithm shows better prediction ability and more stable performance than the baseline algorithms.
Chunyan HOU Chen CHEN Jinsong WANG
In the era of e-commerce, purchase behavior prediction is one of the most important issues to promote both online companies' sales and the consumers' experience. The previous researches usually use the feature engineering and ensemble machine learning algorithms for the prediction. The performance really depends on designed features and the scalability of algorithms because the large-scale data and a lot of categorical features lead to huge samples and the high-dimensional feature. In this study, we explore an alternative to use tree-based Feature Transformation (FT) and simple machine learning algorithms (e.g. Logistic Regression). Random Forest (RF) and Gradient Boosting decision tree (GB) are used for FT. Then, the simple algorithm, rather than ensemble algorithms, is used to predict purchase behavior based on transformed features. Tree-based FT regards the leaves of trees as transformed features, and can learn high-order interactions among original features. Compared with RF, if GB is used for FT, simple algorithms are enough to achieve better performance.
Koki ISHIDA Masamitsu TANAKA Takatsugu ONO Koji INOUE
CMOS microprocessors are limited in their capacity for clock speed improvement because of increasing computing power, i.e., they face a power-wall problem. Single-flux-quantum (SFQ) circuits offer a solution with their ultra-fast-speed and ultra-low-power natures. This paper introduces our contributions towards ultra-high-speed cryogenic SFQ computing. The first step is to design SFQ microprocessors. From qualitatively and quantitatively evaluating past-designed SFQ microprocessors, we have found that revisiting the architecture of SFQ microprocessors and on-chip caches is the first critical challenge. On the basis of cross-layer discussions and analysis, we came to the conclusion that a bit-parallel gate-level pipeline architecture is the best solution for SFQ designs. This paper summarizes our current research results targeting SFQ microprocessors and on-chip cache architectures.
Parfait I. TEBE Yujun KUANG Affum E. AMPOMA Kwasi A. OPARE
In this paper, we provide a novel solution to mitigate pilot contamination in massive MIMO technology. In the proposed approach, we consider seven copilot cells of the first layer of interfering cells of a cellular network. We derive and formulate the worst-case signal-to-interference power ratio (SIR) of a typical user in both downlink and uplink of a pilot contaminated cell. Based on the formulated SIR and other considerations of the system, the total pilot sequence length, the reliability of channel estimation within the cell, the spectral and energy efficiencies are derived and formulated in downlink. The user's transmit power and the achievable sum rate are also derived and formulated in uplink. Our results show that when the cell size is reduced the pilot contamination is significantly mitigated and hence the system performance is improved.
With the recent explosion of geographic data generated by smartphones, sensors, and satellites, a data storage that can handle the massive volume of data and support high-computational spatial queries is becoming essential. Although key-value stores efficiently handle large-scale data, they are not equipped with effective functions for supporting geographic data. To solve this problem, in this paper, we present G-HBase, a high-performance geographical database based on HBase, a standard key-value store. To index geographic data, we first use Geohash as the rowkey in HBase. Then, we present a novel partitioning method, namely binary Geohash rectangle partitioning, to support spatial queries. Our extensive experiments on real datasets have demonstrated an improved performance with k nearest neighbors and range query in G-HBase when compared with SpatialHadoop, a state-of-the-art framework with native support for spatial data. We also observed that performance of spatial join in G-HBase is on par with SpatialHadoop and outperforms SJMR algorithm in HBase.
Alimujiang YASEN Kazunori UEDA
We develop a technique for representing variable names and name binding which is a mechanism of associating a name with an entity in many formal systems including logic, programming languages and mathematics. The idea is to use a general form of graph links (or edges) called hyperlinks to represent variables, graph nodes as constructors of the formal systems, and a graph type called hlground to define substitutions. Our technique is based on simple notions of graph theory in which graph types ensure correct substitutions and keep bound variables distinct. We encode strong reduction of the untyped λ-calculus to introduce our technique. Then we encode a more complex formal system called System F<:, a polymorphic λ-calculus with subtyping that has been one of important theoretical foundations of functional programming languages. The advantage of our technique is that the representation of terms, definition of substitutions, and implementation of formal systems are all straightforward. We formalized the graph type hlground, proved that it ensures correct substitutions in the λ-calculus, and implemented hlground in HyperLMNtal, a modeling language based on hypergraph rewriting. Experiments were conducted to test this technique. By this technique, one can implement formal systems simply by following the steps of their definitions as described in papers.
Marut BURANARACH Chutiporn ANUTARIYA Nopachat KALAYANAPAN Taneth RUANGRAJITPAKORN Vilas WUWONGSE Thepchai SUPNITHI
Knowledge management is important for government agencies in improving service delivery to their customers and data inter-operation within and across organizations. Building organizational knowledge repository for government agency has unique challenges. In this paper, we propose that enterprise ontology can provide support for government agencies in capturing organizational taxonomy, best practices and global data schema. A case study of a large-scale adoption for the Thailand's Excise Department is elaborated. A modular design approach of the enterprise ontology for the excise tax domain is discussed. Two forms of organizational knowledge: global schema and standard practices were captured in form of ontology and rule-based knowledge. The organizational knowledge was deployed to support two KM systems: excise recommender service and linked open data. Finally, we discuss some lessons learned in adopting the framework in the government agency.
Hiroki CHIBA Yuki HYOGO Kazuo MISUE
Spatio-temporal dependent data, such as weather observation data, are data of which the attribute values depend on both time and space. Typical methods for the visualization of such data include plotting the attribute values at each point in time on a map and displaying series of the maps in chronological order with animation, or displaying them by juxtaposing horizontally or vertically. However, these methods are problematic in that they compel readers interested in grasping the spatial changes of the attribute values to memorize the representations on the maps. The problem is exacerbated by considering that the longer the time-period covered by the data, the higher the cognitive load. In order to solve these problems, the authors propose a visualization method capable of overlaying the representations of multiple instantaneous values on a single static map. This paper explains the design of the proposed method and reports two experiments conducted by the authors to investigate the usefulness of the method. The experimental results show that the proposed method is useful in terms of the speed and accuracy with which it reads the spatial changes and its ability to present data with long time series efficiently.
Heemang SONG Seunghoon CHO Kyung-Jin YOU Hyun-Chool SHIN
In this paper, we propose an automotive radar sensor compensation method improving direction of arrival (DOA) and preventing target split tracking. Amplitude and phase mismatching and mutual coupling between radar sensor arrays cause an inaccuracy problem in DOA estimation. By quantifying amplitude and phase distortion levels for each angle, we compensate the sensor distortion. Applying the proposed method to Bartlett, Capon and multiple signal classification (MUSIC) algorithms, we experimentally demonstrate the performance improvement using both experimental data from the chamber and real data obtained in actual road.
Yuan GAO Chengdong WU Xiaosheng YU Wei ZHOU Jiahui WU
Efficient optic disc (OD) segmentation plays a significant role in retinal image analysis and retinal disease screening. In this paper, we present a full-automatic segmentation approach called double boundary extraction for the OD segmentation. The proposed approach consists of the following two stages: first, we utilize an unsupervised learning technology and statistical method based on OD boundary information to obtain the initial contour adaptively. Second, the final optic disc boundary is extracted using the proposed LSO model. The performance of the proposed method is tested on the public DIARETDB1 database and the experimental results demonstrate the effectiveness and advantage of the proposed method.
Hao ZHENG Xingan XU Changwei LV Yuanfang SHANG Guodong WANG Chunlin JI
Concatenated zigzag (CZ) codes are classified as one kind of parallel-concatenated codes with powerful performance and low complexity. This kind of codes has flexible implementation methods and a good application prospect. We propose a modified turbo-type decoder and adaptive extrinsic information scaling method based on the Max-Log-APP (MLA) algorithm, which can provide a performance improvement also under the relatively low decoding complexity. Simulation results show that the proposed method can effectively help the sub-optimal MLA algorithm to approach the optimal performance. Some contrasts with low-density parity-check (LDPC) codes are also presented in this paper.
Mohd SHAHRIZAN OTHMAN Aleksandar SHURBEVSKI Hiroshi NAGAMOCHI
Given an edge-weighted bipartite digraph G=(A,B;E), the Bipartite Traveling Salesman Problem (BTSP) asks to find the minimum cost of a Hamiltonian cycle of G, or determine that none exists. When |A|=|B|=n, the BTSP can be solved using polynomial space in O*(42nnlog n) time by using the divide-and-conquer algorithm of Gurevich and Shelah (SIAM Journal of Computation, 16(3), pp.486-502, 1987). We adapt their algorithm for the bipartite case, and show an improved time bound of O*(42n), saving the nlog n factor.
Erasure codes have been considered as one of the most promising techniques for data reliability enhancement and storage efficiency in modern distributed storage systems. However, erasure codes often suffer from a time-consuming coding process which makes them nearly impractical. The opportunity to solve this problem probably rely on the parallelization of erasure-code-based application on the modern multi-/many-core processors to fully take advantage of the adequate hardware resources on those platforms. However, the complicated data allocation and limited I/O throughput pose a great challenge on the parallelization. To address this challenge, we propose a general multi-threaded parallel coding approach in this work. The approach consists of a general multi-threaded parallel coding model named as MTPerasure, and two detailed parallel coding algorithms, named as sdaParallel and ddaParallel, respectively, adapting to different I/O circumstances. MTPerasure is a general parallel coding model focusing on the high level data allocation, and it is applicable for all erasure codes and can be implemented without any modifications of the low level coding algorithms. The sdaParallel divides the data into several parts and the data parts are allocated to different threads statically in order to eliminate synchronization latency among multiple threads, which improves the parallel coding performance under the dummy I/O mode. The ddaParallel employs two threads to execute the I/O reading and writing on the basis of small pieces independently, which increases the I/O throughput. Furthermore, the data pieces are assigned to the coding thread dynamically. A special thread scheduling algorithm is also proposed to reduce thread migration latency. To evaluate our proposal, we parallelize the popular open source library jerasure based on our approach. And a detailed performance comparison with the original sequential coding program indicates that the proposed parallel approach outperforms the original sequential program by an extraordinary speedups from 1.4x up to 7x, and achieves better utilization of the computation and I/O resources.
Hongbin LIN Zheng WU Dong LEI Wei WANG Xiuping PENG
This letter presents a novel tensor voting mechanism — analytic tensor voting (ATV), to get rid of the difficulties in original tensor voting, especially the efficiency. One of the main advantages is its explicit voting formulations, which benefit the completion of tensor voting theory and computational efficiency. Firstly, new decaying function was designed following the basic spirit of decaying function in original tensor voting (OTV). Secondly, analytic stick tensor voting (ASTV) was formulated using the new decaying function. Thirdly, analytic plate and ball tensor voting (APTV, ABTV) were formulated through controllable stick tensor construction and tensorial integration. These make the each voting of tensor can be computed by several non-iterative matrix operations, improving the efficiency of tensor voting remarkably. Experimental results validate the effectiveness of proposed method.
Abu Hena Al MUKTADIR Ved P. KAFLE Pedro MARTINEZ-JULIA Hiroaki HARAI
Network virtualization and slicing technologies create opportunity for infrastructure-less virtual network operators (VNOs) to enter the market anytime and provide diverse services. Multiple VNOs compete to provide the same kinds of services to end users (EUs). VNOs lease virtual resources from the infrastructure provider (InP) and sell services to the EUs by using the leased resources. The difference between the selling and leasing is the gross profit for the VNOs. A VNO that leases resources without precise knowledge of future demand, may not consume all the leased resources through service offers to EUs. Consequently, the VNO experiences loss and resources remain unused. In order to improve resource utilization and ensure that new entrant VNOs do not face losses, proper estimation of initial resource demand is important. In this paper, we propose a Bayesian game with Cournot oligopoly model to properly estimate the optimal initial resource demands for multiple entrant competing VNOs (players) with the objective of maximizing the expected profit for each VNO. The VNOs offer the same kinds of services to EUs with different qualities (player's type), which are public information. The exact service quality with which a VNO competes in the market is private information. Therefore, a VNO assumes the type of its opponent VNOs with certain probability. We derive the Bayesian Nash equilibrium (BNE) of the presented game and evaluate numerically the effect of service qualities and prices on the expected profit and market share of the VNOs.
A convenient formula for the estimation of the clutter rank of the diving platform radar is derived. Brennan's rule provides a general formula to estimate the clutter rank for the side looking radar with a linear array, which is normally called one-dimensional (1D) estimation problem. With the help of the clutter wavenumber spectrum, the traditional estimation of the clutter rank is extended to the diving scenario and the estimation problem is two-dimensional (2D). The proposed rule is verified by the numerical simulations.
Toshihiro KATASHITA Masakazu HIOKI Yohei HORI Hanpei KOIKE
Field-programmable gate array (FPGA) devices are applied for accelerating specific calculations and reducing power consumption in a wide range of areas. One of the challenges associated with FPGAs is reducing static power for enforcing their power effectiveness. We propose a method involving fine-grained reconfiguration of body biases of logic and net resources to reduce the static power of FPGA devices. In addition, we develop an FPGA device called Flex Power FPGA with SOTB technology and demonstrate its power reduction function with a 32-bit counter circuit. In this paper, we describe the construction of an experimental platform to precisely evaluate power consumption and the maximum operating frequency of the device under various operating voltages and body biases with various practical circuits. Using the abovementioned platform, we evaluate the Flex Power FPGA chip at operating voltages of 0.5-1.0 V and at body biases of 0.0-0.5 V. In the evaluation, we use a 32-bit adder, 16-bit multiplier, and an SBOX circuit for AES cryptography. We operate the chip virtually with uniformed body bias voltage to drive all of the logic resources with the same threshold voltage. We demonstrate the advantage of the Flex Power FPGA by comparing its performance with non-reconfigurable biasing.
David FERNÁNDEZ HERMIDA Miguel RODELGO LACRUZ Cristina LÓPEZ BRAVO Francisco Javier GONZÁLEZ-CASTAO
The growth of Internet traffic and the variety of traffic classes make network performance extremely difficult to evaluate. Even though most current methods rely on complex or costly hardware, recent research on bandwidth sharing has suggested the possibility of defining evaluation methods that simply require basic statistics on aggregated link utilization, such as mean and variance. This would greatly simplify monitoring systems as these statistics are easily calculable from Simple Network Management Protocol (SNMP) calls. However, existing methods require knowledge of certain fixed information about the network being monitored (e.g. link capacities). This is usually unavailable when the operator's view is limited to its share of leased links or when shared links carry traffic with different priorities. In this paper, departing from the analysis of aggregated link utilization statistics obtainable from SNMP requests, we propose a method that detects traffic degradation based on link utilization samples. It does not require knowledge of the capacity of the aggregated link or any other network parameters, giving network operators the possibility to control network performance in a more reliable and cost-effective way.
This paper studies a simultaneous wireless information and power transfer (SWIPT) system in which the transmitter not only sends data and energy to many types of wireless users, such as multiple information decoding users, multiple hybrid power-splitting users (i.e., users with a power-splitting structure to receive both information and energy), and multiple energy harvesting users, but also prevents information from being intercepted by a passive eavesdropper. The transmitter is equipped with multiple antennas, whereas all users and the eavesdropper are assumed to be equipped with a single antenna. Since the transmitter does not have any channel state information (CSI) about the eavesdropper, artificial noise (AN) power is maximized to mask information as well as to interfere with the eavesdropper as much as possible. The non-convex optimization problem is formulated to minimize the transmit power satisfying all signal-to-interference-plus-noise (SINR) and harvested energy requirements for all users so that the remaining power for generating AN is maximized. With perfect CSI, a semidefinite relaxation (SDR) technique is applied, and the optimal solution is proven to be tight. With imperfect CSI, SDR and a Gaussian randomization algorithm are proposed to find the suboptimal solution. Finally, numerical performance with respect to the maximum SINR at the eavesdropper is determined by a Monte-Carlo simulation to compare the proposed AN scenario with a no-AN scenario, as well as to compare perfect CSI with imperfect CSI.
Limin CHEN Jing XU Peter Xiaoping LIU Hui YU
Compressive spectral imaging (CSI) systems capture the 3D spatiospectral data by measuring the 2D compressed focal plane array (FPA) coded projection with the help of reconstruction algorithms exploiting the sparsity of signals. However, the contradiction between the multi-dimension of the scenes and the limited dimension of the sensors has limited improvement of recovery performance. In order to solve the problem, a novel CSI system based on a coded aperture snapshot spectral imager, RGB-CASSI, is proposed, which has two branches, one for CASSI, another for RGB images. In addition, considering that conventional reconstruction algorithms lead to oversmoothing, a RGB-guided low-rank (RGBLR) method for compressive hyperspectral image reconstruction based on compressed sensing and coded aperture spectral imaging system is presented, in which the available additional RGB information is used to guide the reconstruction and a low-rank regularization for compressive sensing and a non-convex surrogate of the rank is also used instead of nuclear norm for seeking a preferable solution. Experiments show that the proposed algorithm performs better in both PSNR and subjective effects compared with other state-of-art methods.