Qian LI Xiaojuan LI Bin WU Yunpeng XIAO
In social networks, predicting user behavior under social hotspots can aid in understanding the development trend of a topic. In this paper, we propose a retweeting prediction method for social hotspots based on tensor decomposition, using user information, relationship and behavioral data. The method can be used to predict the behavior of users and analyze the evolvement of topics. Firstly, we propose a tensor-based mechanism for mining user interaction, and then we propose that the tensor be used to solve the problem of inaccuracy that arises when interactively calculating intensity for sparse user interaction data. At the same time, we can analyze the influence of the following relationship on the interaction between users based on characteristics of the tensor in data space conversion and projection. Secondly, time decay function is introduced for the tensor to quantify further the evolution of user behavior in current social hotspots. That function can be fit to the behavior of a user dynamically, and can also solve the problem of interaction between users with time decay. Finally, we invoke time slices and discretization of the topic life cycle and construct a user retweeting prediction model based on logistic regression. In this way, we can both explore the temporal characteristics of user behavior in social hotspots and also solve the problem of uneven interaction behavior between users. Experiments show that the proposed method can improve the accuracy of user behavior prediction effectively and aid in understanding the development trend of a topic.
Lei CHEN Wei LU Ergude BAO Liqiang WANG Weiwei XING Yuanyuan CAI
MapReduce is an effective framework for processing large datasets in parallel over a cluster. Data locality and data skew on the reduce side are two essential issues in MapReduce. Improving data locality can decrease network traffic by moving reduce tasks to the nodes where the reducer input data is located. Data skew will lead to load imbalance among reducer nodes. Partitioning is an important feature of MapReduce because it determines the reducer nodes to which map output results will be sent. Therefore, an effective partitioner can improve MapReduce performance by increasing data locality and decreasing data skew on the reduce side. Previous studies considering both essential issues can be divided into two categories: those that preferentially improve data locality, such as LEEN, and those that preferentially improve load balance, such as CLP. However, all these studies ignore the fact that for different types of jobs, the priority of data locality and data skew on the reduce side may produce different effects on the execution time. In this paper, we propose a naive Bayes classifier based partitioner, namely, BAPM, which achieves better performance because it can automatically choose the proper algorithm (LEEN or CLP) by leveraging the naive Bayes classifier, i.e., considering job type and bandwidth as classification attributes. Our experiments are performed in a Hadoop cluster, and the results show that BAPM boosts the computing performance of MapReduce. The selection accuracy reaches 95.15%. Further, compared with other popular algorithms, under specific bandwidths, the improvement BAPM achieved is up to 31.31%.
Chunyan HOU Chen CHEN Jinsong WANG
In the era of e-commerce, purchase behavior prediction is one of the most important issues to promote both online companies' sales and the consumers' experience. The previous researches usually use the feature engineering and ensemble machine learning algorithms for the prediction. The performance really depends on designed features and the scalability of algorithms because the large-scale data and a lot of categorical features lead to huge samples and the high-dimensional feature. In this study, we explore an alternative to use tree-based Feature Transformation (FT) and simple machine learning algorithms (e.g. Logistic Regression). Random Forest (RF) and Gradient Boosting decision tree (GB) are used for FT. Then, the simple algorithm, rather than ensemble algorithms, is used to predict purchase behavior based on transformed features. Tree-based FT regards the leaves of trees as transformed features, and can learn high-order interactions among original features. Compared with RF, if GB is used for FT, simple algorithms are enough to achieve better performance.
This paper studies a wireless powered communication network (WPCN) with non-orthogonal multiple access (NOMA) under successive interference cancellation (SIC) constraints, where the users first harvest energy from the power station and then transmit data to the information receiver simultaneously. Under this setup, we investigate the system throughput maximization problem. We first formulate an optimization problem for a general case, which is non-convex. To derive the optimal solution, new variables are introduced to transform the initial problem into a convex optimization problem. For a special case, i.e., two-user case, the optimal solution is derived as a closed-form expression. Simulations on the effect of SIC constraints show the importance of the distinctness among users' channels for the proposed model.
Recently, the join processing of large-scale datasets in MapReduce environments has become an important issue. However, the existing MapReduce-based join algorithms suffer from too much overhead for constructing and updating the data index. Moreover, the similarity computation cost is high because the existing algorithms partition data without considering the data distribution. In this paper, we propose two grid-based join algorithms for MapReduce. First, we propose a similarity join algorithm that evenly distributes join candidates using a dynamic grid index, which partitions data considering data density and similarity threshold. We use a bottom-up approach by merging initial grid cells into partitions and assigning them to MapReduce jobs. Second, we propose a k-NN join query processing algorithm for MapReduce. To reduce the data transmission cost, we determine an optimal grid cell size by considering the data distribution of randomly selected samples. Then, we perform kNN join by assigning the only related join data to a reducer. From performance analysis, we show that our similarity join query processing algorithm and our k-NN join algorithm outperform existing algorithms by up to 10 times, in terms of query processing time.
KyungRak LEE SungRyung CHO JaeWon LEE Inwhee JOE
This paper proposes the mesh-topology based wireless-powered communication network (MT-WPCN), which consists of a hybrid-access point (H-AP) and nodes. The H-AP broadcasts energy to all nodes by wireless, and the nodes harvest the energy and then communicate with other nodes including the H-AP. For the communication in the MT-WPCN, we propose the harvest-then-transceive protocol to ensure that the nodes can harvest energy from the H-AP and transmit information selectively to the H-AP or other nodes, which is not supported in most protocols proposed for the conventional WPCN. In the proposed protocol, we consider that the energy harvesting can be interrupted at nodes, since the nodes cannot harvest energy during transmission or reception. We also consider that the harvested energy is consumed by the reception of information from other nodes. In addition, the energy reservation model is required to guarantee the QoS, which reserves the infimum energy to receive information reliably by the transmission power control. Under these considerations, first, we design the half harvest-then-transceive protocol, which indicates that a node transmits information only to other nodes which do not transmit information yet, for investing the effect of the energy harvesting interruption. Secondly, we also design the full harvest-then-transceive protocol for the information exchange among nodes and compatibility with the conventional star-topology based WPCN, which indicates that a node can transmit information to any network unit, i.e., the H-AP and all nodes. We study the sum-throughput maximization in the MT-WPCN based on the half and full harvest-then-transceive protocols, respectively. Furthermore, the amount of harvested energy is analytically compared according to the energy harvesting interruption in the protocols. Simulation results show that the proposed MT-WPCN outperforms the conventional star-topology based WPCN in terms of the sum-throughput maximization, when wireless information transmission among nodes occurs frequently.
Tao LIANG Flavia GRASSI Giordano SPADACINI Sergio Amedeo PIGNARI
This work presents a hybrid formulation of the stochastic reduced order model (SROM) algorithm, which makes use of Gauss quadrature, a key ingredient of the stochastic collocation method, to avoid the cumbersome optimization process required by SROM for optimal extraction of the sample set. With respect to classic SROM algorithms, the proposed formulation allows a significant reduction in computation time and burden as well as a remarkable improvement in the accuracy and convergence rate in the estimation of statistical moments. The method is here applied to a specific case study, that is the prediction of crosstalk in a two-conductor wiring structure with electrical and geometrical parameters not perfectly known. Both univariate and multivariate analyses are carried out, with the final objective being to compare the performance of the two SROM formulations with respected to Monte Carlo simulations.
Takashi MATSUBARA Ryo AKITA Kuniaki UEHARA
In this study, we propose a deep neural generative model for predicting daily stock price movements given news articles. Approaches involving conventional technical analysis have been investigated to identify certain patterns in past price movements, which in turn helps to predict future price movements. However, the financial market is highly sensitive to specific events, including corporate buyouts, product releases, and the like. Therefore, recent research has focused on modeling relationships between these events that appear in the news articles and future price movements; however, a very large number of news articles are published daily, each article containing rich information, which results in overfitting to past price movements used for parameter adjustment. Given the above, we propose a model based on a generative model of news articles that includes price movement as a condition, thereby avoiding excessive overfitting thanks to the nature of the generative model. We evaluate our proposed model using historical price movements of Nikkei 225 and Standard & Poor's 500 Stock Index, confirming that our model predicts future price movements better than such conventional classifiers as support vector machines and multilayer perceptrons. Further, our proposed model extracts significant words from news articles that are directly related to future stock price movements.
Jianbin ZHOU Dajiang ZHOU Takeshi YOSHIMURA Satoshi GOTO
Compressed Sensing based CMOS image sensor (CS-CIS) is a new generation of CMOS image sensor that significantly reduces the power consumption. For CS-CIS, the image quality and data volume of output are two important issues to concern. In this paper, we first proposed an algorithm to generate a series of deterministic and ternary matrices, which improves the image quality, reduces the data volume and are compatible with CS-CIS. Proposed matrices are derived from the approximate DCT and trimmed in 2D-zigzag order, thus preserving the energy compaction property as DCT does. Moreover, we proposed matrix row operations adaptive to the proposed matrix to further compress data (measurements) without any image quality loss. At last, a low-cost VLSI architecture of measurements compression with proposed matrix row operations is implemented. Experiment results show our proposed matrix significantly improve the coding efficiency by BD-PSNR increase of 4.2 dB, comparing with the random binary matrix used in the-state-of-art CS-CIS. The proposed matrix row operations for measurement compression further increases the coding efficiency by 0.24 dB BD-PSNR (4.8% BD-rate reduction). The VLSI architecture is only 4.3 K gates in area and 0.3 mW in power consumption.
This paper introduces a filter level pruning method based on similar feature extraction for compressing and accelerating the convolutional neural networks by k-means++ algorithm. In contrast to other pruning methods, the proposed method would analyze the similarities in recognizing features among filters rather than evaluate the importance of filters to prune the redundant ones. This strategy would be more reasonable and effective. Furthermore, our method does not result in unstructured network. As a result, it needs not extra sparse representation and could be efficiently supported by any off-the-shelf deep learning libraries. Experimental results show that our filter pruning method could reduce the number of parameters and the amount of computational costs in Lenet-5 by a factor of 17.9× with only 0.3% accuracy loss.
Yu YU Stepan KUCERA Yuto LIM Yasuo TAN
In mobile and wireless networks, controlling data delivery latency is one of open problems due to the stochastic nature of wireless channels, which are inherently unreliable. This paper explores how the current best-effort throughput-oriented wireless services might evolve into latency-sensitive enablers of new mobile applications such as remote three-dimensional (3D) graphical rendering for interactive virtual/augmented-reality overlay. Assuming that the signal propagation delay and achievable throughput meet the standard latency requirements of the user application, we examine the idea of trading excess/federated bandwidth for the elimination of non-negligible delay of data re-ordering, caused by temporal transmission failures and buffer overflows. The general system design is based on (i) spatially diverse data delivery over multiple paths with uncorrelated outage likelihoods; and (ii) forward packet-loss protection (FPP), creating encoding redundancy for proactive recovery of intolerably delayed data without end-to-end retransmissions. Analysis and evaluation are based on traces of real life traffic, which is measured in live carrier-grade long term evolution (LTE) networks and campus WiFi networks, due to no such system/environment yet to verify the importance of spatial diversity and encoding redundancy. Analysis and evaluation reveal the seriousness of the latency problem and that the proposed FPP with spatial diversity and encoding redundancy can minimize the delay of re-ordering. Moreover, a novel FPP effectiveness coefficient is proposed to explicitly represent the effectiveness of EPP implementation.
Haiyang LIU Yan LI Lianrong MA
The separating redundancy is an important concept in the analysis of the error-and-erasure decoding of a linear block code using a parity-check matrix of the code. In this letter, we derive new constructive upper bounds on the second separating redundancies of low-density parity-check (LDPC) codes constructed from projective and Euclidean planes over the field Fq with q even.
Miseon HAN Yeoul NA Dongha JUNG Hokyoon LEE Seon WOOK KIM Youngsun HAN
A memory controller refreshes DRAM rows periodically in order to prevent DRAM cells from losing data over time. Refreshes consume a large amount of energy, and the problem becomes worse with the future larger DRAM capacity. Previously proposed selective refreshing techniques are either conservative in exploiting the opportunity or expensive in terms of required implementation overhead. In this paper, we propose a novel DRAM selective refresh technique by using page residence in a memory hierarchy of hardware-managed TLB. Our technique maximizes the opportunity to optimize refreshing by activating/deactivating refreshes for DRAM pages when their PTEs are inserted to/evicted from TLB or data caches, while the implementation cost is minimized by slightly extending the existing infrastructure. Our experiment shows that the proposed technique can reduce DRAM refresh power 43.6% on average and EDP 3.5% with small amount of hardware overhead.
Shujiao LIAO Qingxin ZHU Rui LIANG
Rough set theory is an important branch of data mining and granular computing, among which neighborhood rough set is presented to deal with numerical data and hybrid data. In this paper, we propose a new concept called inconsistent neighborhood, which extracts inconsistent objects from a traditional neighborhood. Firstly, a series of interesting properties are obtained for inconsistent neighborhoods. Specially, some properties generate new solutions to compute the quantities in neighborhood rough set. Then, a fast forward attribute reduction algorithm is proposed by applying the obtained properties. Experiments undertaken on twelve UCI datasets show that the proposed algorithm can get the same attribute reduction results as the existing algorithms in neighborhood rough set domain, and it runs much faster than the existing ones. This validates that employing inconsistent neighborhoods is advantageous in the applications of neighborhood rough set. The study would provide a new insight into neighborhood rough set theory.
Kazuyuki ISHIKAWA Naoki HAYASHI Shigemasa TAKAI
This paper proposes a consensus-based distributed Particle Swarm Optimization (PSO) algorithm with event-triggered communications for a non-convex and non-differentiable optimization problem. We consider a multi-agent system whose local communications among agents are represented by a fixed and connected graph. Each agent has multiple particles as estimated solutions of global optima and updates positions of particles by an average consensus dynamics on an auxiliary variable that accumulates the past information of the own objective function. In contrast to the existing time-triggered approach, the local communications are carried out only when the difference between the current auxiliary variable and the variable at the last communication exceeds a threshold. We show that the global best can be estimated in a distributed way by the proposed event-triggered PSO algorithm under a diminishing condition of the threshold for the trigger condition.
Tracking-by-detection methods consider tracking task as a continuous detection problem applied over video frames. Modern tracking-by-detection trackers have online learning ability; the update stage is essential because it determines how to modify the classifier inherent in a tracker. However, most trackers search for the target within a fixed region centered at the previous object position; thus, they lack spatiotemporal consistency. This becomes a problem when the tracker detects an incorrect object during short-term occlusion. In addition, the scale of the bounding box that contains the target object is usually assumed not to change. This assumption is unrealistic for long-term tracking, where the scale of the target varies as the distance between the target and the camera changes. The accumulation of errors resulting from these shortcomings results in the drift problem, i.e. drifting away from the target object. To resolve this problem, we present a drift-free, online learning-based tracking-by-detection method using a single static camera. We improve the latent structured support vector machine (SVM) tracker by designing a more robust tracker update step by incorporating two Kalman filter modules: the first is used to predict an adaptive search region in consideration of the object motion; the second is used to adjust the scale of the bounding box by accounting for the background model. We propose a hierarchical search strategy that combines Bhattacharyya coefficient similarity analysis and Kalman predictors. This strategy facilitates overcoming occlusion and increases tracking efficiency. We evaluate this work using publicly available videos thoroughly. Experimental results show that the proposed method outperforms the state-of-the-art trackers.
Yasunori SUZUKI Junya OHKAWARA Shoichi NARAHASHI
This paper proposes a method for reducing the peak-to-average power ratio (PAPR) at the output signal of a digital predistortion linearizer (DPDL) that compensates for frequency dependent intermodulation distortion (IMD) components. The proposed method controls the amplitude and phase values of the frequency components corresponding to the transmission bandwidth of the output signal. A DPDL employing the proposed method simultaneously provides IMD component cancellation of out-of-band components and PAPR reduction at the output signal. This paper identifies the amplitude and phase conditions to minimize the PAPR. Experimental results based on a 2-GHz band 1-W class power amplifier show that the proposed method improves the drain efficiency of the power amplifier when degradation is allowed in the error vector magnitude. To the best knowledge of the authors, this is the first PAPR reduction method for DPDL that reduces the PAPR while simultaneously compensating for IMD components.
Shumpei YOSHIKAWA Koichi KOBAYASHI Yuh YAMASHITA
Event-triggered control is a method that the control input is updated only when a certain triggering condition is satisfied. In networked control systems, quantization errors via A/D conversion should be considered. In this paper, a new method for quantized event-triggered control with switching triggering conditions is proposed. For a discrete-time linear system, we consider the problem of finding a state-feedback controller such that the closed-loop system is uniformly ultimately bounded in a certain ellipsoid. This problem is reduced to an LMI (Linear Matrix Inequality) optimization problem. The volume of the ellipsoid may be adjusted. The effectiveness of the proposed method is presented by a numerical example.
Kenji HOSHINO Manabu MIKAMI Sourabh MAITI Hitoshi YOSHINO
Non-linear precoding (NLP) scheme for downlink multi-user multiple-input multiple-output (DL-MU-MIMO) transmission has received much attention as a promising technology to achieve high capacity within the limited bandwidths available to radio access systems. In order to minimize the required transmission power for DL-MU-MIMO and achieve high spectrum efficiency, Vector Perturbation (VP) was proposed as an optimal NLP scheme. Unfortunately, the original VP suffers from significant computation complexity in detecting the optimal perturbation vector from an infinite number of the candidates. To reduce the complexity with near transmission performance of VP, several recent studies investigated various efficient NLP schemes based on the concept of Tomlinson-Harashima precoding (THP) that applies successive pre-cancellation of inter-user interference (IUI) and offsets the transmission vector based on a modulo operation. In order to attain transmission performance improvement over the original THP, a previous work proposed Minimum Mean Square Error based THP (MMSE-THP) employing IUI successive pre-cancellation based on MMSE criteria. On the other hand, to improve the transmission performance of MMSE-THP, other previous works proposed Ordered MMSE-THP and Lattice-Reduction-Aided MMSE-THP (LRA MMSE-THP). This paper investigates the further transmission performance improvement of Ordered MMSE-THP and LRA MMSE-THP. This paper starts by proposing an extension of MMSE-THP employing a perturbation vector search (PVS), called PVS MMSE-THP as a novel NLP scheme, where the modulo operation is substituted by PVS and a subtraction operation from the transmit signal vector. Then, it introduces an efficient search algorithm of appropriate perturbation vector based on a depth-first branch-and-bound search for PVS MMSE-THP. Next, it also evaluates the transmission performance of PVS MMSE-THP with the appropriate perturbation vector detected by the efficient search algorithm. Computer simulations quantitatively clarify that PVS MMSE-THP achieves better transmission performance than the conventional NLP schemes. Moreover, it also clarifies that PVS MMSE-THP increases the effect of required transmission power reduction with the number of transmit antennas compared to the conventional NLP schemes.
Chow-Yen-Desmond SIM Chih-Chiang CHEN Che-Yu LI Sheng-Yang HUANG
A compact uniplanar antenna design for tablet/laptop applications is proposed. The main design principle of this antenna is the use of the coupling-feed mechanism. The proposed antenna is composed of an inverted L-shaped parasitic element, T-shaped feeding strip, parasitic shorted strip, and a step tuning stub. With its small size of 55mm × 15mm × 0.8mm, the proposed antenna is able to excite dual wideband transmission over the full LTE/WWAN operation ranges of 698-960MHz and 1710-2690MHz. Furthermore, the proposed antenna also exhibits reduced ground effects, such that reducing the ground size of the proposed antenna will not affect its performance.