Zi-fu FAN Chen-chen WEN Zheng-qiang WANG Xiao-yu WAN
In this letter, we investigate the price-based power allocation with rate proportional fairness constraint in downlink non-orthogonal multiple access (NOMA) systems. The Stackelberg game is utilized to model the interaction between the base station (BS) and users. The revenue maximization problem of the BS is first converted to rate allocation problem, then the optimal rate allocation for each user is obtained by variable substitution. Finally, a price-based power allocation with rate proportional fairness (PAPF) algorithm is proposed based on the relationship between rate and transmit power. Simulation results show that the proposed PAPF algorithm is superior to the previous price-based power allocation algorithm in terms of fairness index and minimum normalized user (MNU) rate.
Yuuka HIRAO Yoshinori TAKEUCHI Masaharu IMAI Jaehoon YU
Heart disease is one of the major causes of death in many advanced countries. For prevention or treatment of heart disease, getting an early diagnosis from a long time period of electrocardiogram (ECG) examination is necessary. However, it could be a large burden on medical experts to analyze this large amount of data. To reduce the burden and support the analysis, this paper proposes an arrhythmia detection method based on a deformable part model, which absorbs individual variation of ECG waveform and enables the detection of various arrhythmias. Moreover, to detect the arrhythmia in low processing delay, the proposed method only utilizes time domain features. In an experimental result, the proposed method achieved 0.91 F-measure for arrhythmia detection.
Chenxi LI Lei CAO Xiaoming LIU Xiliang CHEN Zhixiong XU Yongliang ZHANG
As an important method to solve sequential decision-making problems, reinforcement learning learns the policy of tasks through the interaction with environment. But it has difficulties scaling to large-scale problems. One of the reasons is the exploration and exploitation dilemma which may lead to inefficient learning. We present an approach that addresses this shortcoming by introducing qualitative knowledge into reinforcement learning using cloud control systems to represent ‘if-then’ rules. We use it as the heuristics exploration strategy to guide the action selection in deep reinforcement learning. Empirical evaluation results show that our approach can make significant improvement in the learning process.
Soft-thresholding is a sparse modeling method typically applied to wavelet denoising in statistical signal processing. It is also important in machine learning since it is an essential nature of the well-known LASSO (Least Absolute Shrinkage and Selection Operator). It is known that soft-thresholding, thus, LASSO suffers from a problem of dilemma between sparsity and generalization. This is caused by excessive shrinkage at a sparse representation. There are several methods for improving this problem in the field of signal processing and machine learning. In this paper, we considered to extend and analyze a method of scaling of soft-thresholding estimators. In a setting of non-parametric orthogonal regression problem including discrete wavelet transform, we introduced component-wise and data-dependent scaling that is indeed identical to non-negative garrote. We here considered a case where a parameter value of soft-thresholding is chosen from absolute values of the least squares estimates, by which the model selection problem reduces to the determination of the number of non-zero coefficient estimates. In this case, we firstly derived a risk and construct SURE (Stein's unbiased risk estimator) that can be used for determining the number of non-zero coefficient estimates. We also analyzed some properties of the risk curve and found that our scaling method with the derived SURE is possible to yield a model with low risk and high sparsity compared to a naive soft-thresholding method with SURE. This theoretical speculation was verified by a simple numerical experiment of wavelet denoising.
Xushan CHEN Jibin YANG Meng SUN Jianfeng LI
In order to significantly reduce the time and space needed, compressive sensing builds upon the fundamental assumption of sparsity under a suitable discrete dictionary. However, in many signal processing applications there exists mismatch between the assumed and the true sparsity bases, so that the actual representative coefficients do not lie on the finite grid discretized by the assumed dictionary. Unlike previous work this paper introduces the unified compressive measurement operator into atomic norm denoising and investigates the problems of recovering the frequency support of a combination of multiple sinusoids from sub-Nyquist samples. We provide some useful properties to ensure the optimality of the unified framework via semidefinite programming (SDP). We also provide a sufficient condition to guarantee the uniqueness of the optimizer with high probability. Theoretical results demonstrate the proposed method can locate the nonzero coefficients on an infinitely dense grid over a wide range of SNR case.
Image steganalysis can determine whether the image contains the secret messages. In practice, the number of the cover images is far greater than that of the secret images, so it is very important to solve the detection problem in imbalanced image sets. Currently, SMOTE, Borderline-SMOTE and ADASYN are three importantly synthesized algorithms used to solve the imbalanced problem. In these methods, the new sampling point is synthesized based on the minority class samples. But this research is seldom seen in image steganalysis. In this paper, we find that the features of the majority class sample are similar to those of the minority class sample based on the distribution of the image features in steganalysis. So the majority and minority class samples are both used to integrate the new sample points. In experiments, compared with SMOTE, Borderline-SMOTE and ADASYN, this approach improves detection accuracy using the FLD ensemble classifier.
Establishing local visual correspondences between images taken under different conditions is an important and challenging task in computer vision. A common solution for this task is detecting keypoints in images and then matching the keypoints with a feature descriptor. This paper proposes a robust and low-dimensional local feature descriptor named Adaptively Integrated Gradient and Intensity Feature (AIGIF). The proposed AIGIF descriptor partitions the support region surrounding each keypoint into sub-regions, and classifies the sub-regions into two categories: edge-dominated ones and smoothness-dominated ones. For edge-dominated sub-regions, gradient magnitude and orientation features are extracted; for smoothness-dominated sub-regions, intensity feature is extracted. The gradient and intensity features are integrated to generate the descriptor. Experiments on image matching were conducted to evaluate performances of the proposed AIGIF. Compared with SIFT, the proposed AIGIF achieves 75% reduction of feature dimension (from 128 bytes to 32 bytes); compared with SURF, the proposed AIGIF achieves 87.5% reduction of feature dimension (from 256 bytes to 32 bytes); compared with the state-of-the-art ORB descriptor which has the same feature dimension with AIGIF, AIGIF achieves higher accuracy and robustness. In summary, the AIGIF combines the advantages of gradient feature and intensity feature, and achieves relatively high accuracy and robustness with low feature dimension.
Zijie WANG Qin LIU Takeshi IKENAGA
High-dynamic-range imaging (HDRI) technologies aim to extend the dynamic range of luminance against the limitation of camera sensors. Irradiance information of a scene can be reconstructed by fusing multiple low-dynamic-range (LDR) images with different exposures. The key issue is removing ghost artifacts caused by motion of moving objects and handheld cameras. This paper proposes a robust ghost-free HDRI algorithm by visual salience based bilateral motion detection and stack extension based exposure fusion. For ghost areas detection, visual salience is introduced to measure the differences between multiple images; bilateral motion detection is employed to improve the accuracy of labeling motion areas. For exposure fusion, the proposed algorithm reduces the discontinuity of brightness by stack extension and rejects the information of ghost areas to avoid artifacts via fusion masks. Experiment results show that the proposed algorithm can remove ghost artifacts accurately for both static and handheld cameras, remain robust to scenes with complex motion and keep low complexity over recent advances including rank minimization based method and patch based method by 63.6% and 20.4% time savings averagely.
Xina CHENG Norikazu IKOMA Masaaki HONDA Takeshi IKENAGA
The ball state tracking and detection technology plays a significant role in volleyball game analysis, whose performance is limited due to the challenges include: 1) the inaccurate ball trajectory; 2) multiple numbers of the ball event category; 3) the large intra-class difference of one event. With the goal of broadcasting supporting for volleyball games which requires a real time system, this paper proposes a ball state based parallel ball tracking and event detection method based on a sequential estimation method such as particle filter. This method employs a parallel process of the 3D ball tracking and the event detection so that it is friendly for real time system implementation. The 3D ball tracking process uses the same models with the past work [8]. For event detection process, a ball event change estimation based multiple system model, a past trajectory referred hit point likelihood and a court-line distance feature based event type detection are proposed. First, the multiple system model transits the ball event state, which consists the event starting time and the event type, through three models dealing with different ball motion situations in the volleyball game, such as the motion keeping and changing. The mixture of these models is decided by estimation of the ball event change estimation. Secondly, the past trajectory referred hit point likelihood avoids the processing time delay between the ball tracking and the event detection process by evaluating the probability of the ball being hit at certain time without using future ball trajectories. Third, the feature of the distance between the ball and the specific court line are extracted to detect the ball event type. Experimental results based on multi-view HDTV video sequences (2014 Inter High School Men's Volleyball Games, Japan), which contains 606 events in total, show that the detection rate reaches 88.61% while the success rate of 3D ball tracking keeps more than 99%.
Koji TASHIRO Leonardo LANANTE Masayuki KUROSAKI Hiroshi OCHI
High-resolution image and video communication in home networks is highly expected to proliferate with the spread of Wi-Fi devices and the introduction of multiple-input multiple-output (MIMO) systems. This paper proposes a joint transmission and coding scheme for broadcasting high-resolution video streams over multiuser MIMO systems with an eigenbeam-space division multiplexing (E-SDM) technique. Scalable video coding makes it possible to produce the code stream comprised of multiple layers having unequal contribution to image quality. The proposed scheme jointly assigns the data of scalable code streams to subcarriers and spatial streams based on their signal-to-noise ratio (SNR) values in order to transmit visually important data with high reliability. Simulation results show that the proposed scheme surpasses the conventional unequal power allocation (UPA) approach in terms of both peak signal-to-noise ratio (PSNR) of received images and correct decoding probability. PSNR performance of the proposed scheme exceeds 35dB with the probability of over 95% when received SNR is higher than 6dB. The improvement in average PSNR by the proposed scheme compared to the conventional UPA comes up to approx. 20dB at received SNR of 6dB. Furthermore, correct decoding probability reaches 95% when received SNR is greater than 4dB.
Theerat SAKDEJAYONT Chun-Hao LIAO Makoto SUZUKI Hiroyuki MORIKAWA
Real-time and reliable radio communication is essential for wireless control systems (WCS). In WCS, preambles create significant overhead and affect the real-time capability since payloads are typically small. To shorten the preamble transmission time in OFDM systems, previous works have considered adopting either time-direction extrapolation (TDE) or frequency-direction interpolation (FDI) for channel estimation which however result in poor performance in fast fading channels and frequency-selective fading channels, respectively. In this work, we propose a subcarrier-selectable short preamble (SSSP) by introducing selectability to subcarrier sampling patterns of a preamble such that it can provide full sampling coverage of all subcarriers with several preamble transmissions. In addition, we introduce adaptability to a channel estimation algorithm for the SSSP so that it conforms to both fast and frequency-selective channels. Simulation results validate the feasibility of the proposed method in terms of the reliability and real-time capability. In particular, the SSSP scheme shows its advantage in flexibility as it can provide a low error rate and short communication time in various channel conditions.
Bo YIN Shotaro KAMIYA Koji YAMAMOTO Takayuki NISHIO Masahiro MORIKURA Hirantha ABEYSEKERA
Distributed channel selection schemes are proposed in this paper to mitigate the flow-in-the-middle (FIM) starvation in dense wireless local area networks (WLANs). The FIM starvation occurs when the middle transmitter is within the carrier sense range of two exterior transmitters, while the two exterior transmitters are not within the carrier sense range of each other. Since an exterior transmitter sends a frame regardless of the other, the middle transmitter has a high probability of detecting the channel being occupied. Under heavy traffic conditions, the middle transmitter suffers from extremely low transmission opportunities, i.e., throughput starvation. The basic idea of the proposed schemes is to let each access point (AP) select the channel which has less three-node-chain topologies within its two-hop neighborhood. The proposed schemes are formulated in strategic form games. Payoff functions are designed so that they are proved to be potential games. Therefore, the convergence is guaranteed when the proposed schemes are conducted in a distributed manner by using unilateral improvement dynamics. Moreover, we conduct evaluations through graph-based simulations and the ns-3 simulator. Simulations confirm that the FIM starvation has been mitigated since the number of three-node-chain topologies has been significantly reduced. The 5th percentile throughput has been improved.
Jae-Young YANG Ledan WU Yafeng ZHOU Joonho KWON Han-You JEONG
In this paper, we study Wi-Fi mesh networks (WMNs) as a promising candidate for wireless networking infrastructure that interconnects a variety of access networks. The main performance bottleneck of a WMN is their limited capacity due to the packet collision from the contention-based IEEE 802.11s MAC. To mitigate this problem, we present the distributed link-activation (DLA) protocol which activates a set of collision-free links for a fixed amount of time by exchanging a few control packets between neighboring MRs. Through the rigorous proof, it is shown that the upper bound of the DLA rounds is O(Smax), where Smax is the maximum number of (simultaneous) interference-free links in a WMN topology. Based on the DLA, we also design the distributed throughput-maximal scheduling (D-TMS) scheme which overlays the DLA protocol on a new frame architecture based on the IEEE 802.11 power saving mode. To mitigate its high latency, we propose the D-TMS adaptive data-period control (D-TMS-ADPC) that adjusts the data period depending on the traffic load of a WMN. Numerical results show that the D-TMS-ADPC scheme achieves much higher throughput performance than the IEEE 802.11s MAC.
Liangrui TANG Shiyu JI Shimo DU Yun REN Runze WU Xin WU
Network traffic forecasts, as it is well known, can be useful for network resource optimization. In order to minimize the forecast error by maximizing information utilization with low complexity, this paper concerns the difference of traffic trends at large time scales and fits a dual-related model to predict it. First, by analyzing traffic trends based on user behavior, we find both hour-to-hour and day-to-day patterns, which means that models based on either of the single trends are unable to offer precise predictions. Then, a prediction method with the consideration of both daily and hourly traffic patterns, called the dual-related forecasting method, is proposed. Finally, the correlation for traffic data is analyzed based on model parameters. Simulation results demonstrate the proposed model is more effective in reducing forecasting error than other models.
Ju Hong YOON Jungho KIM Youngbae HWANG
In this letter, we propose a robust and fast tracking framework by combining local and global appearance models to cope with partial occlusion and pose variations. The global appearance model is represented by a correlation filter to efficiently estimate the movement of the target and the local appearance model is represented by local feature points to handle partial occlusion and scale variations. Then global and local appearance models are unified via the Bayesian inference in our tracking framework. We experimentally demonstrate the effectiveness of the proposed method in both terms of accuracy and time complexity, which takes 12ms per frame on average for benchmark datasets.
Toru SUMI Yuta INAMURA Yusuke KAMEDA Tomokazu ISHIKAWA Ichiro MATSUDA Susumu ITOH
We previously proposed a lossless image coding scheme using example-based probability modeling, wherein the probability density function of image signals was dynamically modeled pel-by-pel. To appropriately estimate the peak positions of the probability model, several examples, i.e., sets of pels whose neighborhoods are similar to the local texture of the target pel to be encoded, were collected from the already encoded causal area via template matching. This scheme primarily makes use of non-local information in image signals. In this study, we introduce a prediction technique into the probability modeling to offer a better trade-off between the local and non-local information in the image signals.
Akitoshi ITAI Arao FUNASE Andrzej CICHOCKI Hiroshi YASUKAWA
This paper describes the background noise estimation technique of the tensor product expansion with absolute error (TPE-AE) to estimate multiple sources. The electroencephalogram (EEG) signal produced by the saccadic eye movement is adopted to analyze relationship between a brain function and a human activity. The electrooculogram (EOG) generated by eye movements yields significant problems for the EEG analysis. The denoising of EOG artifacts is important task to perform an accurate analysis. In this paper, the two types of TPE-AE are proposed to estimates EOG and other components in EEG during eye movement. One technique estimates two outer products using median filter based TPE-AE. The another technique uses a reference signal to separate the two sources. We show that the proposed method is effective to estimate and separate two sources in EEG.
Wanming HAO Osamu MUTA Haris GACANIN Hiroshi FURUKAWA
Massive MIMO (mMIMO) is a promising technology for smart multimedia and wireless communication fields. In this paper, we investigate pilot allocation problem in two-tier time division duplex (TDD) heterogeneous network (HetNet) with mMIMO. First, we propose a new pilot allocation scheme for maximizing ergodic downlink sum rate of macro users (MUs) and small cell users (SUs), where the uplink pilot overhead and cross-tier interference are jointly considered. Then, we theoretically analyze the formulated problem and propose a low complexity one-dimensional search algorithm to obtain the optimum pilot allocation. In addition, we propose two suboptimal pilot allocation algorithms to simplify the computational process and improve SUs' fairness, respectively. Finally, simulation results show that the performance of the proposed scheme outperforms that of the traditional schemes.
Mitsuji MUNEYASU Nayuta JINDA Yuuya MORITANI Soh YOSHIDA
In this paper, we propose a method of embedding and detecting data in printed images with several formats, such as different resolutions and numbers of blocks, using the camera of a tablet device. To specify the resolution of an image and the number of blocks, invisible markers that are embedded in the amplitude domain of the discrete Fourier transform of the target image are used. The proposed method can increase the variety of images suitable for data embedding.
Xu WANG Julan XIE Zishu HE Qi ZHANG
In the scenario of finite sample size, the performance of the generalized sidelobe canceller (GSC) is still affected by the desired signal even if all signal sources are independent with each other. Firstly, the novel expression of weight vector of the auxiliary array is derived under the circumstances of finite sample size. Utilizing this new weight vector and considering the correlative interferences, the general expression for the interference cancellation ratio (CR) is developed. Then, the impacts of the CR performance are further analyzed for the parameters including the input signal-to-noise ratio (SNR), the auxiliary array size, the correlation coefficient between the desired signal and interference as well as the snapshots of the sample data, respectively. Some guidelines can thus be given for the practical application. Numerical simulations demonstrate the agreement between the simulation results and the analytical results.