Min ZHANG Bo XU Xiaoyun LI Dong FU Jian LIU Baojian WU Kun QIU
The capacity of optical transport networks has been increasing steadily and the networks are becoming more dynamic, complex, and transparent. Though it is common to use worst case assumptions for estimating the quality of transmission (QoT) in the physical layer, over provisioning results in high margin requirements. Accurate estimation on the QoT for to-be-established lightpaths is crucial for reducing provisioning margins. Machine learning (ML) is regarded as one of the most powerful methodological approaches to perform network data analysis and enable automated network self-configuration. In this paper, an artificial neural network (ANN) framework, a branch of ML, to estimate the optical signal-to-noise ratio (OSNR) of to-be-established lightpaths is proposed. It takes account of both nonlinear interference between spectrum neighboring channels and optical monitoring uncertainties. The link information vector of the lightpath is used as input and the OSNR of the lightpath is the target for output of the ANN. The nonlinear interference impact of the number of neighboring channels on the estimation accuracy is considered. Extensive simulation results show that the proposed OSNR estimation scheme can work with any RWA algorithm. High estimation accuracy of over 98% with estimation errors of less than 0.5dB can be achieved given enough training data. ANN model with R=4 neighboring channels should be used to achieve more accurate OSNR estimates. Based on the results, it is expected that the proposed ANN-based OSNR estimation for new lightpath provisioning can be a promising tool for margin reduction and low-cost operation of future optical transport networks.
Tomoya KAGEYAMA Osamu MUTA Haris GACANIN
In this paper, we propose an enhanced selected mapping (e-SLM) technique to improve the performance of OFDM-PLC systems under impulsive noise. At the transmitter, the best transmit sequence is selected from among possible candidates so as to minimize the weighted sum of transmit signal peak power and the estimated receive one, where the received signal peak power is estimated at the transmitter using channel state information (CSI). At the receiver, a nonlinear blanking is applied to hold the impulsive noise under a given threshold, where impulsive noise detection accuracy is improved by the proposed e-SLM. We evaluate the probability of false alarms raised by impulsive noise detection and bit error rate (BER) of OFDM-PLC system using the proposed e-SLM. The results show the effectiveness of the proposed method in OFDM-PLC system compared with the conventional blanking technique.
Zeyun ZHANG Xiaohuan WU Chunguo LI Wei-Ping ZHU
Direction of arrival (DOA) estimation as a fundamental issue in array signal processing has been extensively studied for many applications in military and civilian fields. Many DOA estimation algorithms have been developed for different application scenarios such as low signal-to-noise ratio (SNR), limited snapshots, etc. However, there are still some practical problems that make DOA estimation very difficult. One of them is the correlation between sources. In this paper, we develop a sparsity-based method to estimate the DOA of coherent signals with sparse linear array (SLA). We adopt the off-grid signal model and solve the DOA estimation problem in the sparse Bayesian learning (SBL) framework. By considering the SLA as a ‘missing sensor’ ULA, our proposed method treats the output of the SLA as a partial output of the corresponding virtual uniform linear array (ULA) to make full use of the expanded aperture character of the SLA. Then we employ the expectation-maximization (EM) method to update the hyper-parameters and the output of the virtual ULA in an iterative manner. Numerical results demonstrate that the proposed method has a better performance in correlated signal scenarios than the reference methods in comparison, confirming the advantage of exploiting the extended aperture feature of the SLA.
Toyotaro TOKIMOTO Shintaro TOKIMOTO Kengo FUJII Shogo MORITA Hirotsugu YAMAMOTO
We propose a method to realize a subjective super-resolution on a high-speed LED display, which dynamically shows a set of four neighboring pixels on every LED pixel. We have experimentally confirmed the subjective super-resolution effect. This paper proposes a subjective super-resolution hypothesis in human visual system and reports simulation results with pseudo fixation eye movements.
The estimation of the matrix rank of harmonic components of a music spectrogram provides some useful information, e.g., the determination of the number of basis vectors of the matrix-factorization-based algorithms, which is required for the automatic music transcription or in post-processing. In this work, we develop an algorithm based on Stein's unbiased risk estimator (SURE) algorithm with the matrix factorization model. The noise variance required for the SURE algorithm is estimated by suppressing the harmonic component via median filtering. An evaluation performed using the MIDI-aligned piano sounds (MAPS) database revealed an average estimation error of -0.26 (standard deviation: 4.4) for the proposed algorithm.
Phase-sensitive amplification (PSA) has unique properties, such as the quantum-limited noise figure of 0 dB and the phase clamping effect. This study investigates PSA characteristics when a chirped pulse is incident. The signal gain, the output waveform, and the noise figure for an optical pulse having been chirped through chromatic dispersion or self-phase modulation before amplification are analyzed. The results indicate that the amplification properties for a chirped pulse are different from those of a non-chirped pulse, such that the signal gain is small, the waveform is distorted, and the noise figure is degraded.
Zuopeng ZHAO Hongda ZHANG Yi LIU Nana ZHOU Han ZHENG Shanyi SUN Xiaoman LI Sili XIA
Although correlation filter-based trackers have demonstrated excellent performance for visual object tracking, there remain several challenges to be addressed. In this work, we propose a novel tracker based on the correlation filter framework. Traditional trackers face difficulty in accurately adapting to changes in the scale of the target when the target moves quickly. To address this, we suggest a scale adaptive scheme based on prediction scales. We also incorporate a speed-based adaptive model update method to further improve overall tracking performance. Experiments with samples from the OTB100 and KITTI datasets demonstrate that our method outperforms existing state-of-the-art tracking algorithms in fast motion scenes.
Since first introduced in 2008 with the 1.0 specification, OpenCL has steadily evolved over the decade to increase its support for heterogeneous parallel systems. In this paper, we accelerate stochastic simulation of biochemical reaction networks on modern GPUs (graphics processing units) by means of the OpenCL programming language. In implementing the OpenCL version of the stochastic simulation algorithm, we carefully apply its data-parallel execution model to optimize the performance provided by the underlying hardware parallelism of the modern GPUs. To evaluate our OpenCL implementation of the stochastic simulation algorithm, we perform a comparative analysis in terms of the performance using the CPU-based cluster implementation and the NVidia CUDA implementation. In addition to the initial report on the performance of OpenCL on GPUs, we also discuss applicability and programmability of OpenCL in the context of GPU-based scientific computing.
Krittin INTHARAWIJITR Katsuyoshi IIDA Hiroyuki KOGA Katsunori YAMAOKA
Most of latency-sensitive mobile applications depend on computational resources provided by a cloud computing service. The problem of relying on cloud computing is that, sometimes, the physical locations of cloud servers are distant from mobile users and the communication latency is long. As a result, the concept of distributed cloud service, called mobile edge computing (MEC), is being introduced in the 5G network. However, MEC can reduce only the communication latency. The computing latency in MEC must also be considered to satisfy the required total latency of services. In this research, we study the impact of both latencies in MEC architecture with regard to latency-sensitive services. We also consider a centralized model, in which we use a controller to manage flows between users and mobile edge resources to analyze MEC in a practical architecture. Simulations show that the interval and controller latency trigger some blocking and error in the system. However, the permissive system which relaxes latency constraints and chooses an edge server by the lowest total latency can improve the system performance impressively.
Yuki TAGUCHI Ryota KAWASHIMA Hiroki NAKAYAMA Tsunemasa HAYASHI Hiroshi MATSUO
Many studies have revealed that the performance of software-based Virtual Network Functions (VNFs) is insufficient for mission-critical networks. Scaling-out approaches, such as auto-scaling of VNFs, could handle a huge amount of traffic; however, the exponential traffic growth confronts us the limitations of both expandability of physical resources and complexity of their management. In this paper, we propose a fast datapath processing method called Packet Aggregation Flow (PA-Flow) that is based on hop-by-hop packet aggregation for more efficient Service Function Chaining (SFC). PA-Flow extends a notion of existing intra-node packet aggregation toward network-wide packet aggregation, and we introduce following three novel features. First, packet I/O overheads at intermediate network devices including NFV-nodes are mitigated by reduction of packet amount. Second, aggregated packets are further aggregated as going through the service chain in a hop-by-hop manner. Finally, next-hop aware packet aggregation is realized using OpenFlow-based flow tables. PA-Flow is designed to be available with various VNF forms (e.g. VM/container/baremetal-based) and virtual I/O technologies (e.g. vhost-user/SR-IOV), and its implementation does not bring noticeable delay for aggregation. We conducted two evaluations: (i) a baseline evaluation for understanding fundamental performance characteristics of PA-Flow (ii) a simulation-based SFC evaluation for proving PA-Flow's effect in a realistic environment. The results showed that throughput of short packet forwarding was improved by 4 times. Moreover, the total number of packets was reduced by 93% in a large-scale SFC.
Zheng-qiang WANG Kun-hao HUANG Xiao-yu WAN Zi-fu FAN
In this letter, we investigate the price-based power allocation for non-orthogonal multiple access (NOMA) networks, where the base station (BS) can admit the users to transmit by pricing their power. Stackelberg game is utilized to model the pricing and power purchasing strategies between the BS and the users. Based on backward induction, the pricing problem of the BS is recast into the non-convex power allocation problem, which is equivalent to the rate allocation problem by variable replacement. Based on the equivalence problem, an optimal price-based power allocation algorithm is proposed to maximize the revenue of the BS. Simulation results show that the proposed algorithm is superior to the existing pricing algorithm in items of the revenue of BS and the number of admitted users.
Richard Hsin-Hsyong YANG Chia-Kun LEE Shiunn-Jang CHERN
Continuous phase modulation (CPM) is a very attractive digital modulation scheme, with constant envelope feature and high efficiency in meeting the power and bandwidth requirements. CPM signals with pairs of input sequences that differ in an infinite number of positions and map into pairs of transmitted signals with finite Euclidean distance (ED) are called catastrophic. In the CPM scheme, data sequences that have the catastrophic property are called the catastrophic sequences; they are periodic difference data patterns. The catastrophic sequences are usually with shorter length of the merger. The corresponding minimum normalized squared ED (MNSED) is smaller and below the distance bound. Two important CPM schemes, viz., LREC and LRC schemes, are known to be catastrophic for most cases; they have poor overall power and bandwidth performance. In the literatures, it has been shown that the probability of generating such catastrophic sequences are negligible, therefore, the asymptotic error performance (AEP) of those well-known catastrophic CPM schemes evaluated with the corresponding MNSED, over AWGN channels, might be too negative or pessimistic. To deal with this problem in AWGN channel, this paper presents a new split-merged MNSED and provide criteria to explore which conventional catastrophic CPM scheme could increase the length of mergers with split-merged non-periodic events, effectively. For comparison, we investigate the exact power and bandwidth performance for LREC and LRC CPM for the same bandwidth occupancy. Computer simulation results verify that the AEP evaluating with the split-merged MNSED could achieve up to 3dB gain over the conventional approach.
Xiaomin LI Huali WANG Zhangkai LUO
Parameter estimation theorems for LFM signals have been developed due to the advantages of fractional Fourier transform (FrFT). The traditional estimation methods in the fractional Fourier domain (FrFD) are almost based on two-dimensional search which have the contradiction between estimation performance and complexity. In order to solve this problem, we introduce the orthogonal matching pursuit (OMP) into the FrFD, propose a modified optimization method to estimate initial frequency and final frequency of fractional bandlimited LFM signals. In this algorithm, the differentiation fractional spectrum which is used to form observation matrix in OMP is derived from the spectrum analytical formulations of the LFM signal, and then, based on that the LFM signal has approximate rectangular spectrum in the FrFD and the correlation between the LFM signal and observation matrix yields a maximal value at the edge of the spectrum (see Sect.3.3 for details), the edge spectrum information can be extracted by OMP. Finally, the estimations of initial frequency and final frequency are obtained through multiplying the edge information by the sampling frequency resolution. The proposed method avoids reconstruction and the traditional peak-searching procedure, and the iterations are needed only twice. Thus, the computational complexity is much lower than that of the existing methods. Meanwhile, Since the vectors at the initial frequency and final frequency points both have larger modulus, so that the estimations are closer to the actual values, better normalized root mean squared error (NRMSE) performance can be achieved. Both theoretical analysis and simulation results demonstrate that the proposed algorithm bears a relatively low complexity and its estimation precision is higher than search-based and reconstruction-based algorithms.
Mingu KIM Seungwoo HONG Il Hong SUH
Personalized trip planning is a challenging problem given that places of interest should be selected according to user preferences and sequentially arranged while satisfying various constraints. In this study, we aimed to model various uncertain aspects that should be considered during trip planning and efficiently generate personalized plans that maximize user satisfaction based on preferences and constraints. Specifically, we propose a probabilistic itinerary evaluation model based on a hybrid temporal Bayesian network that determines suitable itineraries considering preferences, constraints, and uncertain environmental variables. The model retrieves the sum of time-weighted user satisfaction, and ant colony optimization generates the trip plan that maximizes the objective function. First, the optimization algorithm generates candidate itineraries and evaluates them using the proposed model. Then, we improve candidate itineraries based on the evaluation results of previous itineraries. To validate the proposed trip planning approach, we conducted an extensive user study by asking participants to choose their preferred trip plans from options created by a human planner and our approach. The results show that our approach provides human-like trip plans, as participants selected our generated plans in 57% of the pairs. We also evaluated the efficiency of the employed ant colony optimization algorithm for trip planning by performance comparisons with other optimization methods.
Xuan WANG Bofeng ZHANG Mingqing HUANG Furong CHANG Zhuocheng ZHOU
When individuals make a purchase from online sources, they may lack first-hand knowledge of the product. In such cases, they will judge the quality of the item by the reviews other consumers have posted. Therefore, it is significant to determine whether comments about a product are credible. Most often, conventional research on comment credibility has employed supervised machine learning methods, which have the disadvantage of needing large quantities of training data. This paper proposes an unsupervised method for judging comment credibility based on the Biterm Sentiment Latent Dirichlet Allocation (BS-LDA) model. Using this approach, first we derived some distributions and calculated each comment's credibility score via them. A comment's credibility was judged based on whether it achieved a threshold score. Our experimental results using comments from Amazon.com demonstrated that the overall performance of our approach can play an important role in determining the credibility of comments in some situation.
Liquid crystal director distributions between strong and weak polar anchoring surfaces in hybrid aligned cells are numerically analyzed. When the anchoring is a critical one, homogeneously or homeotropicly liquid crystal alignment can be obtained. Such cells have no threshold voltage and a driving voltage can be reduced less than 0.5 volt.
Bandhit SUKSIRI Masahiro FUKUMOTO
This paper presents an efficient wideband two-dimensional direction-of-arrival (DOA) estimation for an L-shaped microphone array. We propose a way to construct a wideband sample cross-correlation matrix without any process of DOA preliminary estimation, such as beamforming technique, by exploiting sample cross-correlation matrices of two different frequencies for all frequency bins. Subsequently, wideband DOAs can be estimated by using this wideband matrix along with a scheme of estimating DOA in a narrowband subspace method. Therefore, a contribution of our study is providing an alternative framework for recent narrowband subspace methods to estimating the DOA of wideband sources directly. It means that this framework enables cutting-edge techniques in the existing narrowband subspace methods to implement the wideband direction estimation for reducing the computational complexity and facilitating the estimation algorithm. Theoretical analysis and effectiveness of the proposed method are substantiated through numerical simulations and experiments, which are performed in reverberating environments. The results show that performance of the proposed method performs better than others over a range of signal-to-noise ratio with just a few microphones. All these advantages make the proposed method a powerful tool for navigation systems based on acoustic signal processing.
Tianwen GUO Ping DENG Qiang YU Baoyun WANG
In this letter, we investigate a design of efficient antenna allocation at the full duplex receiver (FDR) in a multi-input multi-output multi-eavesdropper (MIMOME) wiretap channel for physical layer security improvement. Specifically, we propose the allocation which are feasible for the practical scenario with self-interference (SI) taken into account, because the jamming signals from FDR not only confuse the eavesdropper but also inevitably cause SI at the FDR. Due to the nolinear and coupling of the antenna allocation optimization problem, we transform the original problem into an integer programming problem. Then, we derive the optimal solution and the corresponding beamforming matrices in closed-form by means of combining spatial alignment and null-space projection method. Furthermore, we present the feasibility condition and full-protection condition, which offer insight into principles that enable more efficient and effective use of FDR in the wiretap channel for security improvement. From the simulation results, we validate the theoretical analysis and demonstrate the outstanding performance of the proposed antennas allocation at FDR.
Gou HOUBEN Shu FUJITA Keita TAKAHASHI Toshiaki FUJII
Depth (disparity) estimation from a light field (a set of dense multi-view images) is currently attracting much research interest. This paper focuses on how to handle a noisy light field for disparity estimation, because if left as it is, the noise deteriorates the accuracy of estimated disparity maps. Several researchers have worked on this problem, e.g., by introducing disparity cues that are robust to noise. However, it is not easy to break the trade-off between the accuracy and computational speed. To tackle this trade-off, we have integrated a fast denoising scheme in a fast disparity estimation framework that works in the epipolar plane image (EPI) domain. Specifically, we found that a simple 1-D slanted filter is very effective for reducing noise while preserving the underlying structure in an EPI. Moreover, this simple filtering does not require elaborate parameter configurations in accordance with the target noise level. Experimental results including real-world inputs show that our method can achieve good accuracy with much less computational time compared to some state-of-the-art methods.
Yuan ZHOU Yuichi GOTO Jingde CHENG
Many kinds of questionnaires, testing, and voting are performed in some completely electronic ways to do questions and answers on the Internet as Web applications, i.e. e-questionnaire systems, e-testing systems, and e-voting systems. Because there is no unified communication tool among the stakeholders of e-questionnaire, e-testing, and e-voting systems, until now, all the e-questionnaire, e-testing, and e-voting systems are designed, developed, used, and maintained in various ad hoc ways. As a result, the stakeholders are difficult to communicate to implement the systems, because there is neither an exhaustive requirement list to have a grasp of the overall e-questionnaire, e-testing, and e-voting systems nor a standardized terminology for these systems to avoid ambiguity. A general-purpose specification language to provide a unified description way for specifying various e-questionnaire, e-testing, and e-voting systems can solve the problems such that the stakeholders can refer to and use the complete requirements and standardized terminology for better communications, and can easily and unambiguously specify all the requirements of systems and services of e-questionnaire, e-testing, and e-voting, even can implement the systems. In this paper, we propose the first specification language, named “QSL,” with a standardized, consistent, and exhaustive list of requirements for specifying various e-questionnaire, e-testing, and e-voting systems such that the specifications can be used as the precondition of automatically generating e-questionnaire, e-testing, and e-voting systems. The paper presents our design addressing that QSL can specify all the requirements of various e-questionnaire, e-testing, and e-voting systems in a structured way, evaluates its effectiveness, performs real applications using QSL in case of e-questionnaire, e-testing, and e-voting systems, and shows various QSL applications for providing convenient QSL services to stakeholders.