Bandhit SUKSIRI Masahiro FUKUMOTO
This paper presents an efficient wideband two-dimensional direction-of-arrival (DOA) estimation for an L-shaped microphone array. We propose a way to construct a wideband sample cross-correlation matrix without any process of DOA preliminary estimation, such as beamforming technique, by exploiting sample cross-correlation matrices of two different frequencies for all frequency bins. Subsequently, wideband DOAs can be estimated by using this wideband matrix along with a scheme of estimating DOA in a narrowband subspace method. Therefore, a contribution of our study is providing an alternative framework for recent narrowband subspace methods to estimating the DOA of wideband sources directly. It means that this framework enables cutting-edge techniques in the existing narrowband subspace methods to implement the wideband direction estimation for reducing the computational complexity and facilitating the estimation algorithm. Theoretical analysis and effectiveness of the proposed method are substantiated through numerical simulations and experiments, which are performed in reverberating environments. The results show that performance of the proposed method performs better than others over a range of signal-to-noise ratio with just a few microphones. All these advantages make the proposed method a powerful tool for navigation systems based on acoustic signal processing.
Muhammad Ammar MALIK Jae Young CHOI Moonsoo KANG Bumshik LEE
In most supervised learning problems, the labelling quality of datasets plays a paramount role in the learning of high-performance classifiers. The performance of a classifier can significantly be degraded if it is trained with mislabeled data. Therefore, identification of such examples from the dataset is of critical importance. In this study, we proposed an improved majority filtering algorithm, which utilized the ability of a support vector machine in terms of capturing potentially mislabeled examples as support vectors (SVs). The key technical contribution of our work, is that the base (or component) classifiers that construct the ensemble of classifiers are trained using non-SV examples, although at the time of testing, the examples captured as SVs were employed. An example can be tagged as mislabeled if the majority of the base classifiers incorrectly classifies the example. Experimental results confirmed that our algorithm not only showed high-level accuracy with higher F1 scores, for identifying the mislabeled examples, but was also significantly faster than the previous methods.
Tianwen GUO Ping DENG Qiang YU Baoyun WANG
In this letter, we investigate a design of efficient antenna allocation at the full duplex receiver (FDR) in a multi-input multi-output multi-eavesdropper (MIMOME) wiretap channel for physical layer security improvement. Specifically, we propose the allocation which are feasible for the practical scenario with self-interference (SI) taken into account, because the jamming signals from FDR not only confuse the eavesdropper but also inevitably cause SI at the FDR. Due to the nolinear and coupling of the antenna allocation optimization problem, we transform the original problem into an integer programming problem. Then, we derive the optimal solution and the corresponding beamforming matrices in closed-form by means of combining spatial alignment and null-space projection method. Furthermore, we present the feasibility condition and full-protection condition, which offer insight into principles that enable more efficient and effective use of FDR in the wiretap channel for security improvement. From the simulation results, we validate the theoretical analysis and demonstrate the outstanding performance of the proposed antennas allocation at FDR.
Xiaole LI Hua WANG Shanwen YI Linbo ZHAI
The periodic disaster backup activity among geographically distributed multiple datacenters consumes huge network resources and therefore imposes a heavy burden on datacenters and transmission links. Previous work aims at least completion time, maximum utility or minimal cost, without consideration of load balance for limited network resources, likely to result in unfair distribution of backup load or significant impact on daily network services. In this paper, we propose a new progressive forwarding disaster backup strategy in the Software Defined Network scenarios to mitigate forwarding burdens on source datacenters and balance backup loads on backup datacenters and transmission links. We construct a new redundancy-aware time-expanded network model to divide time slots according to redundancy requirement, and propose role-switching method over time to utilize forwarding capability of backup datacenters. In every time slot, we leverage two-step optimization algorithm to realize capacity-constrained backup datacenter selection and fair backup load distribution. Simulations results prove that our strategy achieves good performance in load balance under the condition of guaranteeing transmission completion and backup redundancy.
Gou HOUBEN Shu FUJITA Keita TAKAHASHI Toshiaki FUJII
Depth (disparity) estimation from a light field (a set of dense multi-view images) is currently attracting much research interest. This paper focuses on how to handle a noisy light field for disparity estimation, because if left as it is, the noise deteriorates the accuracy of estimated disparity maps. Several researchers have worked on this problem, e.g., by introducing disparity cues that are robust to noise. However, it is not easy to break the trade-off between the accuracy and computational speed. To tackle this trade-off, we have integrated a fast denoising scheme in a fast disparity estimation framework that works in the epipolar plane image (EPI) domain. Specifically, we found that a simple 1-D slanted filter is very effective for reducing noise while preserving the underlying structure in an EPI. Moreover, this simple filtering does not require elaborate parameter configurations in accordance with the target noise level. Experimental results including real-world inputs show that our method can achieve good accuracy with much less computational time compared to some state-of-the-art methods.
Xuan WANG Bofeng ZHANG Mingqing HUANG Furong CHANG Zhuocheng ZHOU
When individuals make a purchase from online sources, they may lack first-hand knowledge of the product. In such cases, they will judge the quality of the item by the reviews other consumers have posted. Therefore, it is significant to determine whether comments about a product are credible. Most often, conventional research on comment credibility has employed supervised machine learning methods, which have the disadvantage of needing large quantities of training data. This paper proposes an unsupervised method for judging comment credibility based on the Biterm Sentiment Latent Dirichlet Allocation (BS-LDA) model. Using this approach, first we derived some distributions and calculated each comment's credibility score via them. A comment's credibility was judged based on whether it achieved a threshold score. Our experimental results using comments from Amazon.com demonstrated that the overall performance of our approach can play an important role in determining the credibility of comments in some situation.
Yuan ZHOU Yuichi GOTO Jingde CHENG
Many kinds of questionnaires, testing, and voting are performed in some completely electronic ways to do questions and answers on the Internet as Web applications, i.e. e-questionnaire systems, e-testing systems, and e-voting systems. Because there is no unified communication tool among the stakeholders of e-questionnaire, e-testing, and e-voting systems, until now, all the e-questionnaire, e-testing, and e-voting systems are designed, developed, used, and maintained in various ad hoc ways. As a result, the stakeholders are difficult to communicate to implement the systems, because there is neither an exhaustive requirement list to have a grasp of the overall e-questionnaire, e-testing, and e-voting systems nor a standardized terminology for these systems to avoid ambiguity. A general-purpose specification language to provide a unified description way for specifying various e-questionnaire, e-testing, and e-voting systems can solve the problems such that the stakeholders can refer to and use the complete requirements and standardized terminology for better communications, and can easily and unambiguously specify all the requirements of systems and services of e-questionnaire, e-testing, and e-voting, even can implement the systems. In this paper, we propose the first specification language, named “QSL,” with a standardized, consistent, and exhaustive list of requirements for specifying various e-questionnaire, e-testing, and e-voting systems such that the specifications can be used as the precondition of automatically generating e-questionnaire, e-testing, and e-voting systems. The paper presents our design addressing that QSL can specify all the requirements of various e-questionnaire, e-testing, and e-voting systems in a structured way, evaluates its effectiveness, performs real applications using QSL in case of e-questionnaire, e-testing, and e-voting systems, and shows various QSL applications for providing convenient QSL services to stakeholders.
Chen CHANG Jianjun CAO Qin FENG Nianfeng WENG Yuling SHANG
Most existing truth discovery approaches are designed for structured data, and cannot meet the strong need to extract trustworthy information from raw text data for its unique characteristics such as multifactorial property of text answers (i.e., an answer may contain multiple key factors) and the diversity of word usages (i.e., different words may have the same semantic meaning). As for text answers, there are no absolute correctness or errors, most answers may be partially correct, which is quite different from the situation of traditional truth discovery. To solve these challenges, we propose an optimization-based text truth discovery model which jointly groups keywords extracted from the answers of the specific question into a set of multiple factors. Then, we select the subset of multiple factors as identified truth set for each question by parallel ant colony synchronization optimization algorithm. After that, the answers to each question can be ranked based on the similarities between factors answer provided and identified truth factors. The experiment results on real dataset show that though text data structures are complex, our model can still find reliable answers compared with retrieval-based and state-of-the-art approaches.
Qinglan ZHAO Dong ZHENG Baodong QIN Rui GUO
Semi-bent functions have important applications in cryptography and coding theory. 2-rotation symmetric semi-bent functions are a class of semi-bent functions with the simplicity for efficient computation because of their invariance under 2-cyclic shift. However, no construction of 2-rotation symmetric semi-bent functions with algebraic degree bigger than 2 has been presented in the literature. In this paper, we introduce four classes of 2m-variable 2-rotation symmetric semi-bent functions including balanced ones. Two classes of 2-rotation symmetric semi-bent functions have algebraic degree from 3 to m for odd m≥3, and the other two classes have algebraic degree from 3 to m/2 for even m≥6 with m/2 being odd.
Lu TANG Zhigong WANG Tiantian FAN Faen LIU Changchun ZHANG
In this paper, an improved charge pump (CP) and a modified nonlinear phase frequency detector (PFD) are designed and fabricated in a 90-nm CMOS process. The CP is optimized with a combination of circuit techniques such as pedestal error cancel scheme to eliminate the charge injection and the other non-ideal characteristics. The nonlinear PFD is based on a modified circuit topology to enhance the acquisition capability of the PLL. The optimized CP and nonlinear PFD are integrated into a Ka-band PLL. The measured output current mismatch ratio of the improved CP is less than 1% when the output voltage Vout fluctuates between 0.2 to 1.1V from a 1.2V power supply. The measured phase error detection range of the modified nonlinear PFD is between -2π and 2π. Owing to the modified CP and PFD, the measured reference spur of the Ka-band PLL frequency synthesizer containing the optimized CP and PFD is only -56.409dBc at 30-GHz at the locked state.
Takuya MIYASAKA Hiroshi SATO Masaharu TAKAHASHI
MIMO technology, which uses multiple antennas, has been introduced to the mobile terminal to increase communication capacity per unit frequency. However, MIMO suffers from the problem of mutual coupling. If MIMO antennas are closely packed, as in a small wireless terminal, a strong mutual coupling occurs. The mutual coupling degrades radiation efficiency and channel capacity. As modern terminals are likely to use three MIMO antennas, reducing the mutual coupling 3×3 MIMO is essential. Some decoupling methods for three elements have been proposed. Unfortunately, these methods demand that the elements be cross-wired, which complicates fabrication and raises the cost. In this paper, we propose a non-connected decoupling method that uses short stubs and an insertion inductor and confirms that the proposed model offers excellent decoupling and increased radiation efficiency.
Yuki TAGUCHI Ryota KAWASHIMA Hiroki NAKAYAMA Tsunemasa HAYASHI Hiroshi MATSUO
Many studies have revealed that the performance of software-based Virtual Network Functions (VNFs) is insufficient for mission-critical networks. Scaling-out approaches, such as auto-scaling of VNFs, could handle a huge amount of traffic; however, the exponential traffic growth confronts us the limitations of both expandability of physical resources and complexity of their management. In this paper, we propose a fast datapath processing method called Packet Aggregation Flow (PA-Flow) that is based on hop-by-hop packet aggregation for more efficient Service Function Chaining (SFC). PA-Flow extends a notion of existing intra-node packet aggregation toward network-wide packet aggregation, and we introduce following three novel features. First, packet I/O overheads at intermediate network devices including NFV-nodes are mitigated by reduction of packet amount. Second, aggregated packets are further aggregated as going through the service chain in a hop-by-hop manner. Finally, next-hop aware packet aggregation is realized using OpenFlow-based flow tables. PA-Flow is designed to be available with various VNF forms (e.g. VM/container/baremetal-based) and virtual I/O technologies (e.g. vhost-user/SR-IOV), and its implementation does not bring noticeable delay for aggregation. We conducted two evaluations: (i) a baseline evaluation for understanding fundamental performance characteristics of PA-Flow (ii) a simulation-based SFC evaluation for proving PA-Flow's effect in a realistic environment. The results showed that throughput of short packet forwarding was improved by 4 times. Moreover, the total number of packets was reduced by 93% in a large-scale SFC.
Kazumi TAKEMURA Toshiyuki YOSHIDA
This paper proposes a novel Depth From Defocus (DFD) technique based on the property that two images having different focus settings coincide if they are reblurred with the opposite focus setting, which is referred to as the “cross reblurring” property in this paper. Based on the property, the proposed technique estimates the block-wise depth profile for a target object by minimizing the mean squared error between the cross-reblurred images. Unlike existing DFD techniques, the proposed technique is free of lens parameters and independent of point spread function models. A compensation technique for a possible pixel disalignment between images is also proposed to improve the depth estimation accuracy. The experimental results and comparisons with the other DFD techniques show the advantages of our technique.
Menghan JIA Feiteng LI Zhijian CHEN Xiaoyan XIANG Xiaolang YAN
An R-peak detection method with a high noise tolerance is presented in this paper. This method utilizes a customized deep convolution neural network (DCNN) to extract morphological and temporal features from sliced electrocardiogram (ECG) signals. The proposed network adopts multiple parallel dilated convolution layers to analyze features from diverse fields of view. A sliding window slices the original ECG signals into segments, and then the network calculates one segment at a time and outputs every point's probability of belonging to the R-peak regions. After a binarization and a deburring operation, the occurrence time of the R-peaks can be located. Experimental results based on the MIT-BIH database show that the R-peak detection accuracies can be significantly improved under high intensity of the electrode motion artifact or muscle artifact noise, which reveals a higher performance than state-of-the-art methods.
Min ZHANG Bo XU Xiaoyun LI Dong FU Jian LIU Baojian WU Kun QIU
The capacity of optical transport networks has been increasing steadily and the networks are becoming more dynamic, complex, and transparent. Though it is common to use worst case assumptions for estimating the quality of transmission (QoT) in the physical layer, over provisioning results in high margin requirements. Accurate estimation on the QoT for to-be-established lightpaths is crucial for reducing provisioning margins. Machine learning (ML) is regarded as one of the most powerful methodological approaches to perform network data analysis and enable automated network self-configuration. In this paper, an artificial neural network (ANN) framework, a branch of ML, to estimate the optical signal-to-noise ratio (OSNR) of to-be-established lightpaths is proposed. It takes account of both nonlinear interference between spectrum neighboring channels and optical monitoring uncertainties. The link information vector of the lightpath is used as input and the OSNR of the lightpath is the target for output of the ANN. The nonlinear interference impact of the number of neighboring channels on the estimation accuracy is considered. Extensive simulation results show that the proposed OSNR estimation scheme can work with any RWA algorithm. High estimation accuracy of over 98% with estimation errors of less than 0.5dB can be achieved given enough training data. ANN model with R=4 neighboring channels should be used to achieve more accurate OSNR estimates. Based on the results, it is expected that the proposed ANN-based OSNR estimation for new lightpath provisioning can be a promising tool for margin reduction and low-cost operation of future optical transport networks.
Siaw-Lang WONG Raveendran PARAMESRAN Ibuki YOSHIDA Akira TAGUCHI
Light scattering and absorption of light in water cause underwater images to be poorly contrasted, haze and dominated by a single color cast. A solution to this is to find methods to improve the quality of the image that eventually leads to better visualization. We propose an integrated approach using Adaptive Gray World (AGW) and Differential Gray-Levels Histogram Equalization for Color Images (DHECI) to remove the color cast as well as improve the contrast and colorfulness of the underwater image. The AGW is an adaptive version of the GW method where apart from computing the global mean, the local mean of each channel of an image is taken into consideration and both are weighted before combining them. It is applied to remove the color cast, thereafter the DHECI is used to improve the contrast and colorfulness of the underwater image. The results of the proposed method are compared with seven state-of-the-art methods using qualitative and quantitative measures. The experimental results showed that in most cases the proposed method produced better quantitative scores than the compared methods.
This paper deals with the problem of minimizing roundoff noise and pole sensitivity simultaneously subject to l2-scaling constraints for state-space digital filters. A novel measure for evaluating roundoff noise and pole sensitivity is proposed, and an efficient technique for minimizing this measure by jointly optimizing state-space realization and error feedback is explored, namely, the constrained optimization problem at hand is converted into an unconstrained problem and then the resultant problem is solved by employing a quasi-Newton algorithm. A numerical example is presented to demonstrate the validity and effectiveness of the proposed technique.
Fill-a-Pix is a pencil-and-paper puzzle, which is popular worldwide since announced by Conceptis in 2003. It provides a rectangular grid of squares that must be filled in to create a picture. Precisely, we are given a rectangular grid of squares some of which has an integer from 0 to 9 in it, and our task is to paint some squares black so that every square with an integer has the same number of painted squares around it including the square itself. Despite its popularity, computational complexity of Fill-a-Pix has not been known. We in this paper show that the puzzle is NP-complete, ASP-complete, and #P-complete via a parsimonious reduction from the Boolean satisfiability problem. We also consider the fewest clues problem of Fill-a-Pix, where the fewest clues problem is recently introduced by Demaine et al. for analyzing computational complexity of designing “good” puzzles. We show that the fewest clues problem of Fill-a-Pix is Σ2P-complete.
This letter proposes a novel fractional frequency reuse (FFR) scheme with hybrid-beam trisector cell (HBTC) architectures that combine narrow-beam trisector cell (NBTC) and wide-beam trisector cell (WBTC) architectures to increase the system performance of cellular mobile networks. In the proposed FFR scheme, the macro base station first divides its macro user equipments (MUEs) into two groups, i.e., inner group (IG) and outer group (OG), based on the signal to interference plus noise ratio (SINR) of MUEs and then assigns subchannels to the MUEs in the IG and OG using the NBTC and WBTC antennas, respectively. Through simulation results, it is shown that the proposed FFR scheme outperforms other FFR schemes in terms of the SINR and capacity of MUEs.
Xinbo REN Haiyuan WU Qian CHEN Toshiyuki IMAI Takashi KUBO Takashi AKASAKA
Clinical researches show that the morbidity of coronary artery disease (CAD) is gradually increasing in many countries every year, and it causes hundreds of thousands of people all over the world dying for each year. As the optical coherence tomography with high resolution and better contrast applied to the lesion tissue investigation of human vessel, many more micro-structures of the vessel could be easily and clearly visible to doctors, which help to improve the CAD treatment effect. Manual qualitative analysis and classification of vessel lesion tissue are time-consuming to doctors because a single-time intravascular optical coherence (IVOCT) data set of a patient usually contains hundreds of in-vivo vessel images. To overcome this problem, we focus on the investigation of the superficial layer of the lesion region and propose a model based on local multi-layer region for vessel lesion components (lipid, fibrous and calcified plaque) features characterization and extraction. At the pre-processing stage, we applied two novel automatic methods to remove the catheter and guide-wire respectively. Based on the detected lumen boundary, the multi-layer model in the proximity lumen boundary region (PLBR) was built. In the multi-layer model, features extracted from the A-line sub-region (ALSR) of each layer was employed to characterize the type of the tissue existing in the ALSR. We used 7 human datasets containing total 490 OCT images to assess our tissue classification method. Validation was obtained by comparing the manual assessment with the automatic results derived by our method. The proposed automatic tissue classification method achieved an average accuracy of 89.53%, 93.81% and 91.78% for fibrous, calcified and lipid plaque respectively.