The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] (42807hit)

3661-3680hit(42807hit)

  • A Highly Efficient Wideband Two-Dimensional Direction Estimation Method with L-Shaped Microphone Array

    Bandhit SUKSIRI  Masahiro FUKUMOTO  

     
    PAPER-Digital Signal Processing

      Vol:
    E102-A No:11
      Page(s):
    1457-1472

    This paper presents an efficient wideband two-dimensional direction-of-arrival (DOA) estimation for an L-shaped microphone array. We propose a way to construct a wideband sample cross-correlation matrix without any process of DOA preliminary estimation, such as beamforming technique, by exploiting sample cross-correlation matrices of two different frequencies for all frequency bins. Subsequently, wideband DOAs can be estimated by using this wideband matrix along with a scheme of estimating DOA in a narrowband subspace method. Therefore, a contribution of our study is providing an alternative framework for recent narrowband subspace methods to estimating the DOA of wideband sources directly. It means that this framework enables cutting-edge techniques in the existing narrowband subspace methods to implement the wideband direction estimation for reducing the computational complexity and facilitating the estimation algorithm. Theoretical analysis and effectiveness of the proposed method are substantiated through numerical simulations and experiments, which are performed in reverberating environments. The results show that performance of the proposed method performs better than others over a range of signal-to-noise ratio with just a few microphones. All these advantages make the proposed method a powerful tool for navigation systems based on acoustic signal processing.

  • Improved Majority Filtering Algorithm for Cleaning Class Label Noise in Supervised Learning

    Muhammad Ammar MALIK  Jae Young CHOI  Moonsoo KANG  Bumshik LEE  

     
    LETTER-Digital Signal Processing

      Vol:
    E102-A No:11
      Page(s):
    1556-1559

    In most supervised learning problems, the labelling quality of datasets plays a paramount role in the learning of high-performance classifiers. The performance of a classifier can significantly be degraded if it is trained with mislabeled data. Therefore, identification of such examples from the dataset is of critical importance. In this study, we proposed an improved majority filtering algorithm, which utilized the ability of a support vector machine in terms of capturing potentially mislabeled examples as support vectors (SVs). The key technical contribution of our work, is that the base (or component) classifiers that construct the ensemble of classifiers are trained using non-SV examples, although at the time of testing, the examples captured as SVs were employed. An example can be tagged as mislabeled if the majority of the base classifiers incorrectly classifies the example. Experimental results confirmed that our algorithm not only showed high-level accuracy with higher F1 scores, for identifying the mislabeled examples, but was also significantly faster than the previous methods.

  • Antenna Allocation of Full Duplex Receiver for Security Improvement of the MIMOME Wiretap Channel with Self-Interference Cancellation

    Tianwen GUO  Ping DENG  Qiang YU  Baoyun WANG  

     
    LETTER-Digital Signal Processing

      Vol:
    E102-A No:11
      Page(s):
    1560-1565

    In this letter, we investigate a design of efficient antenna allocation at the full duplex receiver (FDR) in a multi-input multi-output multi-eavesdropper (MIMOME) wiretap channel for physical layer security improvement. Specifically, we propose the allocation which are feasible for the practical scenario with self-interference (SI) taken into account, because the jamming signals from FDR not only confuse the eavesdropper but also inevitably cause SI at the FDR. Due to the nolinear and coupling of the antenna allocation optimization problem, we transform the original problem into an integer programming problem. Then, we derive the optimal solution and the corresponding beamforming matrices in closed-form by means of combining spatial alignment and null-space projection method. Furthermore, we present the feasibility condition and full-protection condition, which offer insight into principles that enable more efficient and effective use of FDR in the wiretap channel for security improvement. From the simulation results, we validate the theoretical analysis and demonstrate the outstanding performance of the proposed antennas allocation at FDR.

  • Progressive Forwarding Disaster Backup among Cloud Datacenters

    Xiaole LI  Hua WANG  Shanwen YI  Linbo ZHAI  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2019/08/19
      Vol:
    E102-D No:11
      Page(s):
    2135-2147

    The periodic disaster backup activity among geographically distributed multiple datacenters consumes huge network resources and therefore imposes a heavy burden on datacenters and transmission links. Previous work aims at least completion time, maximum utility or minimal cost, without consideration of load balance for limited network resources, likely to result in unfair distribution of backup load or significant impact on daily network services. In this paper, we propose a new progressive forwarding disaster backup strategy in the Software Defined Network scenarios to mitigate forwarding burdens on source datacenters and balance backup loads on backup datacenters and transmission links. We construct a new redundancy-aware time-expanded network model to divide time slots according to redundancy requirement, and propose role-switching method over time to utilize forwarding capability of backup datacenters. In every time slot, we leverage two-step optimization algorithm to realize capacity-constrained backup datacenter selection and fair backup load distribution. Simulations results prove that our strategy achieves good performance in load balance under the condition of guaranteeing transmission completion and backup redundancy.

  • Fast and Robust Disparity Estimation from Noisy Light Fields Using 1-D Slanted Filters

    Gou HOUBEN  Shu FUJITA  Keita TAKAHASHI  Toshiaki FUJII  

     
    PAPER

      Pubricized:
    2019/07/03
      Vol:
    E102-D No:11
      Page(s):
    2101-2109

    Depth (disparity) estimation from a light field (a set of dense multi-view images) is currently attracting much research interest. This paper focuses on how to handle a noisy light field for disparity estimation, because if left as it is, the noise deteriorates the accuracy of estimated disparity maps. Several researchers have worked on this problem, e.g., by introducing disparity cues that are robust to noise. However, it is not easy to break the trade-off between the accuracy and computational speed. To tackle this trade-off, we have integrated a fast denoising scheme in a fast disparity estimation framework that works in the epipolar plane image (EPI) domain. Specifically, we found that a simple 1-D slanted filter is very effective for reducing noise while preserving the underlying structure in an EPI. Moreover, this simple filtering does not require elaborate parameter configurations in accordance with the target noise level. Experimental results including real-world inputs show that our method can achieve good accuracy with much less computational time compared to some state-of-the-art methods.

  • Improved LDA Model for Credibility Evaluation of Online Product Reviews

    Xuan WANG  Bofeng ZHANG  Mingqing HUANG  Furong CHANG  Zhuocheng ZHOU  

     
    PAPER-Data Engineering, Web Information Systems

      Pubricized:
    2019/08/22
      Vol:
    E102-D No:11
      Page(s):
    2148-2158

    When individuals make a purchase from online sources, they may lack first-hand knowledge of the product. In such cases, they will judge the quality of the item by the reviews other consumers have posted. Therefore, it is significant to determine whether comments about a product are credible. Most often, conventional research on comment credibility has employed supervised machine learning methods, which have the disadvantage of needing large quantities of training data. This paper proposes an unsupervised method for judging comment credibility based on the Biterm Sentiment Latent Dirichlet Allocation (BS-LDA) model. Using this approach, first we derived some distributions and calculated each comment's credibility score via them. A comment's credibility was judged based on whether it achieved a threshold score. Our experimental results using comments from Amazon.com demonstrated that the overall performance of our approach can play an important role in determining the credibility of comments in some situation.

  • QSL: A Specification Language for E-Questionnaire, E-Testing, and E-Voting Systems

    Yuan ZHOU  Yuichi GOTO  Jingde CHENG  

     
    PAPER-Data Engineering, Web Information Systems

      Pubricized:
    2019/08/19
      Vol:
    E102-D No:11
      Page(s):
    2159-2175

    Many kinds of questionnaires, testing, and voting are performed in some completely electronic ways to do questions and answers on the Internet as Web applications, i.e. e-questionnaire systems, e-testing systems, and e-voting systems. Because there is no unified communication tool among the stakeholders of e-questionnaire, e-testing, and e-voting systems, until now, all the e-questionnaire, e-testing, and e-voting systems are designed, developed, used, and maintained in various ad hoc ways. As a result, the stakeholders are difficult to communicate to implement the systems, because there is neither an exhaustive requirement list to have a grasp of the overall e-questionnaire, e-testing, and e-voting systems nor a standardized terminology for these systems to avoid ambiguity. A general-purpose specification language to provide a unified description way for specifying various e-questionnaire, e-testing, and e-voting systems can solve the problems such that the stakeholders can refer to and use the complete requirements and standardized terminology for better communications, and can easily and unambiguously specify all the requirements of systems and services of e-questionnaire, e-testing, and e-voting, even can implement the systems. In this paper, we propose the first specification language, named “QSL,” with a standardized, consistent, and exhaustive list of requirements for specifying various e-questionnaire, e-testing, and e-voting systems such that the specifications can be used as the precondition of automatically generating e-questionnaire, e-testing, and e-voting systems. The paper presents our design addressing that QSL can specify all the requirements of various e-questionnaire, e-testing, and e-voting systems in a structured way, evaluates its effectiveness, performs real applications using QSL in case of e-questionnaire, e-testing, and e-voting systems, and shows various QSL applications for providing convenient QSL services to stakeholders.

  • Truth Discovery of Multi-Source Text Data

    Chen CHANG  Jianjun CAO  Qin FENG  Nianfeng WENG  Yuling SHANG  

     
    LETTER-Fundamentals of Information Systems

      Pubricized:
    2019/08/22
      Vol:
    E102-D No:11
      Page(s):
    2249-2252

    Most existing truth discovery approaches are designed for structured data, and cannot meet the strong need to extract trustworthy information from raw text data for its unique characteristics such as multifactorial property of text answers (i.e., an answer may contain multiple key factors) and the diversity of word usages (i.e., different words may have the same semantic meaning). As for text answers, there are no absolute correctness or errors, most answers may be partially correct, which is quite different from the situation of traditional truth discovery. To solve these challenges, we propose an optimization-based text truth discovery model which jointly groups keywords extracted from the answers of the specific question into a set of multiple factors. Then, we select the subset of multiple factors as identified truth set for each question by parallel ant colony synchronization optimization algorithm. After that, the answers to each question can be ranked based on the similarities between factors answer provided and identified truth factors. The experiment results on real dataset show that though text data structures are complex, our model can still find reliable answers compared with retrieval-based and state-of-the-art approaches.

  • Constructions of 2-Rotation Symmetric Semi-Bent Functions with Degree Bigger than 2

    Qinglan ZHAO  Dong ZHENG  Baodong QIN   Rui GUO  

     
    PAPER-Cryptography and Information Security

      Vol:
    E102-A No:11
      Page(s):
    1497-1503

    Semi-bent functions have important applications in cryptography and coding theory. 2-rotation symmetric semi-bent functions are a class of semi-bent functions with the simplicity for efficient computation because of their invariance under 2-cyclic shift. However, no construction of 2-rotation symmetric semi-bent functions with algebraic degree bigger than 2 has been presented in the literature. In this paper, we introduce four classes of 2m-variable 2-rotation symmetric semi-bent functions including balanced ones. Two classes of 2-rotation symmetric semi-bent functions have algebraic degree from 3 to m for odd m≥3, and the other two classes have algebraic degree from 3 to m/2 for even m≥6 with m/2 being odd.

  • Optimized Charge Pump and Nonlinear Phase Frequency Detector for a Ka-Band Phase-Locked Loop in 90-nm CMOS Process

    Lu TANG  Zhigong WANG  Tiantian FAN  Faen LIU  Changchun ZHANG  

     
    PAPER-Electronic Circuits

      Pubricized:
    2019/06/07
      Vol:
    E102-C No:11
      Page(s):
    825-832

    In this paper, an improved charge pump (CP) and a modified nonlinear phase frequency detector (PFD) are designed and fabricated in a 90-nm CMOS process. The CP is optimized with a combination of circuit techniques such as pedestal error cancel scheme to eliminate the charge injection and the other non-ideal characteristics. The nonlinear PFD is based on a modified circuit topology to enhance the acquisition capability of the PLL. The optimized CP and nonlinear PFD are integrated into a Ka-band PLL. The measured output current mismatch ratio of the improved CP is less than 1% when the output voltage Vout fluctuates between 0.2 to 1.1V from a 1.2V power supply. The measured phase error detection range of the modified nonlinear PFD is between -2π and 2π. Owing to the modified CP and PFD, the measured reference spur of the Ka-band PLL frequency synthesizer containing the optimized CP and PFD is only -56.409dBc at 30-GHz at the locked state.

  • A Non-Connected Decoupling Method for Three Element MIMO Antennas by Using Short Stubs Open Access

    Takuya MIYASAKA  Hiroshi SATO  Masaharu TAKAHASHI  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2019/05/07
      Vol:
    E102-B No:11
      Page(s):
    2167-2173

    MIMO technology, which uses multiple antennas, has been introduced to the mobile terminal to increase communication capacity per unit frequency. However, MIMO suffers from the problem of mutual coupling. If MIMO antennas are closely packed, as in a small wireless terminal, a strong mutual coupling occurs. The mutual coupling degrades radiation efficiency and channel capacity. As modern terminals are likely to use three MIMO antennas, reducing the mutual coupling 3×3 MIMO is essential. Some decoupling methods for three elements have been proposed. Unfortunately, these methods demand that the elements be cross-wired, which complicates fabrication and raises the cost. In this paper, we propose a non-connected decoupling method that uses short stubs and an insertion inductor and confirms that the proposed model offers excellent decoupling and increased radiation efficiency.

  • Fast Datapath Processing Based on Hop-by-Hop Packet Aggregation for Service Function Chaining Open Access

    Yuki TAGUCHI  Ryota KAWASHIMA  Hiroki NAKAYAMA  Tsunemasa HAYASHI  Hiroshi MATSUO  

     
    PAPER-Information Network

      Pubricized:
    2019/08/22
      Vol:
    E102-D No:11
      Page(s):
    2184-2194

    Many studies have revealed that the performance of software-based Virtual Network Functions (VNFs) is insufficient for mission-critical networks. Scaling-out approaches, such as auto-scaling of VNFs, could handle a huge amount of traffic; however, the exponential traffic growth confronts us the limitations of both expandability of physical resources and complexity of their management. In this paper, we propose a fast datapath processing method called Packet Aggregation Flow (PA-Flow) that is based on hop-by-hop packet aggregation for more efficient Service Function Chaining (SFC). PA-Flow extends a notion of existing intra-node packet aggregation toward network-wide packet aggregation, and we introduce following three novel features. First, packet I/O overheads at intermediate network devices including NFV-nodes are mitigated by reduction of packet amount. Second, aggregated packets are further aggregated as going through the service chain in a hop-by-hop manner. Finally, next-hop aware packet aggregation is realized using OpenFlow-based flow tables. PA-Flow is designed to be available with various VNF forms (e.g. VM/container/baremetal-based) and virtual I/O technologies (e.g. vhost-user/SR-IOV), and its implementation does not bring noticeable delay for aggregation. We conducted two evaluations: (i) a baseline evaluation for understanding fundamental performance characteristics of PA-Flow (ii) a simulation-based SFC evaluation for proving PA-Flow's effect in a realistic environment. The results showed that throughput of short packet forwarding was improved by 4 times. Moreover, the total number of packets was reduced by 93% in a large-scale SFC.

  • Depth from Defocus Technique Based on Cross Reblurring

    Kazumi TAKEMURA  Toshiyuki YOSHIDA  

     
    PAPER

      Pubricized:
    2019/07/11
      Vol:
    E102-D No:11
      Page(s):
    2083-2092

    This paper proposes a novel Depth From Defocus (DFD) technique based on the property that two images having different focus settings coincide if they are reblurred with the opposite focus setting, which is referred to as the “cross reblurring” property in this paper. Based on the property, the proposed technique estimates the block-wise depth profile for a target object by minimizing the mean squared error between the cross-reblurred images. Unlike existing DFD techniques, the proposed technique is free of lens parameters and independent of point spread function models. A compensation technique for a possible pixel disalignment between images is also proposed to improve the depth estimation accuracy. The experimental results and comparisons with the other DFD techniques show the advantages of our technique.

  • High Noise Tolerant R-Peak Detection Method Based on Deep Convolution Neural Network

    Menghan JIA  Feiteng LI  Zhijian CHEN  Xiaoyan XIANG  Xiaolang YAN  

     
    LETTER-Biological Engineering

      Pubricized:
    2019/08/02
      Vol:
    E102-D No:11
      Page(s):
    2272-2275

    An R-peak detection method with a high noise tolerance is presented in this paper. This method utilizes a customized deep convolution neural network (DCNN) to extract morphological and temporal features from sliced electrocardiogram (ECG) signals. The proposed network adopts multiple parallel dilated convolution layers to analyze features from diverse fields of view. A sliding window slices the original ECG signals into segments, and then the network calculates one segment at a time and outputs every point's probability of belonging to the R-peak regions. After a binarization and a deburring operation, the occurrence time of the R-peaks can be located. Experimental results based on the MIT-BIH database show that the R-peak detection accuracies can be significantly improved under high intensity of the electrode motion artifact or muscle artifact noise, which reveals a higher performance than state-of-the-art methods.

  • Artificial Neural Network-Based QoT Estimation for Lightpath Provisioning in Optical Networks

    Min ZHANG  Bo XU  Xiaoyun LI  Dong FU  Jian LIU  Baojian WU  Kun QIU  

     
    PAPER-Network

      Pubricized:
    2019/05/16
      Vol:
    E102-B No:11
      Page(s):
    2104-2112

    The capacity of optical transport networks has been increasing steadily and the networks are becoming more dynamic, complex, and transparent. Though it is common to use worst case assumptions for estimating the quality of transmission (QoT) in the physical layer, over provisioning results in high margin requirements. Accurate estimation on the QoT for to-be-established lightpaths is crucial for reducing provisioning margins. Machine learning (ML) is regarded as one of the most powerful methodological approaches to perform network data analysis and enable automated network self-configuration. In this paper, an artificial neural network (ANN) framework, a branch of ML, to estimate the optical signal-to-noise ratio (OSNR) of to-be-established lightpaths is proposed. It takes account of both nonlinear interference between spectrum neighboring channels and optical monitoring uncertainties. The link information vector of the lightpath is used as input and the OSNR of the lightpath is the target for output of the ANN. The nonlinear interference impact of the number of neighboring channels on the estimation accuracy is considered. Extensive simulation results show that the proposed OSNR estimation scheme can work with any RWA algorithm. High estimation accuracy of over 98% with estimation errors of less than 0.5dB can be achieved given enough training data. ANN model with R=4 neighboring channels should be used to achieve more accurate OSNR estimates. Based on the results, it is expected that the proposed ANN-based OSNR estimation for new lightpath provisioning can be a promising tool for margin reduction and low-cost operation of future optical transport networks.

  • An Integrated Method to Remove Color Cast and Contrast Enhancement for Underwater Image Open Access

    Siaw-Lang WONG  Raveendran PARAMESRAN  Ibuki YOSHIDA  Akira TAGUCHI  

     
    PAPER-Image

      Vol:
    E102-A No:11
      Page(s):
    1524-1532

    Light scattering and absorption of light in water cause underwater images to be poorly contrasted, haze and dominated by a single color cast. A solution to this is to find methods to improve the quality of the image that eventually leads to better visualization. We propose an integrated approach using Adaptive Gray World (AGW) and Differential Gray-Levels Histogram Equalization for Color Images (DHECI) to remove the color cast as well as improve the contrast and colorfulness of the underwater image. The AGW is an adaptive version of the GW method where apart from computing the global mean, the local mean of each channel of an image is taken into consideration and both are weighted before combining them. It is applied to remove the color cast, thereafter the DHECI is used to improve the contrast and colorfulness of the underwater image. The results of the proposed method are compared with seven state-of-the-art methods using qualitative and quantitative measures. The experimental results showed that in most cases the proposed method produced better quantitative scores than the compared methods.

  • Weighted Minimization of Roundoff Noise and Pole Sensitivity Subject to l2-Scaling Constraints for State-Space Digital Filters

    Yoichi HINAMOTO  Akimitsu DOI  

     
    PAPER-Digital Signal Processing

      Vol:
    E102-A No:11
      Page(s):
    1473-1480

    This paper deals with the problem of minimizing roundoff noise and pole sensitivity simultaneously subject to l2-scaling constraints for state-space digital filters. A novel measure for evaluating roundoff noise and pole sensitivity is proposed, and an efficient technique for minimizing this measure by jointly optimizing state-space realization and error feedback is explored, namely, the constrained optimization problem at hand is converted into an unconstrained problem and then the resultant problem is solved by employing a quasi-Newton algorithm. A numerical example is presented to demonstrate the validity and effectiveness of the proposed technique.

  • NP-Completeness of Fill-a-Pix and ΣP2-Completeness of Its Fewest Clues Problem

    Yuta HIGUCHI  Kei KIMURA  

     
    PAPER-Algorithms and Data Structures

      Vol:
    E102-A No:11
      Page(s):
    1490-1496

    Fill-a-Pix is a pencil-and-paper puzzle, which is popular worldwide since announced by Conceptis in 2003. It provides a rectangular grid of squares that must be filled in to create a picture. Precisely, we are given a rectangular grid of squares some of which has an integer from 0 to 9 in it, and our task is to paint some squares black so that every square with an integer has the same number of painted squares around it including the square itself. Despite its popularity, computational complexity of Fill-a-Pix has not been known. We in this paper show that the puzzle is NP-complete, ASP-complete, and #P-complete via a parsimonious reduction from the Boolean satisfiability problem. We also consider the fewest clues problem of Fill-a-Pix, where the fewest clues problem is recently introduced by Demaine et al. for analyzing computational complexity of designing “good” puzzles. We show that the fewest clues problem of Fill-a-Pix is Σ2P-complete.

  • Fractional Frequency Reuse with Hybrid-Beam Trisector Cell Architectures for Cellular Mobile Networks

    Ilhak BAN  Se-Jin KIM  

     
    LETTER-Digital Signal Processing

      Vol:
    E102-A No:11
      Page(s):
    1586-1589

    This letter proposes a novel fractional frequency reuse (FFR) scheme with hybrid-beam trisector cell (HBTC) architectures that combine narrow-beam trisector cell (NBTC) and wide-beam trisector cell (WBTC) architectures to increase the system performance of cellular mobile networks. In the proposed FFR scheme, the macro base station first divides its macro user equipments (MUEs) into two groups, i.e., inner group (IG) and outer group (OG), based on the signal to interference plus noise ratio (SINR) of MUEs and then assigns subchannels to the MUEs in the IG and OG using the NBTC and WBTC antennas, respectively. Through simulation results, it is shown that the proposed FFR scheme outperforms other FFR schemes in terms of the SINR and capacity of MUEs.

  • A Local Multi-Layer Model for Tissue Classification of in-vivo Atherosclerotic Plaques in Intravascular Optical Coherence Tomography

    Xinbo REN  Haiyuan WU  Qian CHEN  Toshiyuki IMAI  Takashi KUBO  Takashi AKASAKA  

     
    PAPER-Biological Engineering

      Pubricized:
    2019/08/15
      Vol:
    E102-D No:11
      Page(s):
    2238-2248

    Clinical researches show that the morbidity of coronary artery disease (CAD) is gradually increasing in many countries every year, and it causes hundreds of thousands of people all over the world dying for each year. As the optical coherence tomography with high resolution and better contrast applied to the lesion tissue investigation of human vessel, many more micro-structures of the vessel could be easily and clearly visible to doctors, which help to improve the CAD treatment effect. Manual qualitative analysis and classification of vessel lesion tissue are time-consuming to doctors because a single-time intravascular optical coherence (IVOCT) data set of a patient usually contains hundreds of in-vivo vessel images. To overcome this problem, we focus on the investigation of the superficial layer of the lesion region and propose a model based on local multi-layer region for vessel lesion components (lipid, fibrous and calcified plaque) features characterization and extraction. At the pre-processing stage, we applied two novel automatic methods to remove the catheter and guide-wire respectively. Based on the detected lumen boundary, the multi-layer model in the proximity lumen boundary region (PLBR) was built. In the multi-layer model, features extracted from the A-line sub-region (ALSR) of each layer was employed to characterize the type of the tissue existing in the ALSR. We used 7 human datasets containing total 490 OCT images to assess our tissue classification method. Validation was obtained by comparing the manual assessment with the automatic results derived by our method. The proposed automatic tissue classification method achieved an average accuracy of 89.53%, 93.81% and 91.78% for fibrous, calcified and lipid plaque respectively.

3661-3680hit(42807hit)