The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] (42807hit)

4981-5000hit(42807hit)

  • Enhancement of Video Streaming QoE by Considering Burst Loss in Wireless LANs

    Toshiro NUNOME  Yuta MATSUI  

     
    PAPER

      Pubricized:
    2018/01/22
      Vol:
    E101-B No:7
      Page(s):
    1653-1660

    In order to enhance QoE of audio and video IP transmission, this paper proposes a method for mitigating the spatial quality impairment during burst loss periods over the wireless networks in the video output scheme SCS, which is a QoE-based video output scheme. SCS switches between two common video output schemes: frame skipping and error concealment. The proposed method pauses video output with an undamaged frame during the burst loss period in order not to pause video output on a degraded frame. We perform an experiment with constant thresholds, the table-lookup method, and the proposed method under various network conditions. The result shows that the effect of the proposed method on QoE can differ with the contents and GOP structures.

  • Toward In-Network Deep Machine Learning for Identifying Mobile Applications and Enabling Application Specific Network Slicing Open Access

    Akihiro NAKAO  Ping DU  

     
    INVITED PAPER

      Pubricized:
    2018/01/22
      Vol:
    E101-B No:7
      Page(s):
    1536-1543

    In this paper, we posit that, in future mobile network, network softwarization will be prevalent, and it becomes important to utilize deep machine learning within network to classify mobile traffic into fine grained slices, by identifying application types and devices so that we can apply Quality-of-Service (QoS) control, mobile edge/multi-access computing, and various network function per application and per device. This paper reports our initial attempt to apply deep machine learning for identifying application types from actual mobile network traffic captured from an MVNO, mobile virtual network operator and to design the system for classifying it to application specific slices.

  • Crowd Gathering Detection Based on the Foreground Stillness Model

    Chun-Yu LIU  Wei-Hao LIAO  Shanq-Jang RUAN  

     
    LETTER-Image Recognition, Computer Vision

      Pubricized:
    2018/03/30
      Vol:
    E101-D No:7
      Page(s):
    1968-1971

    The abnormal crowd behavior detection is an important research topic in computer vision to improve the response time of critical events. In this letter, we introduce a novel method to detect and localize the crowd gathering in surveillance videos. The proposed foreground stillness model is based on the foreground object mask and the dense optical flow to measure the instantaneous crowd stillness level. Further, we obtain the long-term crowd stillness level by the leaky bucket model, and the crowd gathering behavior can be detected by the threshold analysis. Experimental results indicate that our proposed approach can detect and locate crowd gathering events, and it is capable of distinguishing between standing and walking crowd. The experiments in realistic scenes with 88.65% accuracy for detection of gathering frames show that our method is effective for crowd gathering behavior detection.

  • Fast Time-Aware Sparse Trajectories Prediction with Tensor Factorization

    Lei ZHANG  Qingfu FAN  Guoxing ZHANG  Zhizheng LIANG  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2018/04/13
      Vol:
    E101-D No:7
      Page(s):
    1959-1962

    Existing trajectory prediction methods suffer from the “data sparsity” and neglect “time awareness”, which leads to low accuracy. Aiming to the problem, we propose a fast time-aware sparse trajectories prediction with tensor factorization method (TSTP-TF). Firstly, we do trajectory synthesis based on trajectory entropy and put synthesized trajectories into the original trajectory space. It resolves the sparse problem of trajectory data and makes the new trajectory space more reliable. Then, we introduce multidimensional tensor modeling into Markov model to add the time dimension. Tensor factorization is adopted to infer the missing regions transition probabilities to further solve the problem of data sparsity. Due to the scale of the tensor, we design a divide and conquer tensor factorization model to reduce memory consumption and speed up decomposition. Experiments with real dataset show that TSTP-TF improves prediction accuracy generally by as much as 9% and 2% compared to the Baseline algorithm and ESTP-MF algorithm, respectively.

  • Improved Wolf Pack Algorithm Based on Differential Evolution Elite Set

    Xiayang CHEN  Chaojing TANG  Jian WANG  Lei ZHANG  Qingkun MENG  

     
    LETTER-Fundamentals of Information Systems

      Pubricized:
    2018/03/30
      Vol:
    E101-D No:7
      Page(s):
    1946-1949

    Although Wolf Pack Algorithm (WPA) is a novel optimal algorithm with good performance, there is still room for improvement with respect to its convergence. In order to speed up its convergence and strengthen the search ability, we improve WPA with the Differential Evolution (DE) elite set strategy. The new proposed algorithm is called the WPADEES for short. WPADEES is faster than WPA in convergence, and it has a more feasible adaptability for various optimizations. Six standard benchmark functions are applied to verify the effects of these improvements. Our experiments show that the performance of WPADEES is superior to the standard WPA and other intelligence optimal algorithms, such as GA, DE, PSO, and ABC, in several situations.

  • Identifying Core Objects for Trace Summarization by Analyzing Reference Relations and Dynamic Properties

    Kunihiro NODA  Takashi KOBAYASHI  Noritoshi ATSUMI  

     
    PAPER

      Pubricized:
    2018/04/20
      Vol:
    E101-D No:7
      Page(s):
    1751-1765

    Behaviors of an object-oriented system can be visualized as reverse-engineered sequence diagrams from execution traces. This approach is a valuable tool for program comprehension tasks. However, owing to the massiveness of information contained in an execution trace, a reverse-engineered sequence diagram is often afflicted by a scalability issue. To address this issue, many trace summarization techniques have been proposed. Most of the previous techniques focused on reducing the vertical size of the diagram. To cope with the scalability issue, decreasing the horizontal size of the diagram is also very important. Nonetheless, few studies have addressed this point; thus, there is a lot of needs for further development of horizontal summarization techniques. We present in this paper a method for identifying core objects for trace summarization by analyzing reference relations and dynamic properties. Visualizing only interactions related to core objects, we can obtain a horizontally compactified reverse-engineered sequence diagram that contains system's key behaviors. To identify core objects, first, we detect and eliminate temporary objects that are trivial for a system by analyzing reference relations and lifetimes of objects. Then, estimating the importance of each non-trivial object based on their dynamic properties, we identify highly important ones (i.e., core objects). We implemented our technique in our tool and evaluated it by using traces from various open-source software systems. The results showed that our technique was much more effective in terms of the horizontal reduction of a reverse-engineered sequence diagram, compared with the state-of-the-art trace summarization technique. The horizontal compression ratio of our technique was 134.6 on average, whereas that of the state-of-the-art technique was 11.5. The runtime overhead imposed by our technique was 167.6% on average. This overhead is relatively small compared with recent scalable dynamic analysis techniques, which shows the practicality of our technique. Overall, our technique can achieve a significant reduction of the horizontal size of a reverse-engineered sequence diagram with a small overhead and is expected to be a valuable tool for program comprehension.

  • Character Feature Learning for Named Entity Recognition

    Ping ZENG  Qingping TAN  Haoyu ZHANG  Xiankai MENG  Zhuo ZHANG  Jianjun XU  Yan LEI  

     
    LETTER

      Pubricized:
    2018/04/20
      Vol:
    E101-D No:7
      Page(s):
    1811-1815

    The deep neural named entity recognition model automatically learns and extracts the features of entities and solves the problem of the traditional model relying heavily on complex feature engineering and obscure professional knowledge. This issue has become a hot topic in recent years. Existing deep neural models only involve simple character learning and extraction methods, which limit their capability. To further explore the performance of deep neural models, we propose two character feature learning models based on convolution neural network and long short-term memory network. These two models consider the local semantic and position features of word characters. Experiments conducted on the CoNLL-2003 dataset show that the proposed models outperform traditional ones and demonstrate excellent performance.

  • Stereophonic Music Separation Based on Non-Negative Tensor Factorization with Cepstral Distance Regularization

    Shogo SEKI  Tomoki TODA  Kazuya TAKEDA  

     
    PAPER-Engineering Acoustics

      Vol:
    E101-A No:7
      Page(s):
    1057-1064

    This paper proposes a semi-supervised source separation method for stereophonic music signals containing multiple recorded or processed signals, where synthesized music is focused on the stereophonic music. As the synthesized music signals are often generated as linear combinations of many individual source signals and their respective mixing gains, phase or phase difference information between inter-channel signals, which represent spatial characteristics of recording environments, cannot be utilized as acoustic clues for source separation. Non-negative Tensor Factorization (NTF) is an effective technique which can be used to resolve this problem by decomposing amplitude spectrograms of stereo channel music signals into basis vectors and activations of individual music source signals, along with their corresponding mixing gains. However, it is difficult to achieve sufficient separation performance using this method alone, as the acoustic clues available for separation are limited. To address this issue, this paper proposes a Cepstral Distance Regularization (CDR) method for NTF-based stereo channel separation, which involves making the cepstrum of the separated source signals follow Gaussian Mixture Models (GMMs) of the corresponding the music source signal. These GMMs are trained in advance using available samples. Experimental evaluations separating three and four sound sources are conducted to investigate the effectiveness of the proposed method in both supervised and semi-supervised separation frameworks, and performance is also compared with that of a conventional NTF method. Experimental results demonstrate that the proposed method yields significant improvements within both separation frameworks, and that cepstral distance regularization provides better separation parameters.

  • Active Contours Driven by Local Rayleigh Distribution Fitting Energy for Ultrasound Image Segmentation

    Hui BI  Yibo JIANG  Hui LI  Xuan SHA  Yi WANG  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2018/02/08
      Vol:
    E101-D No:7
      Page(s):
    1933-1937

    The ultrasound image segmentation is a crucial task in many clinical applications. However, the ultrasound image is difficult to segment due to image inhomogeneity caused by the ultrasound imaging technique. In this paper, to deal with image inhomogeneity with considering ultrasound image properties the Local Rayleigh Distribution Fitting (LRDF) energy term is introduced into the traditional level set method newly. While the curve evolution equation is derived for energy minimization, and self-driven uterus contour is achieved on the ultrasound images. The experimental segmentation results on synthetic images and in-vivo ultrasound images present that the proposed approach is effective and accurate, with the Dice Score Coefficient (DSC) of 0.95 ± 0.02.

  • Secrecy Energy Efficiency Optimization for MIMO SWIPT Systems

    Yewang QIAN  Tingting ZHANG  Haiyang ZHANG  

     
    LETTER-Communication Theory and Signals

      Vol:
    E101-A No:7
      Page(s):
    1141-1145

    In this letter, we consider a multiple-input multiple-output (MIMO) simultaneous wireless information and power transfer (SWIPT) system, in which the confidential message intended for the information receiver (IR) should be kept secret from the energy receiver (ER). Our goal is to design the optimal transmit covariance matrix so as to maximize the secrecy energy efficiency (SEE) of the system while guaranteeing the secrecy rate, energy harvesting and transmit power constraints. To deal with the original non-convex optimization problem, we propose an alternating optimization (AO)- based algorithm and also prove its convergence. Simulation results show that the proposed algorithm outperforms conventional design methods in terms of SEE.

  • An Investigative Study on How Developers Filter and Prioritize Code Smells

    Natthawute SAE-LIM  Shinpei HAYASHI  Motoshi SAEKI  

     
    PAPER

      Pubricized:
    2018/04/20
      Vol:
    E101-D No:7
      Page(s):
    1733-1742

    Code smells are indicators of design flaws or problems in the source code. Various tools and techniques have been proposed for detecting code smells. These tools generally detect a large number of code smells, so approaches have also been developed for prioritizing and filtering code smells. However, lack of empirical data detailing how developers filter and prioritize code smells hinders improvements to these approaches. In this study, we investigated ten professional developers to determine the factors they use for filtering and prioritizing code smells in an open source project under the condition that they complete a list of five tasks. In total, we obtained 69 responses for code smell filtration and 50 responses for code smell prioritization from the ten professional developers. We found that Task relevance and Smell severity were most commonly considered during code smell filtration, while Module importance and Task relevance were employed most often for code smell prioritization. These results may facilitate further research into code smell detection, prioritization, and filtration to better focus on the actual needs of developers.

  • Analysis of a Plasmonic Pole-Absorber Using a Periodic Structure Open Access

    Junji YAMAUCHI  Shintaro OHKI  Yudai NAKAGOMI  Hisamatsu NAKANO  

     
    INVITED PAPER

      Vol:
    E101-C No:7
      Page(s):
    495-500

    A plasmonic black pole (PBP) consisting of a series of touching spherical metal surfaces is analyzed using the finite-difference time-domain (FDTD) method with the periodic boundary condition. First, the wavelength characteristics of the PBP are studied under the assumption that the PBP is omnidirectionally illuminated. It is found that partial truncation of each metal sphere reduces the reflectivity over a wide wavelength range. Next, we consider the case where the PBP is illuminated with a cylindrical wave from a specific direction. It is shown that an absorptivity of more than 80% is obtained over a wavelength range of λ=500 nm to 1000 nm. Calculation regarding the Poynting vector distribution also shows that the incident wave is bent and absorbed towards the center axis of the PBP.

  • Stochastic Number Duplicators Based on Bit Re-Arrangement Using Randomized Bit Streams

    Ryota ISHIKAWA  Masashi TAWADA  Masao YANAGISAWA  Nozomu TOGAWA  

     
    PAPER

      Vol:
    E101-A No:7
      Page(s):
    1002-1013

    Recently, stochastic computing based on stochastic numbers attracts attention as an effective computation method, which realizes arithmetic operations by simple logic circuits with a tolerance of bit errors. When we input two or more identical values to a stochastic circuit, we require to duplicate a stochastic number. However, if bit streams of duplicated stochastic numbers are dependent on each other, their arithmetic operation results can be inaccurate. In this paper, we propose two stochastic number duplicators, called FSR and RRR. The stochastic numbers duplicated by the FSR and RRR duplicators have the equivalent values but have independent bit streams, effectively utilizing bit re-arrangement using randomized bit streams. Experimental evaluation results demonstrate that the RRR duplicator, in particular, obtains more accurate results even if a circuit has re-convergence paths, reducing the mean square errors by 20%-89% compared to a conventional stochastic number duplicator.

  • Using Scattered X-Rays to Improve the Estimation Accuracy of Attenuation Coefficients: A Fundamental Analysis

    Naohiro TODA  Tetsuya NAKAGAMI  Yoichi YAMAZAKI  Hiroki YOSHIOKA  Shuji KOYAMA  

     
    PAPER-Measurement Technology

      Vol:
    E101-A No:7
      Page(s):
    1101-1114

    In X-ray computed tomography, scattered X-rays are generally removed by using a post-patient collimator located in front of the detector. In this paper, we show that the scattered X-rays have the potential to improve the estimation accuracy of the attenuation coefficient in computed tomography. In order to clarify the problem, we simplified the geometry of the computed tomography into a thin cylinder composed of a homogeneous material so that only one attenuation coefficient needs to be estimated. We then conducted a Monte Carlo numerical experiment on improving the estimation accuracy of attenuation coefficient by measuring the scattered X-rays with several dedicated toroidal detectors around the cylinder in addition to the primary X-rays. We further present a theoretical analysis to explain the experimental results. We employed a model that uses a T-junction (i.e., T-junction model) to divide the photon transport into primary and scattered components. This division is processed with respect to the attenuation coefficient. Using several T-junction models connected in series, we modeled the case of several scatter detectors. The estimation accuracy was evaluated according to the variance of the efficient estimator, i.e., the Cramer-Rao lower bound. We confirmed that the variance decreases as the number of scatter detectors increases, which implies that using scattered X-rays can reduce the irradiation dose for patients.

  • Cross-Layer Design for Exposed Node Reduction in Ad Hoc WLANs

    Emilia WEYULU  Masaki HANADA  Hidehiro KANEMITSU  Eun-Chan PARK  Moo Wan KIM  

     
    PAPER

      Pubricized:
    2018/01/22
      Vol:
    E101-B No:7
      Page(s):
    1575-1588

    Interference in ad hoc WLANs is a common occurrence as there is no centralized access point controlling device access to the wireless channel. IEEE 802.11 WLANs use carrier sense multiple access with collision avoidance (CSMA/CA) which initiates the Request to Send/Clear to Send (RTS/CTS) handshaking mechanism to solve the hidden node problem. While it solves the hidden node problem, RTS/CTS triggers the exposed node problem. In this paper, we present an evaluation of a method for reducing exposed nodes in 802.11 ad hoc WLANs. Using asymmetric transmission ranges for RTS and CTS frames, a cross-layer design is implemented between Layer 2 and 3 of the OSI model. Information obtained by the AODV routing protocol is utilized in adjusting the RTS transmission range at the MAC Layer. The proposed method is evaluated with the NS-2 simulator and we observe significant throughput improvement, and confirm the effectiveness of the proposed method. Especially when the mobile nodes are randomly distributed, the throughput gain of the Asymmetric RTS/CTS method is up to 30% over the Standard RTS/CTS method.

  • Implementing Adaptive Decisions in Stochastic Simulations via AOP

    Pilsung KANG  

     
    LETTER-Software Engineering

      Pubricized:
    2018/04/05
      Vol:
    E101-D No:7
      Page(s):
    1950-1953

    We present a modular way of implementing adaptive decisions in performing scientific simulations. The proposed method employs modern software engineering mechanisms to allow for better software management in scientific computing, where software adaptation has often been implemented manually by the programmer or by using in-house tools, which complicates software management over time. By applying the aspect-oriented programming (AOP) paradigm, we consider software adaptation as a separate concern and, using popular AOP constructs, implement adaptive decision separately from the original code base, thereby improving software management. We demonstrate the effectiveness of our approach with applications to stochastic simulation software.

  • Error Correction for Search Engine by Mining Bad Case

    Jianyong DUAN  Tianxiao JI  Hao WANG  

     
    PAPER-Natural Language Processing

      Pubricized:
    2018/03/26
      Vol:
    E101-D No:7
      Page(s):
    1938-1945

    Automatic error correction of users' search terms for search engines is an important aspect of improving search engine retrieval efficiency, accuracy and user experience. In the era of big data, we can analyze and mine massive search engine logs to release the hidden mind with big data ideas. It can obtain better results through statistical modeling of query errors in search engine log data. But when we cannot find the error query in the log, we can't make good use of the information in the log to correct the query result. These undiscovered error queries are called Bad Case. This paper combines the error correction algorithm model and search engine query log mining analysis. First, we explored Bad Cases in the query error correction process through the search engine query logs. Then we quantified the characteristics of these Bad Cases and built a model to allow search engines to automatically mine Bad Cases with these features. Finally, we applied Bad Cases to the N-gram error correction algorithm model to check the impact of Bad Case mining on error correction. The experimental results show that the error correction based on Bad Case mining makes the precision rate and recall rate of the automatic error correction improved obviously. Users experience is improved and the interaction becomes more friendly.

  • Effect of User Antenna Selection on Block Beamforming Algorithms for Suppressing Inter-User Interference in Multiuser MIMO System Open Access

    Nobuyoshi KIKUMA  Kentaro NISHIMORI  Takefumi HIRAGURI  

     
    INVITED PAPER

      Pubricized:
    2018/01/22
      Vol:
    E101-B No:7
      Page(s):
    1523-1535

    Multiuser MIMO (MU-MIMO) improves the system channel capacity by generating a large virtual MIMO channel between a base station and multiple user terminals (UTs) with effective utilization of wireless resources. Block beamforming algorithms such as Block Diagonalization (BD) and Block Maximum Signal-to-Noise ratio (BMSN) have been proposed in order to realize MU-MIMO broadcast transmission. The BD algorithm cancels inter-user interference (IUI) by creating the weights so that the channel matrices for the other users are set to be zero matrices. The BMSN algorithm has a function of maintaining a high gain response for each desired user in addition to IUI cancellation. Therefore, the BMSN algorithm generally outperforms the BD algorithm. However, when the number of transmit antennas is equal to the total number of receive antennas, the transmission rate by both BD and BMSN algorithms is decreased. This is because the eigenvalues of channel matrices are too small to support data transmission. To resolve the issue, this paper focuses on an antenna selection (AS) method at the UTs. The AS method reduces the number of pattern nulls for the other users except an intended user in the BD and BMSN algorithms. It is verified via bit error rate (BER) evaluation that the AS method is effective in the BD and BMSN algorithms, especially, when the number of user antennas with a low bit rate (i.e., low signal-to-noise power ratio) is increased. Moreover, this paper evaluates the achievable bit rate and throughput including an actual channel state information feedback based on IEEE802.11ac standard. Although the number of equivalent receive antenna is reduced to only one by the AS method when the number of antennas at the UT is two, it is shown that the throughputs by BD and BMSN with the AS method (BD-AS and BMSN-AS) are higher than those by the conventional BD and BMSN algorithms.

  • Towards an Improvement of Bug Report Summarization Using Two-Layer Semantic Information

    Cheng-Zen YANG  Cheng-Min AO  Yu-Han CHUNG  

     
    PAPER

      Pubricized:
    2018/04/20
      Vol:
    E101-D No:7
      Page(s):
    1743-1750

    Bug report summarization has been explored in past research to help developers comprehend important information for bug resolution process. As text mining technology advances, many summarization approaches have been proposed to provide substantial summaries on bug reports. In this paper, we propose an enhanced summarization approach called TSM by first extending a semantic model used in AUSUM with the anthropogenic and procedural information in bug reports and then integrating the extended semantic model with the shallow textual information used in BRC. We have conducted experiments with a dataset of realistic software projects. Compared with the baseline approaches BRC and AUSUM, TSM demonstrates the enhanced performance in achieving relative improvements of 34.3% and 7.4% in the F1 measure, respectively. The experimental results show that TSM can effectively improve the performance.

  • Growth Mechanism of Polar-Plane-Free Faceted InGaN Quantum Wells Open Access

    Yoshinobu MATSUDA  Mitsuru FUNATO  Yoichi KAWAKAMI  

     
    INVITED PAPER

      Vol:
    E101-C No:7
      Page(s):
    532-536

    The growth mechanisms of three-dimensionally (3D) faceted InGaN quantum wells (QWs) on (=1=12=2) GaN substrates are discussed. The structure is composed of (=1=12=2), {=110=1}, and {=1100} planes, and the cross sectional shape is similar to that of 3D QWs on (0001). However, the 3D QWs on (=1=12=2) and (0001) show quite different inter-facet variation of In compositions. To clarify this observation, the local thicknesses of constituent InN and GaN on the 3D GaN are fitted with a formula derived from the diffusion equation. It is suggested that the difference in the In incorporation efficiency of each crystallographic plane strongly affects the surface In adatom migration.

4981-5000hit(42807hit)