The search functionality is under construction.

Keyword Search Result

[Keyword] performance evaluation(105hit)

1-20hit(105hit)

  • Theoretical Analysis of Fully Wireless-Power-Transfer Node Networks Open Access

    Hiroshi SAITO  

     
    PAPER-Fundamental Theories for Communications

      Pubricized:
    2023/05/10
      Vol:
    E106-B No:10
      Page(s):
    864-872

    The performance of a fully wireless-power-transfer (WPT) node network, in which each node transfers (and receives) energy through a wireless channel when it has sufficient (and insufficient) energy in its battery, was theoretically analyzed. The lost job ratio (LJR), namely, is the ratio of (i) the amount of jobs that cannot be done due to battery of a node running out to (ii) the amount of jobs that should be done, is used as a performance metric. It describes the effect of the battery of each node running out and how much additional energy is needed. Although it is known that WPT can reduce the probability of the battery running out among a few nodes within a small area, the performance of a fully WPT network has not been clarified. By using stochastic geometry and first-passage-time analysis for a diffusion process, the expected LJR was theoretically derived. Numerical examples demonstrate that the key parameters determining the performance of the network are node density, threshold switching of statuses between “transferring energy” and “receiving energy,” and the parameters of power conversion. They also demonstrate the followings: (1) The mean energy stored in the node battery decreases in the networks because of the loss caused by WPT, and a fully WPT network cannot decrease the probability of the battery running out under the current WPT efficiency. (2) When the saturation value of power conversion increases, a fully WPT network can decrease the probability of the battery running out although the mean energy stored in the node battery still decreases in the networks. This result is explained by the fact that the variance of stored energy in each node battery becomes smaller due to transfer of energy from nodes of sufficient energy to nodes of insufficient energy.

  • Workload-Driven Analysis on the Performance Characteristics of GPU-Accelerated DBMSes

    Junyoung AN  Young-Kyoon SUH  Byungchul TAK  

     
    LETTER-Data Engineering, Web Information Systems

      Vol:
    E105-D No:11
      Page(s):
    1984-1989

    This letter conducts an in-depth empirical analysis of the influence of various query characteristics on the performance of modern GPU DBMSes. Our analysis reveals that, although they can efficiently process concurrent queries, the GPU DBMSes we consider still should address various performance concerns including n-way joins, aggregates, and selective scans.

  • How to Extend CTRT for AES-256 and AES-192

    SeongHan SHIN  Shota YAMADA  Goichiro HANAOKA  Yusuke ISHIDA  Atsushi KUNII  Junichi OKETANI  Shimpei KUNII  Kiyoshi TOMOMURA  

     
    PAPER-Cryptography and Information Security

      Pubricized:
    2022/02/16
      Vol:
    E105-A No:8
      Page(s):
    1121-1133

    AONT (All-or-Nothing Transform) is a kind of (n, n)-threshold secret sharing scheme that distributes a message m into a set of n shares such that the message m can be reconstructed if and only if n shares are collected. At CRYPTO 2000, Desai proposed a simple and faster AONT based on the CTR mode of encryption (called CTRT) and proved its security in the ideal cipher model. Though AES-128, whose key length k = 128 and block length l = 128, can be used in CTRT as a block cipher, AES-256 and AES-192 cannot be used due to its intrinsic restriction of k ≤ l. In this paper, we propose an extended CTRT (for short, XCTRT) suitable for AES-256. By thoroughly evaluating all the tricky cases, we prove that XCTRT is secure in the ideal cipher model under the same CTRT security definition. Also, we discuss the security result of XCTRT in concrete parameter settings. For more flexibility of key length, we propose a variant of XCTRT dealing with l

  • An Efficient Aircraft Boarding Strategy Considering Implementation

    Kenji UEHARA  Kunihiko HIRAISHI  Kokolo IKEDA  

     
    PAPER-Mathematical Systems Science

      Pubricized:
    2021/01/22
      Vol:
    E104-A No:8
      Page(s):
    1051-1058

    Boarding is the last step of aircraft turnaround and its completion in the shortest possible time is desired. In this paper, we propose a new boarding strategy that outperforms conventional strategies such as the back-to-front strategy and the outside-in strategy. The Steffen method is known as one of the most efficient boarding strategies in literature, but it is hard to be realized in the real situation because the complete sorting of passengers in a prescribed order is required. The proposed strategy shows a performance close to that of the Steffen method and can be easily implemented by using a special gate system.

  • Node-Disjoint Paths Problems in Directed Bijective Connection Graphs

    Keiichi KANEKO  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2019/09/26
      Vol:
    E103-D No:1
      Page(s):
    93-100

    In this paper, we extend the notion of bijective connection graphs to introduce directed bijective connection graphs. We propose algorithms that solve the node-to-set node-disjoint paths problem and the node-to-node node-disjoint paths problem in a directed bijective connection graph. The time complexities of the algorithms are both O(n4), and the maximum path lengths are both 2n-1.

  • An Efficient Parallel Coding Scheme in Erasure-Coded Storage Systems

    Wenrui DONG  Guangming LIU  

     
    PAPER-Computer System

      Pubricized:
    2017/12/12
      Vol:
    E101-D No:3
      Page(s):
    627-643

    Erasure codes have been considered as one of the most promising techniques for data reliability enhancement and storage efficiency in modern distributed storage systems. However, erasure codes often suffer from a time-consuming coding process which makes them nearly impractical. The opportunity to solve this problem probably rely on the parallelization of erasure-code-based application on the modern multi-/many-core processors to fully take advantage of the adequate hardware resources on those platforms. However, the complicated data allocation and limited I/O throughput pose a great challenge on the parallelization. To address this challenge, we propose a general multi-threaded parallel coding approach in this work. The approach consists of a general multi-threaded parallel coding model named as MTPerasure, and two detailed parallel coding algorithms, named as sdaParallel and ddaParallel, respectively, adapting to different I/O circumstances. MTPerasure is a general parallel coding model focusing on the high level data allocation, and it is applicable for all erasure codes and can be implemented without any modifications of the low level coding algorithms. The sdaParallel divides the data into several parts and the data parts are allocated to different threads statically in order to eliminate synchronization latency among multiple threads, which improves the parallel coding performance under the dummy I/O mode. The ddaParallel employs two threads to execute the I/O reading and writing on the basis of small pieces independently, which increases the I/O throughput. Furthermore, the data pieces are assigned to the coding thread dynamically. A special thread scheduling algorithm is also proposed to reduce thread migration latency. To evaluate our proposal, we parallelize the popular open source library jerasure based on our approach. And a detailed performance comparison with the original sequential coding program indicates that the proposed parallel approach outperforms the original sequential program by an extraordinary speedups from 1.4x up to 7x, and achieves better utilization of the computation and I/O resources.

  • A New Algorithm to Determine Covariance in Statistical Maximum for Gaussian Mixture Model

    Daiki AZUMA  Shuji TSUKIYAMA  

     
    PAPER

      Vol:
    E100-A No:12
      Page(s):
    2834-2841

    In statistical approaches such as statistical static timing analysis, the distribution of the maximum of plural distributions is computed by repeating a maximum operation of two distributions. Moreover, since each distribution is represented by a linear combination of several explanatory random variables so as to handle correlations efficiently, sensitivity of the maximum of two distributions to each explanatory random variable, that is, covariance between the maximum and an explanatory random variable, must be calculated in every maximum operation. Since distribution of the maximum of two Gaussian distributions is not a Gaussian, Gaussian mixture model is used for representing a distribution. However, if Gaussian mixture models are used, then it is not always possible to make both variance and covariance of the maximum correct simultaneously. We propose a new algorithm to determine covariance without deteriorating the accuracy of variance of the maximum, and show experimental results to evaluate its performance.

  • Node-to-Node Disjoint Paths Problem in Möbius Cubes

    David KOCIK  Keiichi KANEKO  

     
    PAPER-Dependable Computing

      Pubricized:
    2017/04/25
      Vol:
    E100-D No:8
      Page(s):
    1837-1843

    The Möbius cube is a variant of the hypercube. Its advantage is that it can connect the same number of nodes as a hypercube but with almost half the diameter of the hypercube. We propose an algorithm to solve the node-to-node disjoint paths problem in n-Möbius cubes in polynomial-order time of n. We provide a proof of correctness of the algorithm and estimate that the time complexity is O(n2) and the maximum path length is 3n-5.

  • A New Model of Effective Carrier-to-Noise Ratio for Global Navigation Satellite System Receiver in the Presence of Continuous Wave Interference

    Jian LI  Junwei NIE  Huaming CHEN  Guangfu SUN  Feixue WANG  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2016/12/13
      Vol:
    E100-B No:6
      Page(s):
    1003-1009

    In this paper, a new effective C/N0 theoretical model for global navigation satellite system (GNSS) receiver is proposed, in the presence of continuous wave interference (CWI). The proposed model is derived based on an interesting finding, correlator may output direct current (DC) in the presence of CWI. The DC introduced by CWI eventually leads to increase of carrier power estimation. It is totally different from current assumption that interference just causes noise power increase after correlation. The proposed model is verified by simulation.

  • Simulation Study of Low Latency Network Architecture Using Mobile Edge Computing

    Krittin INTHARAWIJITR  Katsuyoshi IIDA  Hiroyuki KOGA  

     
    PAPER

      Pubricized:
    2017/02/08
      Vol:
    E100-D No:5
      Page(s):
    963-972

    Attaining extremely low latency service in 5G cellular networks is an important challenge in the communication research field. A higher QoS in the next-generation network could enable several unprecedented services, such as Tactile Internet, Augmented Reality, and Virtual Reality. However, these services will all need support from powerful computational resources provided through cloud computing. Unfortunately, the geolocation of cloud data centers could be insufficient to satisfy the latency aimed for in 5G networks. The physical distance between servers and users will sometimes be too great to enable quick reaction within the service time boundary. The problem of long latency resulting from long communication distances can be solved by Mobile Edge Computing (MEC), though, which places many servers along the edges of networks. MEC can provide shorter communication latency, but total latency consists of both the transmission and the processing times. Always selecting the closest edge server will lead to a longer computing latency in many cases, especially when there is a mass of users around particular edge servers. Therefore, the research studies the effects of both latencies. The communication latency is represented by hop count, and the computation latency is modeled by processor sharing (PS). An optimization model and selection policies are also proposed. Quantitative evaluations using simulations show that selecting a server according to the lowest total latency leads to the best performance, and permitting an over-latency barrier would further improve results.

  • Reversible Audio Data Hiding Based on Variable Error-Expansion of Linear Prediction for Segmental Audio and G.711 Speech

    Akira NISHIMURA  

     
    PAPER

      Pubricized:
    2015/10/21
      Vol:
    E99-D No:1
      Page(s):
    83-91

    Reversible data hiding is a technique in which hidden data are embedded in host data such that the consistency of the host is perfectly preserved and its data are restored during extraction of the hidden data. In this paper, a linear prediction technique for reversible data hiding of audio waveforms is improved. The proposed variable expansion method is able to control the payload size through varying the expansion factor. The proposed technique is combined with the prediction error expansion method. Reversible embedding, perfect payload detection, and perfect recovery of the host signal are achieved for a framed audio signal. A smaller expansion factor results in a smaller payload size and less degradation in the stego audio quality. Computer simulations reveal that embedding a random-bit payload of less than 0.4 bits per sample into CD-format music signals provide stego audio with acceptable objective quality. The method is also applied to G.711 µ-law-coded speech signals. Computer simulations reveal that embedding a random-bit payload of less than 0.1 bits per sample into speech signals provide stego speech with good objective quality.

  • Long-Term Performance Evaluation of Hadoop Jobs in Public and Community Clouds

    Kento AIDA  Omar ABDUL-RAHMAN  Eisaku SAKANE  Kazutaka MOTOYAMA  

     
    PAPER-Computer System

      Pubricized:
    2015/02/25
      Vol:
    E98-D No:6
      Page(s):
    1176-1184

    Cloud computing is a widely used computing platform in business and academic communities. Performance is an important issue when a user runs an application in the cloud. The user may want to estimate the application-execution time beforehand to guarantee the application performance or to choose the most suitable cloud. Moreover, the cloud system architect and the designer need to understand the application performance characteristics, such as the scalability or the utilization of cloud platforms, to improve performance. However, because the application performance in clouds sometime fluctuates, estimation of the application performance is difficult. In this paper, we discuss the performance fluctuation of Hadoop jobs in both a public cloud and a community cloud for one to three months. The experimental results indicate phenomena that we cannot see without long-term experiments and phenomena inherent in Hadoop. The results suggest better ways to estimate Hadoop application performances in clouds. For example, we should be aware of application characteristics (CPU intensive or communication intensive), datacenter characteristics (busy or not), and time frame (time of day and day of the week) to estimate the performance fluctuation due to workload congestion in cloud platforms. Furthermore, we should be aware of performance degradation due to task re-execution in Hadoop applications.

  • Research and Modeling on Performance Evaluation of IEEE 802.15.6

    Yali WANG  Lan CHEN  Chao LV  

     
    PAPER-Network

      Vol:
    E97-B No:11
      Page(s):
    2378-2385

    IEEE 802.15.6 provides PHY and MAC layer profiles for wearable and implanted Wireless Body Area Networks (WBANs). The critical requirements of QoS guarantee and ultra-low-power are severe challenges when implementing IEEE 802.15.6. In this paper, the key problem in IEEE 802.15.6 standard that “How to allocate EAP (Exclusive Access Phase)?” is solved for the first time: An analysis of network performance indicates that too much EAP allocation can not promote traffic performance obviously and effectually. However, since EAP allocation plays an important role in guaranteeing quality of service, a customized and quantitative EAP allocation solution is proposed. Simulation results show that the solution can obtain the optimal network performance. Furthermore, the estimated models of delay and energy are developed, which help to design the WBAN according to application requirements and analyze the network performance according to the traffic characteristics. The models are simple, effective, and relatively accurate. Results show that the models have approximated mean and the correlation coefficient is greater than 0.95 compared with the simulations of IEEE 802.15.6 using NS2 platform. The work of this paper can solve crucial practical problems in using IEEE 802.15.6, and will propel WBANs applications widely.

  • Implementation and Performance Evaluation of a Distributed TV White Space Sensing System

    Ha-Nguyen TRAN  Yohannes D. ALEMSEGED  Hiroshi HARADA  

     
    PAPER

      Vol:
    E97-B No:2
      Page(s):
    305-313

    Spectrum sensing is one of the methods to identify available white spaces for secondary usage which was specified by the regulators. However, signal quality to be sensed can plunge to a very low signal-to-noise-ratio due to signal propagation and hence readings from individual sensors will be unreliable. Distributed sensing by the cooperation of multiple sensors is one way to cope with this problem because the diversity gain due to the combining effect of data captured at different position will assist in detecting signals that might otherwise not be detected by a single sensor. In effect, the probability of detection can be improved. We have implemented a distributed sensing system to evaluate the performance of different cooperative sensing algorithms. In this paper we describe our implementation and measurement experience which include the system design, specification of the system, measurement method, the issues and solutions. This paper also confirms the performance enhancement offered by distributed sensing algorithms, and describes several ideas for further enhancement of the sensing quality.

  • Two Heuristic Algorithms for the Minimum Initial Marking Problem of Timed Petri Nets

    Satoru OCHIIWA  Satoshi TAOKA  Masahiro YAMAUCHI  Toshimasa WATANABE  

     
    PAPER-Algorithms and Data Structures

      Vol:
    E96-A No:2
      Page(s):
    540-553

    A timed Petri net, an extended model of an ordinary Petri net with introduction of discrete time delay in firing activity, is practically useful in performance evaluation of real-time systems and so on. Unfortunately though, it is often too difficult to solve (efficiently) even most basic problems in timed Petri net theory. This motivates us to do research on analyzing complexity of Petri net problems and on designing efficient and/or heuristic algorithms. The minimum initial marking problem of timed Petri nets (TPMIM) is defined as follows: “Given a timed Petri net, a firing count vector X and a nonnegative integer π, find a minimum initial marking (an initial marking with the minimum total token number) among those initial ones M each of which satisfies that there is a firing scheduling which is legal on M with respect to X and whose completion time is no more than π, and, if any, find such a firing scheduling.” In a production system like factory automation, economical distribution of initial resources, from which a schedule of job-processings is executable, can be formulated as TPMIM. The subject of the paper is to propose two pseudo-polynomial time algorithms TPM and TMDLO for TPMIM, and to evaluate them by means of computer experiment. Each of the two algorithms finds an initial marking and a firing sequence by means of algorithms for MIM (the initial marking problem for non-timed Petri nets), and then converts it to a firing scheduling of a given timed Petri net. It is shown through our computer experiments that TPM has highest capability among our implemented algorithms including TPM and TMDLO.

  • ER-TCP (Exponential Recovery-TCP): High-Performance TCP for Satellite Networks

    Mankyu PARK  Minsu SHIN  Deockgil OH  Doseob AHN  Byungchul KIM  Jaeyong LEE  

     
    PAPER-Network

      Vol:
    E95-B No:5
      Page(s):
    1679-1688

    A transmission control protocol (TCP) using an additive increase multiplicative decrease (AIMD) algorithm for congestion control plays a leading role in advanced Internet services. However, the AIMD method shows only low link utilization in lossy networks with long delay such as satellite networks. This is because the cwnd dynamics of TCP are reduced by long propagation delay, and TCP uses an inadequate congestion control algorithm, which does not distinguish packet loss from wireless errors from that due to congestion of the wireless networks. To overcome these problems, we propose an exponential recovery (ER) TCP that uses the exponential recovery function for rapidly occupying available bandwidth during a congestion avoidance period, and an adaptive congestion window decrease scheme using timestamp base available bandwidth estimation (TABE) to cope with wireless channel errors. We simulate the proposed ER-TCP under various test scenarios using the ns-2 network simulator to verify its performance enhancement. Simulation results show that the proposal is a more suitable TCP than the several TCP variants under long delay and heavy loss probability environments of satellite networks.

  • Performance Evaluation of an Option-Based Learning Algorithm in Multi-Car Elevator Systems

    Alex VALDIVIELSO CHIAN  Toshiyuki MIYAMOTO  

     
    LETTER-Concurrent Systems

      Vol:
    E95-A No:4
      Page(s):
    835-839

    In this letter, we present the evaluation of an option-based learning algorithm, developed to perform a conflict-free allocation of calls among cars in a multi-car elevator system. We evaluate its performance in terms of the service time, its flexibility in the task-allocation, and the load balancing.

  • On Improving the Reliability and Performance of the YAFFS Flash File System

    Seungjae BAEK  Heekwon PARK  Jongmoo CHOI  

     
    LETTER-Software System

      Vol:
    E94-D No:12
      Page(s):
    2528-2532

    In this paper, we propose three techniques to improve the performance of YAFFS (Yet Another Flash File System), while enhancing the reliability of the system. Specifically, we first propose to manage metadata and user data separately on segregated blocks. This modification not only leads to the reduction of the mount time but also reduces the garbage collection time. Second, we tailor the wear-leveling to the segregated metadata and user data blocks. That is, worn out blocks between the segregated blocks are swapped, which leads to more evenly worn out blocks increasing the lifetime of the system. Finally, we devise an analytic model to predict the expected garbage collection time. By accurately predicting the garbage collection time, the system can perform garbage collection at more opportune times when the user's perceived performance may not be negatively affected. Performance evaluation results based on real implementations show that our modifications enhance performance and reliability without incurring additional overheads. Specifically, the YAFFS with our proposed techniques outperforms the original YAFFS by six times in terms of mount speed and five times in terms of benchmark performance, while reducing the average erase count of blocks by 14%.

  • Empirical Performance Evaluation of Raster-to-Vector Conversion Methods: A Study on Multi-Level Interactions between Different Factors

    Hasan S.M. AL-KHAFFAF  Abdullah Z. TALIB  Rosalina ABDUL SALAM  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E94-D No:6
      Page(s):
    1278-1288

    Many factors, such as noise level in the original image and the noise-removal methods that clean the image prior to performing a vectorization, may play an important role in affecting the line detection of raster-to-vector conversion methods. In this paper, we propose an empirical performance evaluation methodology that is coupled with a robust statistical analysis method to study many factors that may affect the quality of line detection. Three factors are studied: noise level, noise-removal method, and the raster-to-vector conversion method. Eleven mechanical engineering drawings, three salt-and-pepper noise levels, six noise-removal methods, and three commercial vectorization methods were used in the experiment. The Vector Recovery Index (VRI) of the detected vectors was the criterion used for the quality of line detection. A repeated measure ANOVA analyzed the VRI scores. The statistical analysis shows that all the studied factors affected the quality of line detection. It also shows that two-way interactions between the studied factors affected line detection.

  • Performance Evaluation of Routing Schemes for the Energy-Constrained DTN with Selfish Nodes

    Yong LI  Depeng JIN  Li SU  Lieguang ZENG  

     
    LETTER-Network

      Vol:
    E94-B No:5
      Page(s):
    1442-1446

    Due to the lack of end-to-end paths between the communication source and destination, the routing of Delay Tolerant Networks (DTN) exploits the store-carry-and-forward mechanism. This mechanism requires nodes with sufficient energy to relay and forward messages in a cooperative and selfless way. However, in the real world, the energy is constrained and most of the nodes exhibit selfish behaviors. In this paper, we investigate the performance of DTN routing schemes considering both the energy constraint and selfish behaviors of nodes. First, we model the two-hop relay and epidemic routing based on a two-dimensional continuous time Markov chain. Then, we obtain the system performance of message delivery delay and delivery cost by explicit expressions. Numerical results show that both the energy constraint and node selfishness reduce the message delivery cost at the expense of increasing the message delivery delay. Furthermore, we demonstrate that the energy constraint plays a more important role in the performance of epidemic routing than that of two-hop relay.

1-20hit(105hit)