The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] engine(173hit)

61-80hit(173hit)

  • Channel Allocation and DBA Mechanisms for Improvement of Bandwidth Utilization Performance of ECDM-PON Systems

    Yoshihiro NAKAHIRA  Ryuichi WATANABE  Masayuki KASHIMA  

     
    PAPER-Fiber-Optic Transmission for Communications

      Vol:
    E94-B No:2
      Page(s):
    437-443

    This paper describes a novel channel allocation and DBA (Dynamic Bandwidth Allocation) mechanism for ECDM-PON (Electric Code Division Multiplex -- Passive Optical Network) systems. In the current ECDM-PON systems, each ONU (Optical Network Unit) is limited to 2 or 3 CDM channels. This is because (fixed channel) CDM transmitters are expensive, and tunable CDM transmitters even more expensive. With a small number of CDM channels, the bandwidth utilization ratio is restricted by channel blocking. Our proposed mechanisms can reduce the channel blocking ratio without increasing the number of CDM transmitters or using tunable CDM transmitters. To clarify the advantages of the proposed system performance, we have evaluated the channel non-blocking ratio (Rn) and wasted resource ratio (Rw) when some users request bandwidth more than 100%. Evaluation of the non-blocking ratio, Rn shows that the proposed mechanisms approach the performance of a system with tunable CDM transmitters when the number of ONUs with over 100% traffic load is small. We have also simulated throughput for uniform traffic. In addition to these evaluations, we implemented our proposed mechanism on an FPGA (Field Programming Gate Array) and evaluated the calculation speed to allocate timeslots on CDM channels and a timeline.

  • Component Identification and Evaluation for Legacy Systems--An Empirical Study--

    JianFeng CUI  HeungSeok CHAE  

     
    PAPER-Software Engineering

      Vol:
    E93-D No:12
      Page(s):
    3306-3320

    In the field of software reengineering, many component identification approaches have been proposed for evolving legacy systems into component-based systems. Understanding the behaviors of various component identification approaches is the first important step to meaningfully employ them for legacy systems evolution, therefore we performed an empirical study on component identification technology with considerations of their similarity measures, clustering approaches and stopping criteria. We proposed a set of evaluation criteria and developed the tool CIETool to automate the process of component identification and evaluation. The experimental results revealed that many components of poor quality were produced by the employed component identification approaches; that is, many of the identified components were tightly coupled, weakly cohesive, or had inappropriate numbers of implementation classes and interface operations. Finally, we presented an analysis on the component identification approaches according to the proposed evaluation criteria, which suggested that the weaknesses of these clustering approaches were the major reasons that caused components of poor-quality.

  • Architecture and Circuit Optimization of Hardwired Integer Motion Estimation Engine for H.264/AVC

    Zhenyu LIU  Dongsheng WANG  Takeshi IKENAGA  

     
    PAPER-Image Processing

      Vol:
    E93-A No:11
      Page(s):
    2065-2073

    Variable block size motion estimation developed by the latest video coding standard H.264/AVC is the efficient approach to reduce the temporal redundancies. The intensive computational complexity coming from the variable block size technique makes the hardwired accelerator essential, for real-time applications. Propagate partial sums of absolute differences (Propagate Partial SAD) and SAD Tree hardwired engines outperform other counterparts, especially considering the impact of supporting variable block size technique. In this paper, the authors apply the architecture-level and the circuit-level approaches to improve the maximum operating frequency and reduce the hardware overhead of Propagate Partial SAD and SAD Tree, while other metrics, in terms of latency, memory bandwidth and hardware utilization, of the original architectures are maintained. Experiments demonstrate that by using the proposed approaches, at 110.8 MHz operating frequency, compared with the original architectures, 14.7% and 18.0% gate count can be saved for Propagate Partial SAD and SAD Tree, respectively. With TSMC 0.18 µm 1P6M CMOS technology, the proposed Propagate Partial SAD architecture achieves 231.6 MHz operating frequency at a cost of 84.1 k gates. Correspondingly, the maximum work frequency of the optimized SAD Tree architecture is improved to 204.8 MHz, which is almost two times of the original one, while its hardware overhead is merely 88.5 k-gate.

  • Autonomous Traffic Engineering for Boosting Application Fidelity in Wireless Sensor Networks

    Md. Abdur RAZZAQUE  Choong Seon HONG  Sungwon LEE  

     
    PAPER-Network

      Vol:
    E93-B No:11
      Page(s):
    2990-3003

    This paper presents an autonomous traffic engineering framework, named ATE, a highly efficient data dissemination mechanism for multipath data forwarding in Wireless Sensor Networks (WSNs). The proposed ATE has several salient features. First, ATE utilizes three coordinating schemes: an incipient congestion inference scheme, an accurate link quality estimation scheme and a dynamic traffic diversion scheme. It significantly minimizes packet drops due to congestion by dynamically and adaptively controlling the data traffic over congested nodes and/or poorer quality links, and by opportunistically exploiting under-utilized nodes for traffic diversion, while minimizing the estimation and measurement overhead. Second, ATE can provide with high application fidelity of the network even for increasing values of bit error rates and node failures. The proposed link quality estimation and congestion inference schemes are light weight and distributed, improving the energy efficiency of the network. Autonomous Traffic Engineering has been evaluated extensively via NS-2 simulations, and the results have shown that ATE provides a better performance with minimum overhead than those of existing approaches.

  • A Case Study of Requirements Elicitation Process with Changes

    Takako NAKATANI  Shouzo HORI  Naoyasu UBAYASHI  Keiichi KATAMINE  Masaaki HASHIMOTO  

     
    PAPER-Software Engineering

      Vol:
    E93-D No:8
      Page(s):
    2182-2189

    Requirements changes sometimes cause a project to fail. A lot of projects now follow incremental development processes so that new requirements and requirements changes can be incorporated as soon as possible. These processes are called integrated requirements processes, which function to integrate requirements processes with other developmental processes. We have quantitatively and qualitatively investigated the requirements processes of a specific project from beginning to end. Our focus is to clarify the types of necessary requirements based on the components contained within a certain portion of the software architecture. Further, each type reveals its typical requirements processes through its own rationale. This case study is a system to manage the orders and services of a restaurant. In this paper, we introduce the case and categorize its requirements processes based on the components of the system and the qualitative characteristics of ISO-9126. We could identify seven categories of the typical requirements process to be managed and/or controlled. Each category reveals its typical requirements processes and their characteristics. The case study is our first step of practical integrated requirements engineering.

  • High-Speed Computation of the Kleene Star in Max-Plus Algebraic System Using a Cell Broadband Engine

    Hiroyuki GOTO  

     
    PAPER-Fundamentals of Information Systems

      Vol:
    E93-D No:7
      Page(s):
    1798-1806

    This research addresses a high-speed computation method for the Kleene star of the weighted adjacency matrix in a max-plus algebraic system. We focus on systems whose precedence constraints are represented by a directed acyclic graph and implement it on a Cell Broadband EngineTM (CBE) processor. Since the resulting matrix gives the longest travel times between two adjacent nodes, it is often utilized in scheduling problem solvers for a class of discrete event systems. This research, in particular, attempts to achieve a speedup by using two approaches: parallelization and SIMDization (Single Instruction, Multiple Data), both of which can be accomplished by a CBE processor. The former refers to a parallel computation using multiple cores, while the latter is a method whereby multiple elements are computed by a single instruction. Using the implementation on a Sony PlayStation 3TM equipped with a CBE processor, we found that the SIMDization is effective regardless of the system's size and the number of processor cores used. We also found that the scalability of using multiple cores is remarkable especially for systems with a large number of nodes. In a numerical experiment where the number of nodes is 2000, we achieved a speedup of 20 times compared with the method without the above techniques.

  • Eliminating Cell Broadband EngineTM DMA Buffer Overflows

    Masana MURASE  

     
    PAPER

      Vol:
    E93-D No:5
      Page(s):
    1062-1069

    This paper presents effective and efficient implementation techniques for DMA buffer overflow elimination on the Cell Broadband EngineTM (Cell/B.E.) processor. In the Cell/B.E. programming model, application developers manually issue DMA commands to transfer data from the system memory to the local memories of the Cell/B.E. cores. Although this allows us to eliminate cache misses or cache invalidation overhead, it requires careful management of the buffer arrays for DMA in the application programs to prevent DMA buffer overflows. To guard against DMA buffer overflows, we introduced safe DMA handling functions for the applications to use. To improve and minimize the performance overhead of buffer overflow prevention, we used three different optimization techniques that take advantage of SIMD operations: branch-hint-based optimizations, jump-table-based optimizations and self-modifying-based optimizations. Our optimized implementation prevents all DMA buffer overflows with minimal performance overhead, only 2.93% average slowdown in comparison to code without the buffer overflow protection.

  • A Knowledge-Based Strategy for the Automated Support to Network Management Tasks

    Sameera ABAR  Tetsuo KINOSHITA  

     
    PAPER-Management Techniques

      Vol:
    E93-D No:4
      Page(s):
    774-788

    This paper presents a domain-ontology driven multi-agent based scheme for representing the knowledge of the communication network management system. In the proposed knowledge-intensive framework, the static domain-related concepts are articulated as the domain knowledge ontology. The experiential knowledge for managing the network is represented as the fault-case reasoning models, and it is explicitly encoded as the core knowledge of multi-agent middleware layer as heuristic production-type rules. These task-oriented management expertise manipulates the domain content and structure during the diagnostic sessions. The agents' rules along with the embedded generic java-based problem-solving algorithms and run-time log information, perform the automated management tasks. For the proof of concept, an experimental network system has been implemented in our laboratory, and the deployment of some test-bed scenarios is performed. Experimental results confirm a marked reduction in the management-overhead of the network administrator, as compared to the manual network management techniques, in terms of the time-taken and effort-done during a particular fault-diagnosis session. Validation of the reusability/modifiability aspects of our system, illustrates the flexible manipulation of the knowledge fragments within diverse application contexts. The proposed approach can be regarded as one of the pioneered steps towards representing the network knowledge via reusable domain ontology and intelligent agents for the automated network management support systems.

  • Entropy Decoding Processor for Modern Multimedia Applications

    Sumek WISAYATAKSIN  Dongju LI  Tsuyoshi ISSHIKI  Hiroaki KUNIEDA  

     
    PAPER-Embedded, Real-Time and Reconfigurable Systems

      Vol:
    E92-A No:12
      Page(s):
    3248-3257

    An entropy decoding engine plays an important role in modern multimedia decoders. Previous researches that focused on the decoding performance paid a considerable attention to only one parameter such as the data parsing speed, but they did not consider the performance caused by a table configuration time and memory size. In this paper, we developed a novel method of entropy decoding based on the two step group matching scheme. Our approach achieves the high performance on both data parsing speed and configuration time with small memory needed. We also deployed our decoding scheme to implement an entropy decoding processor, which performs operations based on normal processor instructions and VLD instructions for decoding variable length codes. Several extended VLD instructions are prepared to increase the bitstream parsing process in modern multimedia applications. This processor provides a solution with software flexibility and hardware high speed for stand-alone entropy decoding engines. The VLSI hardware is designed by the Language for Instruction Set Architecture (LISA) with 23 Kgates and 110 MHz maximum clock frequency under TSMC 0.18 µm technology. The experimental simulations revealed that proposed processor achieves the higher performance and suitable for many practical applications such as MPEG-2, MPEG-4, H.264/AVC and AAC.

  • A Technique for Defining Metamodel Translations

    Iván GARCÍA-MAGARIÑO  Rubén FUENTES-FERNÁNDEZ  

     
    PAPER-Fundamentals of Software and Theory of Programs

      Vol:
    E92-D No:10
      Page(s):
    2043-2052

    Model-Driven Engineering and Domain-Specific Modeling Languages are encouraging an increased used of metamodels for the definition of languages and tools. Although the Meta Object Facility language is the standard for metamodeling, there are alternative metamodeling languages that are aimed at satisfying specific requirements. In this context, sharing information throughout different domains and tools requires not only being able to translate models between modeling languages defined with the same metamodeling language, but also between different metamodeling languages. This paper addresses this latter need describing a general technique to define transformations that perform this translation. In this work, two case studies illustrate the application of this process.

  • Counter-Measures for Relay Failures due to Dynamic Welding: A Robust Engineering Design

    Thomas J. SCHOEPF  

     
    PAPER-Electromechanical Devices and Components

      Vol:
    E92-C No:5
      Page(s):
    728-735

    In prior work, contact welding phenomena were observed in automotive relays during break of motor inrush current. The switching performance of the type of relay investigated could be correlated with the parameters: over-travel, coil suppression, and the break current. In the present work the author further explores the impact of both the contact material (silver tin oxide versus fine grain silver) and the contact surface topography (brand new and pre-aged contacts). He further assesses the robustness of the system "relay" with those parameters using the Taguchi methods for robust design. Furthermore, the robustness of two alternative automotive relay types will be discussed.

  • Enhancing Salt-and-Pepper Noise Removal in Binary Images of Engineering Drawing

    Hasan S. M. AL-KHAFFAF  Abdullah Z. TALIB  Rosalina Abdul SALAM  

     
    PAPER-Image Processing and Video Processing

      Vol:
    E92-D No:4
      Page(s):
    689-704

    Noise removal in engineering drawing is an important operation performed before other image analysis tasks. Many algorithms have been developed to remove salt-and-pepper noise from document images. Cleaning algorithms should remove noise while keeping the real part of the image unchanged. Some algorithms have disadvantages in cleaning operation that leads to removing of weak features such as short thin lines. Others leave the image with hairy noise attached to image objects. In this article a noise removal procedure called TrackAndMayDel (TAMD) is developed to enhance the noise removal of salt-and-pepper noise in binary images of engineering drawings. The procedure could be integrated with third party algorithms' logic to enhance their ability to remove noise by investigating the structure of pixels that are part of weak features. It can be integrated with other algorithms as a post-processing step to remove noise remaining in the image such as hairy noise attached with graphical elements. An algorithm is proposed by incorporating TAMD in a third party algorithm. Real scanned images from GREC'03 contest are used in the experiment. The images are corrupted by salt-and-pepper noise at 10%, 15%, and 20% levels. An objective performance measure that correlates with human vision as well as MSE and PSNR are used in this experiment. Performance evaluation of the introduced algorithm shows better-quality images compared to other algorithms.

  • Evolution and Integration of Medical Laboratory Information System in an Asia National Medical Center

    Po-Hsun CHENG  Sao-Jie CHEN  Jin-Shin LAI  

     
    PAPER

      Vol:
    E92-B No:2
      Page(s):
    379-386

    This work elucidates the evolution of three generations of the laboratory information system in the National Taiwan University Hospital, which were respectively implemented in an IBM Series/1 minicomputer, a client/server and a plug-and-play HL7 interface engine environment respectively. The experience of using the HL7 healthcare information exchange in the hospital information system, laboratory information system, and automatic medical instruments over the past two decades are illustrated and discussed. The latest design challenge in developing intelligent laboratory information services is to organize effectively distributed and heterogeneous medical instruments through the message gateways. Such experiences had spread to some governmental information systems for different purposes in Taiwan; besides, the healthcare information exchange standard, software reuse mechanism, and application service provider adopted in developing the plug-and-play laboratory information system are also illustrated.

  • An Accurate Approach to Large-Scale IP Traffic Matrix Estimation

    Dingde JIANG  Guangmin HU  

     
    LETTER-Network

      Vol:
    E92-B No:1
      Page(s):
    322-325

    This letter proposes a novel method of large-scale IP traffic matrix (TM) estimation, called algebraic reconstruction technique inference (ARTI), which is based on the partial flow measurement and Fratar model. In contrast to previous methods, ARTI can accurately capture the spatio-temporal correlations of TM. Moreover, ARTI is computationally simple since it uses the algebraic reconstruction technique. We use the real data from the Abilene network to validate ARTI. Simulation results show that ARTI can accurately estimate large-scale IP TM and track its dynamics.

  • On Applicability of Formal Methods and Tools to Dependable Services Open Access

    Fuyuki ISHIKAWA  Shinichi HONIDEN  

     
    INVITED PAPER

      Vol:
    E92-B No:1
      Page(s):
    9-16

    As a variety of digital services are provided through networks, more and more efforts are made to ensure dependability of software behavior implementing services. Formal methods and tools have been considered as promising means to support dependability in complex software systems during the development. On the other hand, there have been serious doubts on practical applicability of formal methods. This paper overviews the present state of formal methods and discusses their applicability, especially focusing on two representative methods (SPIN and B Method) and their recent industrial applications. This paper also discusses applications of formal methods to dependable networked software.

  • Analysis of Two-Phase Path Management Scheme for MPLS Traffic Engineering

    Hitomi TAMURA  Kenji KAWAHARA  Yuji OIE  

     
    PAPER

      Vol:
    E92-B No:1
      Page(s):
    59-67

    Traffic Engineering (TE) is important for improving QoS in forwarding paths by efficient use of network resources. In fact, MPLS allows several detour paths to be (pre-)established for some source-destination pair as well as its primary path of minimum hops. Thus, we focus on a two-phase path management scheme using these two kinds of paths. In the first phase, each primary path is allocated to a flow on a specific source-destination pair if the path is not congested, i.e., if its utilization is less than some predetermined threshold; otherwise, as the second phase, one of the detour paths is allocated randomly if the path is available. Therefore, in this paper, we analytically evaluate this path management scheme by extending the M/M/c/c queueing system, and through some numerical results we investigate the impact of a threshold on the flow-blocking probability. Through some numerical results, we discuss the adequacy of the path management scheme for MPLS-TE.

  • Estimation of Current Traffic Matrices from Long-Term Traffic Variations

    Yuichi OHSITA  Takashi MIYAMURA  Shin'ichi ARAKAWA  Eiji OKI  Kohei SHIOMOTO  Masayuki MURATA  

     
    PAPER-Network

      Vol:
    E92-B No:1
      Page(s):
    171-183

    Obtaining current traffic matrices is essential to traffic engineering (TE) methods. Because it is difficult to monitor traffic matrices, several methods for estimating them from link loads have been proposed. The models used in these methods, however, are incorrect for some real networks. Thus, methods improving the accuracy of estimation by changing routes also have been proposed. However, existing methods for estimating the traffic matrix by changing routes can only capture long-term variations and cannot obtain current traffic matrices accurately. In this paper, we propose a method for estimating current traffic matrices that uses route changes introduced by a TE method. In this method, we first estimate the long-term variations of traffic by using the link loads monitored at previous times. Then, we adjust the estimated long-term variations so as to fit the current link loads. In addition, when the traffic variation trends change and the estimated long-term variations fail to match the current traffic, our method detects mismatch. Then, so as to capture the current traffic variations, the method re-estimates the long-term variations after removing monitored data corresponding to the end-to-end traffic causing the mismatches. We evaluate our method through simulation. The results show that our method can estimate current traffic matrices accurately even when some end-to-end traffic changes suddenly.

  • Distributed Computing Software Building-Blocks for Ubiquitous Computing Societies

    K.H. (Kane) KIM  

     
    INVITED PAPER

      Vol:
    E91-D No:9
      Page(s):
    2233-2242

    The steady approach of advanced nations toward realization of ubiquitous computing societies has given birth to rapidly growing demands for new-generation distributed computing (DC) applications. Consequently, economic and reliable construction of new-generation DC applications is currently a major issue faced by the software technology research community. What is needed is a new-generation DC software engineering technology which is at least multiple times more effective in constructing new-generation DC applications than the currently practiced technologies are. In particular, this author believes that a new-generation building-block (BB), which is much more advanced than the current-generation DC object that is a small extension of the object model embedded in languages C++, Java, and C#, is needed. Such a BB should enable systematic and economic construction of DC applications that are capable of taking critical actions with 100-microsecond-level or even 10-microsecond-level timing accuracy, fault tolerance, and security enforcement while being easily expandable and taking advantage of all sorts of network connectivity. Some directions considered worth pursuing for finding such BBs are discussed.

  • A Single Camera Motion Capture System for Human-Computer Interaction

    Ryuzo OKADA  Bjorn STENGER  

     
    PAPER

      Vol:
    E91-D No:7
      Page(s):
    1855-1862

    This paper presents a method for markerless human motion capture using a single camera. It uses tree-based filtering to efficiently propagate a probability distribution over poses of a 3D body model. The pose vectors and associated shapes are arranged in a tree, which is constructed by hierarchical pairwise clustering, in order to efficiently evaluate the likelihood in each frame. A new likelihood function based on silhouette matching is proposed that improves the pose estimation of thinner body parts, i.e. the limbs. The dynamic model takes self-occlusion into account by increasing the variance of occluded body-parts, thus allowing for recovery when the body part reappears. We present two applications of our method that work in real-time on a Cell Broadband EngineTM: a computer game and a virtual clothing application.

  • Improvements of HITS Algorithms for Spam Links

    Yasuhito ASANO  Yu TEZUKA  Takao NISHIZEKI  

     
    PAPER-Scoring Algorithms

      Vol:
    E91-D No:2
      Page(s):
    200-208

    The HITS algorithm proposed by Kleinberg is one of the representative methods of scoring Web pages by using hyperlinks. In the days when the algorithm was proposed, most of the pages given high score by the algorithm were really related to a given topic, and hence the algorithm could be used to find related pages. However, the algorithm and the variants including Bharat's improved HITS, abbreviated to BHITS, proposed by Bharat and Henzinger cannot be used to find related pages any more on today's Web, due to an increase of spam links. In this paper, we first propose three methods to find "linkfarms," that is, sets of spam links forming a densely connected subgraph of a Web graph. We then present an algorithm, called a trust-score algorithm, to give high scores to pages which are not spam pages with a high probability. Combining the three methods and the trust-score algorithm with BHITS, we obtain several variants of the HITS algorithm. We ascertain by experiments that one of them, named TaN+BHITS using the trust-score algorithm and the method of finding linkfarms by employing name servers, is most suitable for finding related pages on today's Web. Our algorithms take time and memory no more than those required by the original HITS algorithm, and can be executed on a PC with a small amount of main memory.

61-80hit(173hit)