The search functionality is under construction.

Keyword Search Result

[Keyword] consistency(57hit)

41-57hit(57hit)

  • Single-Particle Approach to Self-Consistent Monte Carlo Device Simulation

    Fabian M. BUFLER  Christoph ZECHNER  Andreas SCHENK  Wolfgang FICHTNER  

     
    PAPER

      Vol:
    E86-C No:3
      Page(s):
    308-313

    The validity and capability of an iterative coupling scheme between single-particle frozen-field Monte Carlo simulations and nonlinear Poisson solutions for achieving self-consistency is investigated. For this purpose, a realistic 0.1 µm lightly-doped-drain (LDD) n-MOSFET with a maximum doping level of about 2.5 1020 cm-3 is simulated. It is found that taking the drift-diffusion (DD) or the hydrodynamic (HD) model as initial simulation leads to the same Monte Carlo result for the drain current. This shows that different electron densities taken either from a DD or a HD simulation in the bulk region, which is never visited by Monte Carlo electrons, have a negligible influence on the solution of the Poisson equation. For the device investigated about ten iterations are necessary to reach the stationary state after which gathering of cumulative averages can begin. Together with the absence of stability problems at high doping levels this makes the self-consistent single-particle approach (SPARTA) a robust and efficient method for the simulation of nanoscale MOSFETs where quasi-ballistic transport is crucial for the on-current.

  • Caching and Concurrency Control in a Wireless Mobile Computing Environment

    SangKeun LEE  

     
    PAPER-Databases

      Vol:
    E85-D No:8
      Page(s):
    1284-1296

    Caching of frequently accessed data has been shown to be a useful technique for reducing congestion on the narrow bandwidth of wireless channels. However, traditional client/server strategies for supporting transactional cache consistency, which require extensive communications between a client and a server, are not appropriate in a wireless mobile database. This paper proposes two, simple but effective, transactional cache consistency protocols for mobile read-only transactions by utilizing the broadcast-based solutions for the problem of invalidating caches. The novelty of our approach is that the consistency check on accessed data and the commitment protocol are implemented in a truly distributed fashion as an integral part of cache invalidation process. The applicability of proposed techniques is also examined by an analytical study.

  • Exploiting Metadata of Absent Objects for Proxy Cache Consistency

    Jooyong KIM  Hyokyung BAHN  Kern KOH  

     
    PAPER-Network

      Vol:
    E84-B No:5
      Page(s):
    1406-1412

    Caching at the Web proxy server plays an important role in reducing the response time, the network traffic, and the load of Web servers. Many recent studies have proposed and examined the replacement and consistency policies for the proxy cache, which plays a central role in the performance of caching components. For better performance, they exploit various metadata of Web objects, such as the reference count, reference time, and modification time information of past behaviors, to estimate the re-reference likelihood and freshness of the objects. However, all of these known to the authors use the metadata only when the actual object is in the cache. We observed from various proxy traces that about 20-30% of clients' requests incurred only the validity checks of cached objects without transferring actual objects from the proxy server. In this case, only the metadata are necessary at the proxy server. This paper proposes a proxy cache consistency policy that uses the metadata even for absent objects. These include the time information of evicted objects from the cache and those out of the header-only replies from Web servers. Trace-driven simulations with public proxy cache traces show that our policy reduces the response time and the number of connections to Web servers significantly.

  • Complexity of the Type-Consistency Problem for Acyclic Object-Oriented Database Schemas

    Shougo SHIMIZU  Yasunori ISHIHARA  Junji YOKOUCHI  Minoru ITO  

     
    PAPER-Databases

      Vol:
    E84-D No:5
      Page(s):
    623-634

    Method invocation mechanism is one of the essential features in object-oriented programming languages. This mechanism contributes to data encapsulation and code reuse, but there is a risk of runtime type errors. In the case of object-oriented databases (OODBs), a runtime error causes rollback. Therefore, it is desirable to ensure that a given OODB schema is consistent, i.e., no runtime error occurs during the execution of queries under any database instance of the OODB schema. This paper discusses the computational complexity of the type-consistency problem. As a model of OODB schemas, we adopt update schemas introduced by Hull et al., which have all of the basic features of OODBs such as class hierarchy, inheritance, complex objects, and so on. The type-consistency problem for update schemas is known to be undecidable. We introduce a subclass of update schemas, called acyclic schemas, and show that the type-consistency problem for acyclic schemas is in coNEXPTIME. Furthermore, we show that the problem for recursion-free acyclic schemas is coNEXPTIME-hard and the problem for retrieval acyclic schemas is PSPACE-complete.

  • Ensuring Latest-Bound Currency of Read-Only Transactions in Mobile Broadcasting Environments

    Boohyung HAN  Sung Kwon CHUNG  Yookun CHO  

     
    PAPER-Databases

      Vol:
    E84-D No:4
      Page(s):
    456-464

    In mobile broadcasting environments, an information server periodically broadcasts a set of data items to a large mobile client population at every broadcast cycle and mobile clients retrieve the data items they need upon arrival at the broadcast channel. In such environments, the cost of data delivery is independent of the number of clients. Many applications such as auctions and stock quotes perform read-only transactions that require the clients to read consistent and current data for accurate result. Previous concurrency control mechanisms designed for such environments ensure that the clients read consistent data, but they cannot ensure latest-bound currency which allows the clients to read the latest or most up-to-date data. In this paper, we propose an efficient concurrency control mechanism that ensures latest-bound currency as well as update consistency, which is appropriate for the mobile broadcasting environments. To ensure latest-bound currency, the server computes control information using "virtual" broadcast cycles. This control information is also used for checking update consistency. Thus, the proposed mechanism allows all data committed in current broadcast cycle to be broadcast. We have performed simulation experiments to measure transaction aborts in order to evaluate the performance of the proposed mechanism. The result confirms that the proposed mechanism produces no stale reads and also shows that the proposed mechanism generates less transaction aborts than previous mechanisms, which implies that we can get higher data currency without increasing data inconsistency.

  • An Optimistic Cache Consistency Protocol Using Preemptive Approach

    SungHo CHO  Jeong-Hyon HWANG  Kyoung Yul BAE  Chong-Sun HWANG  

     
    PAPER-Databases

      Vol:
    E83-D No:9
      Page(s):
    1772-1780

    In Optimistic Two-Phase Locking (O2PL), when a transaction requests a commit, the transaction can not be committed until all requested locks are obtained. By this reason, O2PL leads to unnecessary waits and operations even though it adopts an optimistic approach. This paper suggests an efficient optimistic cache consistency protocol that provides serializability of committed transactions. Our cache consistency scheme, called PCP (Preemptive Cache Protocol), decides whether to commit or abort without waiting when transactions request commits. In PCP, some transactions that read stale data items can not be aborted, because it adopts a re-ordering scheme to enhance the performance. In addition, for re-ordering, PCP stores only one version of each data item. This paper presents a simulation-based analysis on the performance of PCP with other protocols such as O2PL, Optimistic Concurrency Control and Caching Two-Phase Locking. The simulation experiments show that PCP performs as well as or better than other schemes with low overhead.

  • A Supporting Method of Making a Consistent Software Requirements Specification Based on the Dempster and Shafer's Theory

    Hideaki SUGIMOTO  Atsushi OHNISHI  

     
    PAPER-Theory and Methodology

      Vol:
    E83-D No:4
      Page(s):
    659-668

    A software requirements specification (SRS) is a document at the first phase of software development. Since it is difficult to make an accurate SRS at the beginning of software development, we propose a supporting method to detect and interpret the inconsistency of SRS. First, we classify and define the inconsistency of SRS. Next, we describe how to detect and interpret the inconsistency of SRS. We use the Requirements Frame Model to detect the inconsistency of SRS. We apply the Dempster and Shafer's theory to interpret the inconsistency of SRS. We illustrate our method with an example.

  • A False-Sharing Free Distributed Shared Memory Management Scheme

    Alexander I-Chi LAI  Chin-Laung LEI  Hann-Huei CHIOU  

     
    PAPER-Computer Systems

      Vol:
    E83-D No:4
      Page(s):
    777-788

    Distributed shared memory (DSM) systems on top of network of workstations are especially vulnerable to the impact of false sharing because of their higher memory transaction overheads and thus higher false sharing penalties. In this paper we develop a dynamic-granularity shared memory management scheme that eliminates false sharing without sacrificing the transparency to conventional shared-memory applications. Our approach utilizes a special threaded splay tree (TST) for shared memory information management, and a dynamic token-based path-compression synchronization algorithm for data transferring. The combination of the TST and path compression is quite efficient; asymptotically, in an n-processor system with m shared memory segments, synchronizing at most s segments takes O(s log m log n) amortized computation steps and generates O(s log n) communication messages, respectively. Based on the proposed scheme we constructed an experimental DSM prototype which consists of several Ethernet-connected Pentium-based computers running Linux. Preliminary benchmark results on our prototype indicate that our scheme is quite efficient, significantly outperforming traditional schemes and scaling up well.

  • Data Transfer Evaluation of Nomadic Data Consistency Model for Large-Scale Mobile Systems

    Masahiro KURODA  Ryoji ONO  Yoshiki SHIMOTSUMA  Takashi WATANABE  Tadanori MIZUNO  

     
    PAPER

      Vol:
    E82-D No:4
      Page(s):
    822-830

    The optimistic consistency scheme has been established with respect to data consistency and availability in distributed systems. The nomadic data consistency model using version vectors to support data versioning for data synchronization and concurrent conflict detection is suitable for an optimistic replication system that supports large-scale wireless networks. This paper describes the architecture and its data consistency model using data versioning and its access domain control targeted for nomadic data sharing systems, such as collaborative works using database and messaging, and the data transfer optimizations of the model. We evaluate our data versioning scheme comparing with a traditional data versioning and the data transfer optimization by estimation and measurement assuming a mobile worker's job. We generate arithmetic formulas for data transfer estimation using the optimizing techniques and apply them to large-scale data sharing configurations in which collaboration groups are dynamically formed and data is exchanged in each group. The data versioning with an access domain increases flexibility in data sharing configurations, such as mobile collaboration systems and client/server type mobile systems. We confirmed that the combination of the general optimizations and the access domain configurations based on our data consistency model is applicable for large-scale mobile data sharing systems.

  • Selective Write-Update: A Method to Relax Execution Constraints in a Critical Section

    Jae Bum LEE  Chu Shik JHON  

     
    PAPER-Computer Systems

      Vol:
    E81-D No:11
      Page(s):
    1186-1194

    In a shared-memory multiprocessor, shared data are usually accessed in a critical section that is protected by a lock variable. Therefore, the order of accesses by multiple processors to the shared data corresponds to the order of acquiring the ownership of the lock variable. This paper presents a selective write-update protocol, where data modified in a critical section are stored in a write cache and, at a synchronization point, they are transferred only to the processor that will execute the critical section following the current processor. By using QOLB synchronization primitives, the next processor can be determined at the execution time. We prove that the selective write-update protocol ensures data coherency of parallel programs that comply with release consistency, and evaluate the performance of the protocol by analytical modeling and program-driven simulation. The simulation results show that our protocol can reduce the number of coherence misses in a critical section while avoiding the multicast of write-update requests on an interconnection network. In addition, we observe that synchronization latency can be decreased by reducing both the execution time of a critical section and the number of write-update requests. From the simulation results, it is shown that our protocol provides better performance than a write-invalidate protocol and a write-update protocol as the number of processors increases.

  • Blind Identification of IIR Model Based on Output Over-Sampling

    Hajime KAGIWADA  Lianming SUN  Akira SANO  Wenjiang LIU  

     
    PAPER-Digital Signal Processing

      Vol:
    E81-A No:11
      Page(s):
    2350-2360

    A new identification algorithm based on output over-sampling scheme is proposed for a IIR model whose input signal can not be available directly. By using only an output signal sampled at higher rate than unknown input, parameters of the IIR model can be identified. It is clarified that the consistency of the obtained parameter estimates is assured under some specified conditions. Further an efficient recursive algorithm for blind parameter estimation is also given for practical applications. Simulation results demonstrate its effectiveness in both system and channel identification.

  • Testing Type Consistency of Method Schemas

    Hiroyuki SEKI  Yasunori ISHIHARA  Hiroki DODO  

     
    PAPER-Software Theory

      Vol:
    E81-D No:3
      Page(s):
    278-287

    Method schemas were proposed as a formal model of object-oriented languages. A method schema S is called consistent if, for each method invocation during the execution of S, a method definition to be bound to the invoked method name is uniquely determined. However, the consistency testing problem is known to be undecidable in general. This paper presents an algorithm which analyzes the consistency of a given method schema. The algorithm decides the consistency problem in polynomial time for monadic method schemas. We also provide an incremental algorithm for testing consistency after updates of a method schema.

  • Requirements Specification and Analysis of Digital Systems Using FARHDL

    Victor R. L. SHEN  Feng-Ho KUO  Feipei LAI  

     
    PAPER-Artificial Intelligence and Cognitive Science

      Vol:
    E81-D No:3
      Page(s):
    317-328

    As expert system technology gains wider acceptance in digital system design, the need to build and maintain a large scale knowledge base will assume greater importance. However, how to build a correct and efficient rule base is even a hard part in the knowledge-based system development. In this paper, we develop FARHDL (Frame-And-Rule-based Hardware Description Language) to form a knowledge base. The FARHDL is simple but powerful to specify the hardware requirements and can be directly simulated by PROLOG. Through the knowledge base transformed from FARHDL, a formal method can be developed to design, implement, and validate the digital hardware systems. Furthermore, behavioral properties, anomaly properties, structural properties, and timing properties are applied to analyze the requirements specification. The purposes of those properties are used to detect explicit/implicit incorrect specification clauses and to capture some desired requirements, such as completeness and consistency. Finally, the analysis results can be a useful tool for finding obscure problems in tricky digital system designs and can also aid in the development of formal specifications.

  • Adsmith: An Object-Based Distributed Shared Memory System for Networks of Workstations

    Wen-Yew LIANG  Chung-Ta KING  Feipei LAI  

     
    PAPER-Computer Architecture

      Vol:
    E80-D No:9
      Page(s):
    899-908

    This paper introduces an object-based distributed shared memory (DSM) system called Adsmith. The primary goal of Adsmith is to provide a low-cost, portable, and efficient DSM for networks of workstations (NOW). Adsmith achieves this goal by building on top of PVM, a widely supported communication subsystem, as a user-level library and by incorporating many traffic reduction and latency hiding techniques. Issues involved in the design of Adsmith and our solution strategies will be discussed. Preliminary performance evaluation of Adsmith on a network of Pentium computers will be presented. The results show that programs developed with Adsmith can achieve a performance comparable to that developed with PVM.

  • Deferred Locking with Buffer Validation on Demand for Client-Server Database Consistency: DL

    Hyeokmin KWON  Songchun MOON  

     
    PAPER-Databases

      Vol:
    E80-D No:7
      Page(s):
    705-716

    In client-server database management systems (DBMSs), inter-transaction caching is an effective technique for improving the performance. However, inter-transaction caching requires a cache consistency maintenance (CCM) protocol to ensure that cached copies at clients are kept mutually consistent. Such a protocol could be complex to implement and expensive to run, since several rounds of message exchange may be required. In this paper, we propose a new CCM scheme based on the primary-copy locking algorithm. In the proposed scheme, a number of lock requests and a data-shipping request are combined into a single message packet to reduce client-server interactions, which are known to be very critical to the performance of clientserver DBMSs. We examine its performance tradeoffs on the basis of a simulation model under a wide range of workloads. The performance results indicate that the proposed scheme improves the overall system throughput significantly over the caching two-phase locking and the optimistic two-phase locking scheme. Its higher performance mainly results from its lower communication overhead and lower degree of transaction blocking ratio.

  • A Cache-Coherent, Distributed Memory Multiprocessor System and Its Performance Analysis

    Douglas E. MARQUARDT  Hasan S. ALKHATIB  

     
    PAPER-Computer Systems

      Vol:
    E75-D No:3
      Page(s):
    274-290

    The problems of cache coherency in multiprocessor systems are directly related to their architectural structures. Small scale multiprocessor systems have focused on the use of bus based memory interconnection networks using centrally shared memory and a sequential consistency model for coherency. This has limited scalability to but a few tens of processors due to the limited bus bandwidth used for both coherency updates and memory traffic. Recently, large scale multiprocessor systems have been proposed that use general interconnection networks and distributed shared memory. These architectures have been proposed using weak consistency models and various directory map schemes to hide the overhead for coherency maintenance within the memory hieratchy, interconnection network or process context switch latencies. The coherency and memory traffic are still maintained over the same interconnection network. In this paper, we present the architecture of a new general purpose medium scale multiprocessor system. This Cache Coherent Multiprocessor System (C2MP), supports distributed shared memory using a general memory interconnection network for memory traffic and a separate bus based coherency interconnection network for coherency maintenance. Through the use of a special directory based coherency protocol and cache oriented distributed coherency controllers, direct cache-to-cache coherency maintenance is performed over the dedicated coherency bus. This minimizes coherency updates to only those processor nodes needing coherency maintenance. An aggressive sequential coherncy model is used, which reduces the hardware penalty to support an ideal sequential consistency programmers model. The system can scale up to 256-512 processors depending on the degree of shared data and is expected to have higher per processor utilization in this range than currently proposed medium and large scale multiprocessor systems. The C2MP system is analyzed utilizing a Generalized Timed Petri-Net model of a processor node. A stochastic model for internode interactions over the general memory interconnection network and coherency bus are used . The model of the proposed architecture is analyzed under steady-state conditions for varying system work load parameters.

  • A Distributed Mutual Exclusion Algorithm Based on Weak Copy Consistency

    Seoung Sup LEE  Ha Ryoung OH  June Hyoung KIM  Won Ho CHUNG  Myunghwan KIM  

     
    PAPER-Computer Networks

      Vol:
    E75-D No:3
      Page(s):
    298-306

    This paper presents a destributed algorithm that uses weak copy consistency to create mutual exclusion in a distributed computer system. The weak copy consistency is deduced from the uncertainty of state which occurs due to the finite and unpredictable communication delays in a distributed environment. Also the method correlates outdated state information to current state. The average number of messages to enter critical section in the algorithm is n/2 to n messages where n is the number of sites. We show that the algorithm achieves mutual exclusion and the fairness and liveness of the algorithm is proven. We study the performance of the algorithm by simulation technique.

41-57hit(57hit)