The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] CAC(297hit)

201-220hit(297hit)

  • A Randomized Online Algorithm for the File Caching Problem

    Seiichiro TANI  Toshiaki MIYAZAKI  

     
    PAPER-Algorithms

      Vol:
    E86-D No:4
      Page(s):
    686-697

    Caching web files reduces user response time as well as network traffic. When implementing caches, the file caching problem must be addressed; the problem is how to determine which files should be evicted from a cache when there is insufficient space for storing a new file so that the sum of the mis-hit (fault) file costs is minimized. Greedy-Dual-Size (GDS) is the best online algorithm in terms of competitiveness, i. e. , (k)/(k-h+1)-competitive, where k and h are the storage space of, respectively, GDS and an optimal offline algorithm. GDS performs very well even in trace-driven simulations. The worst-case time taken to service a request is another important measure for online file caching algorithms since slow response times render caching meaningless from the client's view point. This paper proposes a fast randomized (k)/(k-h+1)-competitive algorithm that performs in O(2log ^* k) time per file eviction or insertion, whereas GDS takes O(log k) time, where 2log ^* k is a much slower increasing function than log k. To confirm its practicality, we conduct trace driven simulations. Experimental results show that our algorithm attains only slightly worse byte hit rates and sufficiently large reduced latency in comparison with GDS, and our algorithm is a good candidate for caches requiring high-speed processing such as second-level caches in the large networks.

  • Reducing Memory System Energy by Software-Controlled On-Chip Memory

    Masaaki KONDO  Hiroshi NAKAMURA  

     
    PAPER-Architecture and Algorithms

      Vol:
    E86-C No:4
      Page(s):
    580-588

    In recent computer systems, a large portion of energy is consumed by on-chip cache accesses and data movement between cache and off-chip main memory. Reducing these memory system energy is indispensable for future microprocessors because power and thermal issues certainly become a key factor of limiting processor performance. In this paper, we discuss and evaluate how our architecture called SCIMA contributes to energy saving. SCIMA integrates software-controllable memory (SCM) into processor chip. SCIMA can save total memory system energy by using SCM under the support of compiler. The evaluation results reveal that SCIMA can reduce 5-50% of memory system energy and still faster than conventional cache based architecture.

  • Improving the Performance of Linux Operating System via Buffer Cache Partitioning and Prefetching

    Heung Seok JEON  Sam H. NOH  

     
    PAPER-Software Systems

      Vol:
    E86-D No:3
      Page(s):
    616-622

    Buffer caching is an integral part of the operating system. In this paper, we propose a scheme that integrates buffer cache management and prefetching via cache partitioning. The scheme, which we call SA-W2R, is simple to implement, making it a feasible solution in real systems. In its basic form, for buffer replacement, it uses the LRU policy. However, its modular design allows for any replacement policy to be incorporated into the scheme. For prefetching, it uses the LRU-One Block Lookahead (LRU-OBL) approach, eliminating any extra burden that is generally necessary in other prefetching approaches. Implementation studies based on the GNU/Linux kernel version 2.2.14 show that the SA-W2R performs better than the scheme currently used, with a maximum increases of 23% for the workloads considered.

  • Constructing a Cactus for Minimum Cuts of a Graph in O(mn+n2log n) Time and O(m) Space

    Hiroshi NAGAMOCHI  Shuji NAKAMURA  Toshimasa ISHII  

     
    PAPER-Graph Algorithms

      Vol:
    E86-D No:2
      Page(s):
    179-185

    It is known that all minimum cuts in an edge-weighted undirected graph with n vertices and m edges can be represented by a cactus with O(n) vertices and edges, a connected graph in which each edge is contained in an exactly one cycle. In this paper, we show that such a cactus representation can be computed in O(mn+n2log n) time and O(m) space. This improves the previously best complexity of deterministic cactus construction algorithms, and matches with the time bound of the fastest deterministic algorithm for computing a single minimum cut.

  • Call Admission Control with QoS Class Modification

    Toshiaki TSUCHIYA  

     
    PAPER-Packet Transmission

      Vol:
    E86-B No:2
      Page(s):
    682-689

    In IP networks, QoS guarantee is becoming important, to allow real-time services such as voice over IP (VoIP) and video conferencing. To guarantee various QoS requirements from a variety of applications, call admission control (CAC) together with a service differentiation mechanism is useful. Service differentiation enables the network to provide different QoSs to various applications, by assigning priority classes to calls. Then, CAC limits the acceptance of new calls to prevent the QoSs of established calls from degrading. For this environment, we propose a CAC method in which priority classes assigned to already established calls are changed adaptively. Classes may be modified when the CAC function handles the acceptance of a new call if that can decrease the blocking probability. The effect of the method is shown by numerical examples.

  • Performance Analysis of HIPERLAN Channel Access Control Protocol

    KwangOh CHO  HyungCheol SHIN  JongKyu LEE  

     
    PAPER

      Vol:
    E85-B No:10
      Page(s):
    2044-2050

    In this paper, the performance of HIPERLAN (HIgh PErformance Radio Local Area Networks) CAC (Channel Access Control) of ETSI (European Telecommunication Standards Institute) in Europe, as High speed wireless LAN, is analyzed mathematically. The CAC protocol of HIPERLAN is the EY-NPMA (Elimination Yield-Nonpreemptive Priority Multiple Access) in which data is transmitted after prioritization, elimination and yield phase. We analyzed channel contention phase composed of elimination and yield phase and then throughput is inspected by simulation. This result is useful to design and implement of Ad hoc wireless networks.

  • Stochastic Model of Internet Access Patterns: Coexistence of Stationarity and Zipf-Type Distributions

    Masaki AIDA  Tetsuya ABE  

     
    PAPER-Fundamental Theories

      Vol:
    E85-B No:8
      Page(s):
    1469-1478

    This paper investigates the stochastic property of packet destinations in order to describe Internet access patterns. If we assume a sort of stationary condition for the address generation process, the process is an LRU stack model. Although the LRU stack model gives appropriate descriptions of address generation on a medium/long time-scale, address sequences generated from the LRU stack model do not reproduce Zipf-type distributions, which appear frequently in Internet access patterns. This implies that the address generation behavior on a short time-scale has a strong influence on the shape of the distributions that describe frequency of address appearances. This paper proposes an address generation algorithm that does not meet the stationary condition on the short time-scale, but restores it on the medium/long time-scale, and shows that the proposed algorithm reproduces Zipf-type distributions.

  • Caching and Concurrency Control in a Wireless Mobile Computing Environment

    SangKeun LEE  

     
    PAPER-Databases

      Vol:
    E85-D No:8
      Page(s):
    1284-1296

    Caching of frequently accessed data has been shown to be a useful technique for reducing congestion on the narrow bandwidth of wireless channels. However, traditional client/server strategies for supporting transactional cache consistency, which require extensive communications between a client and a server, are not appropriate in a wireless mobile database. This paper proposes two, simple but effective, transactional cache consistency protocols for mobile read-only transactions by utilizing the broadcast-based solutions for the problem of invalidating caches. The novelty of our approach is that the consistency check on accessed data and the commitment protocol are implemented in a truly distributed fashion as an integral part of cache invalidation process. The applicability of proposed techniques is also examined by an analytical study.

  • Trends in High-Performance, Low-Power Cache Memory Architectures

    Koji INOUE  Vasily G. MOSHNYAGA  Kazuaki MURAKAMI  

     
    PAPER-High-Performance Technologies

      Vol:
    E85-C No:2
      Page(s):
    304-314

    One of uncompromising requirements from portable computing is energy efficiency, because that affects directly the battery life. On the other hand, portable computing will target more demanding applications, for example moving pictures, so that higher performance is still required. Cache memories have been employed as one of the most important components of computer systems. In this paper, we briefly survey architectural techniques for high performance, low power cache memories.

  • Omitting Cache Look-up for High-Performance, Low-Power Microprocessors

    Koji INOUE  Vasily G. MOSHNYAGA  Kazuaki MURAKAMI  

     
    PAPER-Low-Power Technologies

      Vol:
    E85-C No:2
      Page(s):
    279-287

    In this paper, we propose a novel architecture for low-power direct-mapped instruction caches, called "history-based tag-comparison (HBTC) cache. " The cache attempts to reuse tag-comparison results for avoiding unnecessary tag checks. Execution footprints are recorded into an extended BTB (Branch Target Buffer). In our evaluation, it is observed that the energy for tag comparison can be reduced by more than 90% in many applications.

  • Asynchronous Cache Invalidation Strategy to Support Read-Only Transaction in Mobile Environments

    SungHun NAM  IlYoung CHUNG  SungHo CHO  ChongSun HWANG  

     
    PAPER-Databases

      Vol:
    E85-D No:2
      Page(s):
    373-385

    The stateless-based cache invalidation schemes for wireless environments can be categorized into either asynchronous or synchronous cache invalidation according to the broadcasting way of invalidation report. However, if the asynchronous cache invalidation scheme attempts to support local processing of read-only transaction, a critical problem may occur; the asynchronous invalidation reports provide no guarantee of waiting time for mobile transactions requesting commit. To solve this problem, the server in our approaches broadcasts two kind of messages, asynchronous invalidation report to reduce transaction latency and periodic guide message to avoid the uncertainty of waiting time for the next invalidation report. This paper presents a simulation-based analysis on the performance of the suggesting algorithms. The simulation experiments show that the local processing algorithms of read-only transaction based on asynchronous cache invalidation scheme get better response time than the algorithm based on synchronous cache invalidation scheme.

  • Designs of Building Blocks for High-Speed, Low-Power Processors

    Tadayoshi ENOMOTO  

     
    PAPER-High-Performance Technologies

      Vol:
    E85-C No:2
      Page(s):
    331-338

    A fast, low-power 16-bit adder, 32-word register file and 512-bit cache SRAM have been developed using 0.25-µm GaAs HEMT technology for future multi-GHz processors. The 16-bit adder, which uses a negative logic binary look-ahead carry structure based on NOR gates, operates at the maximum clock frequency of 1.67 GHz and consumes 134.4 mW at a supply voltage of 0.6 V. The active area is 1.6 mm2 and there are about 1,230 FETs. A new DC/DC level converter has been developed for use in high-speed, low-power storage circuits such as SRAMs and register files. The level converter can increase the DC voltage, which is supplied to an active-load circuit on request, or supply a minimal DC voltage to a load circuit in the stand-by mode. The power dissipation (P) of the 32-word register file with on-chip DC/DC level converters is 459 mW, a reduction to 25.2% of that of an equivalent conventional register file, while the operating frequency (fc) was 5.17 GHz that is 74.8% of fc for the conventional register file. P for the 512-bit cache SRAM with the new DC/DC level converters is 34.3 mW, 89.7% of the value for an equivalent conventional cache SRAM, with the read-access time of 455 psec, only 1.1% longer than that of the conventional cache SRAM.

  • Performance Evaluation of a Load Balancing Routing Algorithm for Clustered Multiple Cache Servers

    Hiroyoshi MIWA  Kazunori KUMAGAI  Shinya NOGAMI  Takeo ABE  Hisao YAMAMOTO  

     
    PAPER

      Vol:
    E85-B No:1
      Page(s):
    147-156

    The explosive growth of World Wide Web usage is causing a number of performance problems, including slow response times, network congestion, and denial of service. Web site that has a huge number of accesses and requires high quality of services, such as a site offering hosting services, or content delivery services, usually uses a cache server to reduce the load on the original server offering the original content. To increase the throughput of the caching process and to improve service availability, multiple cache servers are often positioned in front of the original server. This requires a switch to direct incoming requests to one of the multiple cache servers. In this paper, we propose a routing algorithm for such a switch in front of clustered multiple cache servers and evaluate its performance by simulation. The results show that our routing algorithm is effective when content has request locality and a short period of validity, for example, news, map data, road traffic data, or weather information. We also identify points to consider when the proposed algorithm is applied to a real system.

  • Providing Proportional Loss Rate for Adaptive Traffic: A New Relative DiffServ Model

    Wei WU  Yong REN  Xiuming SHAN  

     
    PAPER

      Vol:
    E85-B No:1
      Page(s):
    129-136

    Most applications can adapt their coding techniques and sending rates according to the network congestion and the resource needed can be provided at the beginning of the transmission. So traditional Differentiated Services (DiffServ) model is too rigid to them. In this paper, we are seeking a balance between the relative DiffServ and the absolute DiffServ and propose a new Diffserv model, a relative Differentiated Service model with admission control, which suits the adaptive application. By providing the proportional differentiated services in core routers and loss-rate based CAC control in edge routers, we can make both the network and the users adaptive: the network is adaptive to the traffic load and the users is adaptive to the network congestion. This model is promising to the elastic but unpredictable traffic, such as IP telephony or other multimedia applications.

  • IN Service Provision Using a Caching-Based Mobile Agent in the Next Generation Network

    Ji-Young LEE  Youngsik MA  Yeon-Joong KIM  Dong-Ho KIM  Sunshin AN  

     
    PAPER-Mobile Service and Technologies

      Vol:
    E84-B No:12
      Page(s):
    3141-3154

    As the network speed becomes faster and requirements about various services are increased, a number of groups are currently developing technologies aimed at evolving and enhancing the capabilities of existing network. A Next-Generation Network (NGN) is defined as a hybrid telecommunications network that employs new distributed processing techniques to provide all types of services. By integrating the Intelligent Network (IN) technology and the Mobile Agent (MA) technology we can support service flexibility and service portability in NGN. In this paper, we propose a caching-based mobile agent model for NGN and analyze the performance of this model. The mobile agent technology increases the service portability and the caching strategy does the service reusability. Each Physical Entity (PE) has MAs within their repository through the caching strategy and processes service requests from users without the control of the central system such as Service Control Point (SCP). Therefore, we can decrease the total network load and the response time for user requests.

  • An Efficient Standard-Compatible Traffic Description Parameter Selection Algorithm for VBR Video Sources

    Heejune AHN  Andrea BAIOCCHI  Jae-kyoon KIM  

     
    LETTER-Fundamental Theories

      Vol:
    E84-B No:12
      Page(s):
    3274-3277

    The international telecommunication standards bodies such as ITU-T, ATM Forum, and IETF recommend the dual leaky bucket for the traffic specifications for VBR service. On the other hand, recent studies have demonstrated multiple time-scale burstiness in compressed video traffic. In order to fill this gap between the current standards and real traffic characteristics, we present a standard-compatible traffic parameter selection method based on the notion of a critical time scale (CTS). The defined algorithm is optimal in the sense that it minimizes the required amount of link capacity for a traffic flow under a maximum delay constraint. Simulation results with compressed video traces demonstrate the efficiency of the defined traffic parameter selection algorithm in resource allocation.

  • Reducing Cache Energy Dissipation by Using Dual Voltage Supply

    Vasily G. MOSHNYAGA  Hiroshi TSUJI  

     
    PAPER-Optimization of Power and Timing

      Vol:
    E84-A No:11
      Page(s):
    2762-2768

    Due to a large capacitance and enormous access rate, caches dissipate about a third of the total energy consumed by today's processors. In this paper we present a new architectural technique to reduce energy consumption in caches. Unlike previous approaches, which have focused on lowering cache capacitance and the number of accesses, our method exploits a new freedom in cache design, namely the voltage per access. Since in modern block-buffered caches, the loading capacitance operated on block-hit is much less than the capacitance operated on miss, the given clock cycle time is inefficiently utilized during the hit. We propose to trade-off this unused time with the supply voltage, lowering the voltage level on the hit and increasing it on the miss. Experiments show that the approach can half the cache energy dissipation without large performance and area overhead.

  • Effective Reference Probability Incorporating the Effect of Expiration Time in Web Cache

    Jeong-Joon LEE  Kyu-Young WHANG  Yang-Sae MOON  Eui-Kyung HONG  

     
    PAPER-Databases

      Vol:
    E84-D No:9
      Page(s):
    1184-1197

    Web caching has become an important problem when addressing the performance issues in Web applications. The expiration time of the Web data item is useful a piece of information for performance enhancement in Web caching. In this paper, we introduce the notion of the effective reference probability that incorporates the effect of expiration time for Web caching. For a formal approach, we propose the continuous independent reference model extending the existing independent reference model. Based on this model, we define formally the effective reference probability and derive it theoretically. By simply replacing the reference probability in the existing cache replacement algorithms with the effective reference probability, we can take the effect of expiration time into account. The results of performance experiments show that the replacement algorithms using the effective reference probability always outperform existing ones. In particular, when the cache fraction is 0.05 and data update is comparatively frequent (i.e., the update frequency is more than 1/10 of the reference frequency), the performance is enhanced by more than 30% in LRU-2 and 13% in Aggarwal's method. The results show that the effective reference probability significantly enhances the performance of Web caching when the expiration time is given.

  • A Design Framework for Online Algorithms Solving the Object Replacement Problem

    Seiichiro TANI  Toshiaki MIYAZAKI  

     
    PAPER-Algorithms

      Vol:
    E84-D No:9
      Page(s):
    1135-1143

    Network caches reduce network traffic as well as user response time. When implementing network caches, the object replacement problem is one of the core problems; The problem is to determine which objects should be evicted from a cache when there is insufficient space. This paper first formalizes the problem and gives a simple but sufficient condition for deterministic online algorithms to be competitive. Based on the condition, a general framework to make a non-competitive algorithm competitive is constructed. As an application of the framework, an online algorithm, called Competitive_SIZE, is proposed. Both event-driven and trace-driven simulations show that Competitive_SIZE is better than previously proposed algorithms such as LRU (Least Recently Used).

  • Stochastic Model of Internet Access Patterns

    Masaki AIDA  Tetsuya ABE  

     
    PAPER-Traffic Measurement and Analysis

      Vol:
    E84-B No:8
      Page(s):
    2142-2150

    This paper investigates the stochastic property of the packet destinations and proposes an address generation algorithm which is applicable for describing various Internet access patterns. We assume that a stochastic process of Internet access satisfies the stationary condition and derive the fundamental structure of the address generation algorithm. Pseudo IP-address sequence generated from our algorithm gives dependable cache performance and reproduces the results obtained from trace-driven simulation. The proposed algorithm is applicable not only to the destination IP address but also to the destination URLs of packets, and is useful for simulation studies of Internet performance, Web caching, DNS, and so on.

201-220hit(297hit)