The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] data durability(2hit)

1-2hit
  • Data Recovery of Distributed Hash Table with Distributed-to-Distributed Data Copy

    Yusuke DOI  Shirou WAKAYAMA  Satoshi OZAKI  

     
    PAPER-Parallel and Distributed Architecture

      Vol:
    E92-D No:10
      Page(s):
    1999-2006

    To realize huge-scale information services, many Distributed Hash Table (DHT) based systems have been proposed. For example, there are some proposals to manage item-level product traceability information with DHTs. In such an application, each entry of a huge number of item-level IDs need to be available on a DHT. To ensure data availability, the soft-state approach has been employed in previous works. However, this does not scale well against the number of entries on a DHT. As we expect 1010 products in the traceability case, the soft-state approach is unacceptable. In this paper, we propose Distributed-to-Distributed Data Copy (D3C). With D3C, users can reconstruct the data as they detect data loss, or even migrate to another DHT system. We show why it scales well against the number of entries on a DHT. We have confirmed our approach with a prototype. Evaluation shows our approach fits well on a DHT with a low rate of failure and a huge number of data entries.

  • Reducing Replication Overhead for Data Durability in DHT Based P2P System

    Kyungbaek KIM  Daeyeon PARK  

     
    LETTER-Dependable Computing

      Vol:
    E90-D No:9
      Page(s):
    1452-1455

    DHT based p2p systems appear to provide scalable storage services with idle resource from many unreliable clients. If a DHT is used in storage intensive applications where data loss must be minimized, quick replication is especially important to replace lost redundancy on other nodes in reaction to failures. To achieve this easily, a simple replication method directly uses a consistent set, such as a leaf set and a successor list. However, this set is tightly coupled to the current state of nodes and the traffic needed to support this replication can be high and bursty under churn. This paper explores efficient replication methods that only glimpse a consistent set to select a new replica. Replicas are loosely coupled to a consistent set and we can eliminate the compulsory replication under churn. Because of a complication of the new replication methods, the careful data management is needed under churn for the correct and efficient data lookup. Results from a simulation study suggest that our methods can reduce network traffic enormously for high data durability.