The search functionality is under construction.

Keyword Search Result

[Keyword] reliability(282hit)

101-120hit(282hit)

  • Empirical-Statistics Analysis for Zero-Failure GaAs MMICs Life Testing Data

    Zheng-Liang HUANG  Fa-Xin YU  Shu-Ting ZHANG  Hao LUO  Ping-Hui WANG  Yao ZHENG  

     
    LETTER-Reliability, Maintainability and Safety Analysis

      Vol:
    E92-A No:9
      Page(s):
    2376-2379

    GaAs MMICs (Monolithic Microwave Integrated Circuits) reliability is a critical part of the overall reliability of the thermal solution in semiconductor devices. With MMICs reliability improved, GaAs MMICs failure rates will reach levels which are impractical to measure with conventional methods in the near future. This letter proposes a methodology to predict the GaAs MMICs reliability by combining empirical and statistical methods based on zero-failure GaAs MMICs life testing data. Besides, we investigate the effect of accelerated factors on MMICs degradation and make a comparison between the Weibull and lognormal distributions. The method has been used in the reliability evaluation of GaAs MMICs successfully.

  • Effects of Data Scrubbing on Reliability in Storage Systems

    Junkil RYU  Chanik PARK  

     
    PAPER-Computer Systems

      Vol:
    E92-D No:9
      Page(s):
    1639-1649

    Silent data corruptions, which are induced by latent sector errors, phantom writes, DMA parity errors and so on, can be detected by explicitly issuing a read command to a disk controller and comparing the corresponding data with their checksums. Because some of the data stored in a storage system may not be accessed for a long time, there is a high chance of silent data corruption occurring undetected, resulting in data loss. Therefore, periodic checking of the entire data in a storage system, known as data scrubbing, is essential to detect such silent data corruptions in time. The errors detected by data scrubbing will be recovered by the replica or the redundant information maintained to protect against permanent data loss. The longer the period between data scrubbings, the higher the probability of a permanent data loss. This paper proposes a Markov failure and repair model to conservatively analyze the effect of data scrubbing on the reliability of a storage system. We show the relationship between the period of a data scrubbing operation and the number of data replicas to manage the reliability of a storage system by using the proposed model.

  • Software Reliability Modeling Based on Capture-Recapture Sampling

    Hiroyuki OKAMURA  Tadashi DOHI  

     
    PAPER

      Vol:
    E92-A No:7
      Page(s):
    1615-1622

    This paper proposes a dynamic capture-recapture (DCR) model to estimate not only the total number of software faults but also quantitative software reliability from observed data. Compared to conventional static capture-recapture (SCR) model and usual software reliability models (SRMs) in the past literature, the DCR model can handle dynamic behavior of software fault-detection processes and can evaluate quantitative software reliability based on capture-recapture sampling of software fault data. This is regarded as a unified modeling framework of SCR and SRM with the Bayesian estimation. Simulation experiments under some plausible testing scenarios show that our models are superior to SCR and SRMs in terms of estimation accuracy.

  • Performability Modeling for Software System with Performance Degradation and Reliability Growth

    Koichi TOKUNO  Shigeru YAMADA  

     
    PAPER

      Vol:
    E92-A No:7
      Page(s):
    1563-1571

    In this paper, we discuss software performability evaluation considering the real-time property; this is defined as the attribute that the system can complete the task within the stipulated response time limit. We assume that the software system has two operational states from the viewpoint of the end users: one is operating with the desirable performance level according to specification and the other is with degraded performance level. The dynamic software reliability growth process with performance degradation is described by the extended Markovian software reliability model with imperfect debugging. Assuming that the software system can process the multiple tasks simultaneously and that the arrival process of the tasks follows a nonhomogeneous Poisson process, we analyze the distribution of the number of tasks whose processes can be completed within the processing time limit with the infinite server queueing model. We derive several software performability measures considering the real-time property; these are given as the functions of time and the number of debugging activities. Finally, we illustrate several numerical examples of the measures to investigate the impact of consideration of the performance degradation on the system performability evaluation.

  • Temperature-Aware NBTI Modeling Techniques in Digital Circuits

    Hong LUO  Yu WANG  Rong LUO  Huazhong YANG  Yuan XIE  

     
    PAPER-Integrated Electronics

      Vol:
    E92-C No:6
      Page(s):
    875-886

    Negative bias temperature instability (NBTI) has become a critical reliability phenomena in advanced CMOS technology. In this paper, we propose an analytical temperature-aware dynamic NBTI model, which can be used in two circuit operation cases: executing tasks with different temperatures, and switching between active and standby mode. A PMOS Vth degradation model and a digital circuits' temporal performance degradation estimation method are developed based on our NBTI model. The simulation results show that: 1) the execution of a low temperature task can decrease ΔVth due to NBTI by 24.5%; 2) switching to standby mode can decrease ΔVth by 52.3%; 3) for ISCAS85 benchmark circuits, the delay degradation can decrease significantly if the circuit execute low temperature task or switch to standby mode; 4) we have also observed the execution time's ratio of different tasks and the ratio of active to standby time both have a considerable impact on NBTI effect.

  • Reliability Inherent in Heterogeneous Multiprocessor Systems and Task Scheduling for Ameliorating Their Reliability

    Makoto SUGIHARA  

     
    PAPER

      Vol:
    E92-A No:4
      Page(s):
    1121-1128

    Utilizing a heterogeneous multiprocessor system has become a popular design paradigm to build an embedded system at a cheap cost within short development time. A reliability issue for embedded systems, which is vulnerability to single event upsets (SEUs), has become a matter of concern as technology proceeds. This paper discusses reliability inherent in heterogeneous multiprocessors and proposes task scheduling for minimizing SEU vulnerability of them. This paper experimentally shows that increasing performance of a CPU core deteriorates its reliability. Based on the experimental observation, we propose task scheduling for reducing SEU vulnerability of a heterogeneous multiprocessor system. The experimental results demonstrate that our task scheduling technique can reduce much of SEU vulnerability under real-time constraints.

  • User-Perceived Reliability of M-for-N (M:N) Shared Protection Systems

    Hirokazu OZAKI  Atsushi KARA  Zixue CHENG  

     
    PAPER-Dependable Computing

      Vol:
    E92-D No:3
      Page(s):
    443-450

    In this paper we investigate the reliability of general type shared protection systems i.e. M for N (M:N) that can typically be applied to various telecommunication network devices. We focus on the reliability that is perceived by an end user of one of N units. We assume that any failed unit is instantly replaced by one of the M units (if available). We describe the effectiveness of such a protection system in a quantitative manner. The mathematical analysis gives the closed-form solution of the availability, the recursive computing algorithm of the MTTFF (Mean Time to First Failure) and the MTTF (Mean Time to Failure) perceived by an arbitrary end user. We also show that, under a certain condition, the probability distribution of TTFF (Time to First Failure) can be approximated by a simple exponential distribution. The analysis provides useful information for the analysis and the design of not only the telecommunication network devices but also other general shared protection systems that are subject to service level agreements (SLA) involving user-perceived reliability measures.

  • CRRT: Congestion-Aware and Rate-Controlled Reliable Transport in Wireless Sensor Networks

    Muhammad Mahbub ALAM  Choong Seon HONG  

     
    PAPER-Network

      Vol:
    E92-B No:1
      Page(s):
    184-199

    For successful data collection in wireless sensor networks, it is important to ensure that the required delivery ratio is maintained while keeping a fair rate for every sensor. Furthermore, emerging high-rate applications might require complete reliability and the transfer of large volume of data, where persistent congestion might occur. These requirements demand a complete but efficient solution for data transport in sensor networks which reliably transports data from many sources to one or more sinks, avoids congestion and maintains fairness. In this paper, we propose congestion-aware and rate-controlled reliable transport (CRRT), an efficient and low-overhead data transport mechanism for sensor networks. CRRT uses efficient MAC retransmission to increase one-hop reliability and end-to-end retransmission for loss recovery. It also controls the total rate of the sources centrally, avoids the congestion in the bottleneck based on congestion notifications from intermediate nodes and centrally assigns the rate to the sources based on rate assignment policy of the applications. Performance of CRRT is evaluated in NS-2 and simulation results demonstrate the effectiveness of CRRT.

  • Design Methodology of a Sensor Network Architecture Supporting Urgent Information and Its Evaluation

    Tetsuya KAWAI  Naoki WAKAMIYA  Masayuki MURATA  

     
    PAPER-Network

      Vol:
    E91-B No:10
      Page(s):
    3232-3240

    Wireless sensor networks are expected to become an important social infrastructure which helps our life to be safe, secure, and comfortable. In this paper, we propose design methodology of an architecture for fast and reliable transmission of urgent information in wireless sensor networks. In this methodology, instead of establishing single complicated monolithic mechanism, several simple and fully-distributed control mechanisms which function in different spatial and temporal levels are incorporated on each node. These mechanisms work autonomously and independently responding to the surrounding situation. We also show an example of a network architecture designed following the methodology. We evaluated the performance of the architecture by extensive simulation and practical experiments and our claim was supported by the results of these experiments.

  • Formulas for Counting the Numbers of Connected Spanning Subgraphs with at Most n+1 Edges in a Complete Graph Kn

    Peng CHENG  Shigeru MASUYAMA  

     
    PAPER

      Vol:
    E91-A No:9
      Page(s):
    2314-2321

    Let Ni be the number of connected spanning subgraphs with i(n-1 i m) edges in an n-vertex m-edge undirected graph G=(V,E). Although Nn-1 is computed in polynomial time by the Matrix-tree theorem, whether Nn is efficiently computed for a graph G is an open problem (see e.g., [2]). On the other hand, whether Nn2≥ Nn-1Nn+1 for a graph G is also open as a part of log concave conjecture (see e.g., [6],[12]). In this paper, for a complete graph Kn, we give the formulas for Nn, Nn+1, by which Nn, Nn+1 are respectively computed in polynomial time on n, and, in particular, prove Nn2> Nn-1Nn+1 as well.

  • Highly Reliable Submicron InP-Based HBTs with over 300-GHz ft

    Norihide KASHIO  Kenji KURISHIMA  Yoshino K. FUKAI  Shoji YAMAHATA  

     
    PAPER-GaAs- and InP-Based Devices

      Vol:
    E91-C No:7
      Page(s):
    1084-1090

    We have developed 0.5-µm-emitter InP-based HBTs with high reliability. The HBTs incorporate a passivation ledge structure and tungsten-based emitter metal. A fabricated HBT exhibits high collector current density and a current gain of 58 at a collector current density of 4 mA/µm2. The results of dc measurements indicate that the ledge layer sufficiently suppresses the recombination current at the emitter-base periphery. The HBT also exhibits an ft of 321 GHz and an fmax of 301 GHz at a collector current density of 4 mA/µm2. The ft does not degrade even though the emitter size is reduced to as small as 0.5 µm2 µm. The results of an accelerated life test show that the degradation of dc current gain is due to thermal degradation of the interfacial quality of semiconductors at the passivation ledge. The activation energy is expected to be around 1.5 eV, and the extrapolated mean time to failure is expected to be over 108 hours at a junction temperature of 125. These results indicate that this InP HBT technology is promising for making over-100-Gbit/s ICs with high reliability.

  • Drouting Architecture: Improvement of Failure Avoidance Capability Using Multipath Routing

    Yasuhiro OHARA  Hiroyuki KUSUMOTO  Osamu NAKAMURA  Jun MURAI  

     
    PAPER-Network

      Vol:
    E91-B No:5
      Page(s):
    1403-1415

    Failure avoidance capability is a desired feature for telecommunication networks, such as the Internet. However, not all failures can be promptly bypassed on the Internet because routing systems that are responsible for detecting and avoiding failures cannot detect all failures. Consequently, failures can interrupt internet communications for a long time, such as a few hours. This paper proposes a novel routing architecture called Drouting that enables flexible failure avoidance. In Drouting, routers calculate multipaths from a source to a destination by constructing Directed Acyclic Graphs (DAGs) that include all links in the intra-domain network graph. IP packets carry packet tags that are set by the end host. The packet tags are used to select a network path from the multipath routes. In this paper, the failure avoidance property of Drouting architecture is evaluated through comparison with another proposal, Deflection, using simulations. Simulations were performed on inferred and synthetic topologies. Drouting exhibits similar performance with Deflection in terms of the number of nexthops, the number of paths and the length of paths, while Drouting outperforms Deflection in the probability of success of failure avoidance.

  • Reliable Cache Architectures and Task Scheduling for Multiprocessor Systems

    Makoto SUGIHARA  Tohru ISHIHARA  Kazuaki MURAKAMI  

     
    PAPER

      Vol:
    E91-C No:4
      Page(s):
    410-417

    This paper proposes a task scheduling approach for reliable cache architectures (RCAs) of multiprocessor systems. The RCAs dynamically switch their operation modes for reducing the usage of vulnerable SRAMs under real-time constraints. A mixed integer programming model has been built for minimizing vulnerability under real-time constraints. Experimental results have shown that our task scheduling approach achieved 47.7-99.9% less vulnerability than a conventional one.

  • Circuit Performance Degradation of Switched-Capacitor Circuit with Bootstrapped Technique due to Gate-Oxide Overstress in a 130-nm CMOS Process

    Jung-Sheng CHEN  Ming-Dou KER  

     
    PAPER-Electronic Circuits

      Vol:
    E91-C No:3
      Page(s):
    378-384

    The MOS switch with bootstrapped technique is widely used in low-voltage switched-capacitor circuit. The switched-capacitor circuit with the bootstrapped technique could be a dangerous design approach in the nano-scale CMOS process due to the gate-oxide transient overstress. The impact of gate-oxide transient overstress on MOS switch in switched-capacitor circuit is investigated in this work with the sample-and-hold amplifier (SHA) in a 130-nm CMOS process. After overstress on the MOS switch of SHA with unity-gain buffer, the circuit performances in time domain and frequency domain are measured to verify the impact of gate-oxide reliability on circuit performances. The oxide breakdown on switch device degrades the circuit performance of bootstrapped switch technique.

  • Inequalities on the Number of Connected Spanning Subgraphs in a Multigraph

    Peng CHENG  Shigeru MASUYAMA  

     
    PAPER-Graphs and Networks

      Vol:
    E91-D No:2
      Page(s):
    178-186

    Consider an undirected multigraph G=(V,E) with n vertices and m edges, and let Ni denote the number of connected spanning subgraphs with i(min) edges in G. Recently, we showed in [3] the validity of (m-i+1)Ni-1>Ni for a simple graph and each i(min). Note that, from this inequality, 2 is easily derived. In this paper, for a multigraph G and all i(min), we prove (m-i+1)Ni-1(i-n+2)Ni, and give a necessary and sufficient condition by which (m-i+1)Ni-1=(i-n+2)Ni. In particular, this means that (m-i+1)Ni-1>Ni is not valid for all multigraphs, in general. Furthermore, we prove 2, which is not straightforwardly derived from (m-i+1)Ni-1(i-n+2)Ni, and also introduce a necessary and sufficent condition by which =2. Moreover, we show a sufficient condition for a multigraph to have Nn2>Nn-1Nn+1. As special cases of the sufficient condition, we show that if G contains at least +1 multiple edges between some pair of vertices, or if its underlying simple graph has no cycle with length more than 4, then Nn2>Nn-1Nn+1.

  • RK-Means Clustering: K-Means with Reliability

    Chunsheng HUA  Qian CHEN  Haiyuan WU  Toshikazu WADA  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E91-D No:1
      Page(s):
    96-104

    This paper presents an RK-means clustering algorithm which is developed for reliable data grouping by introducing a new reliability evaluation to the K-means clustering algorithm. The conventional K-means clustering algorithm has two shortfalls: 1) the clustering result will become unreliable if the assumed number of the clusters is incorrect; 2) during the update of a cluster center, all the data points belong to that cluster are used equally without considering how distant they are to the cluster center. In this paper, we introduce a new reliability evaluation to K-means clustering algorithm by considering the triangular relationship among each data point and its two nearest cluster centers. We applied the proposed algorithm to track objects in video sequence and confirmed its effectiveness and advantages.

  • An Optimal Parallel Algorithm for Finding All Hinge Vertices of a Circular-Arc Graph

    Hirotoshi HONMA  Shigeru MASUYAMA  

     
    PAPER-Algorithms and Data Structures

      Vol:
    E91-A No:1
      Page(s):
    383-391

    Let G =(V, E) be an undirected simple graph with u ∈ V. If there exist any two vertices in G whose distance becomes longer when a vertex u is removed, then u is defined as a hinge vertex. Finding the set of hinge vertices in a graph is useful for identifying critical nodes in an actual network. A number of studies concerning hinge vertices have been made in recent years. In a number of graph problems, it is known that more efficient sequential or parallel algorithms can be developed by restricting classes of graphs. In this paper, we shall propose a parallel algorithm which runs in O(log n) time with O(n/log n) processors on EREW PRAM for finding all hinge vertices of a circular-arc graph.

  • Discrete Program-Size Dependent Software Reliability Assessment: Modeling, Estimation, and Goodness-of-Fit Comparisons

    Shinji INOUE  Shigeru YAMADA  

     
    PAPER-Reliability, Maintainability and Safety Analysis

      Vol:
    E90-A No:12
      Page(s):
    2891-2902

    In this paper we propose a discrete program-size dependent software reliability growth model flexibly describing the software failure-occurrence phenomenon based on a discrete Weibull distribution. We also conduct model comparisons of our discrete SRGM with existing discrete SRGMs by using actual data sets. The program size is one of the important metrics of software complexity. It is known that flexible discrete software reliability growth modeling is difficult due to the mathematical manipulation under a conventional modeling-framework in which the time-dependent behavior of the cumulative number of detected faults is formulated by a difference equation. Our discrete SRGM is developed under an existing unified modeling-framework based on the concept of general order-statistics, and can incorporate the effect of the program size into software reliability assessment. Further, we discuss the method of parameter estimation, and derive software reliability assessment measures of our discrete SRGM. Finally, we show numerical examples of discrete software reliability analysis based on our discrete SRGM by using actual data.

  • Proposal of an Assured Corridor Mechanism for Urgent Information Transmission in Wireless Sensor Networks

    Tetsuya KAWAI  Naoki WAKAMIYA  Masayuki MURATA  

     
    PAPER-Network

      Vol:
    E90-B No:10
      Page(s):
    2817-2826

    Wireless sensor networks are expected to play an essential role as a social infrastructure to realize our safe and secure living environment. In such a network, critical information must be transmitted faster and more reliably than other information. We propose a distributed transmission mechanism which enables emergency packets to be carried with high reliability and low latency along a preferential path, which is called an "assured corridor." In this self-organizing assured corridor mechanism (ACM), which works above the network layer and does not depend on any specific routing or MAC protocol, a corridor is gradually established as the first packet containing urgent information propagates to the base station. The nodes surrounding the corridor suppress the transmission of non-urgent information and nodes in the corridor are kept awake to forward emergency packets. ACM avoids packet loss and possible delay caused by collisions in the wireless transmission and normal sleep scheduling. An acknowledgment and retransmission scheme is incorporated into ACM in order to improve reliability of transmission of urgent information. Simulation experiments showed that, when only one node transmitted urgent information, the retransmission contributed to establish a corridor quickly and that ACM improved the delivery ratio and the delay of the urgent information transmission once a corridor is established. It was proved that ACM was effective to improve the reliability and the latency of urgent information as well in the cases where multiple nodes sent urgent information at once.

  • Fuzzy c-Means Algorithms for Data with Tolerance Based on Opposite Criterions

    Yuchi KANZAWA  Yasunori ENDO  Sadaaki MIYAMOTO  

     
    PAPER-Soft Computing

      Vol:
    E90-A No:10
      Page(s):
    2194-2202

    In this paper, two new clustering algorithms are proposed for the data with some errors. In any of these algorithms, the error is interpreted as one of decision variables -- called "tolerance" -- of a certain optimization problem like the previously proposed algorithm, but the tolerance is determined based on the opposite criterion to its corresponding previously proposed one. Applying our each algorithm together with its corresponding previously proposed one, a reliability of the clustering result is discussed. Through some numerical experiments, the validity of this paper is discussed.

101-120hit(282hit)