The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] software(508hit)

121-140hit(508hit)

  • Software Reliability Assessment with Multiple Changes of Testing-Environment

    Shinji INOUE  Shigeru YAMADA  

     
    PAPER

      Vol:
    E98-A No:10
      Page(s):
    2031-2041

    We discuss software reliability assessment considering multiple changes of software fault-detection phenomenon. The testing-time when the characteristic of the software failure-occurrence or fault-detection phenomenon changes notably in the testing-phase of a software development process is called change-point. It is known that the occurrence of the change-point influences the accuracy for the software reliability assessment based on a software reliability growth models, which are mainly divided into software failure-occurrence time and fault counting models. This paper discusses software reliability growth modeling frameworks considering with the effect of the multiple change-point occurrence on the software reliability growth process in software failure-occurrence time and fault counting modeling. And we show numerical illustrations for the software reliability analyses based on our models by using actual data.

  • One-Step Error Detection and Correction Approach for Voice Word Processor

    Junhwi CHOI  Seonghan RYU  Kyusong LEE  Gary Geunbae LEE  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2015/05/20
      Vol:
    E98-D No:8
      Page(s):
    1517-1525

    We propose a one-step error detection and correction interface for a voice word processor. This correction interface performs analysis region detection, user intention understanding and error correction utterance recognition, all from a single user utterance input. We evaluate the performance of each component first, and then compare the effectiveness of our interface to two previous interfaces. Our evaluation demonstrates that each component is technically superior to the baselines and that our one-step error detection and correction method yields an error correction interface that is more convenient and natural than the two previous interfaces.

  • Effect Analysis of Coding Convention Violations on Readability of Post-Delivered Code

    Taek LEE  Jung-Been LEE  Hoh Peter IN  

     
    PAPER-Software Engineering

      Pubricized:
    2015/04/10
      Vol:
    E98-D No:7
      Page(s):
    1286-1296

    Adherence to coding conventions during the code production stage of software development is essential. Benefits include enabling programmers to quickly understand the context of shared code, communicate with one another in a consistent manner, and easily maintain the source code at low costs. In reality, however, programmers tend to doubt or ignore the degree to which the quality of their code is affected by adherence to these guidelines. This paper addresses research questions such as “Do violations of coding conventions affect the readability of the produced code?”, “What kinds of coding violations reduce code readability?”, and “How much do variable factors such as developer experience, project size, team size, and project maturity influence coding violations?” To respond to these research questions, we explored 210 open-source Java projects with 117 coding conventions from the Sun standard checklist. We believe our findings and the analysis approach used in the paper will encourage programmers and QA managers to develop their own customized and effective coding style guidelines.

  • Electromagnetic Analysis against Public-Key Cryptographic Software on Embedded OS

    Hajime UNO  Sho ENDO  Naofumi HOMMA  Yu-ichi HAYASHI  Takafumi AOKI  

     
    PAPER

      Vol:
    E98-B No:7
      Page(s):
    1242-1249

    Electromagnetic analysis (EMA) against public-key cryptographic software on an embedded OS is presented in this paper. First, we propose a method for finding an observation point for EMA, where the EM radiation caused by cryptographic operations can be observed with low noise. The basic idea is to find specific EM radiation patterns produced by cryptographic operations given specific input pattern. During the operations, we scan the surface of the target device(s) with a micro magnetic probe. The scan is optimized in advanced using another compatible device that has the same central processing unit (CPU) and OS as the target device. We demonstrate the validity of the proposed EMAs through some EMA experiments with two types of RSA software on an embedded OS platform. The two types of RSA software have different implementations for modular multiplication algorithms: one is a typical and ready-made implementation using BigInteger class on Java standard library, and another is a custom-made implementation based on the Montgomery multiplication algorithm. We conduct experiments of chosen-message EMA using our scanning method, and show such EMAs successfully reveal the secret key of RSA software even under the noisy condition of the embedded OS platform. We also discuss some countermeasures against the above EMAs.

  • Software Maintenance Evaluation of Agile Software Development Method Based on OpenStack

    Yoji YAMATO  Shinichiro KATSURAGI  Shinji NAGAO  Norihiro MIURA  

     
    LETTER-Software Engineering

      Pubricized:
    2015/04/20
      Vol:
    E98-D No:7
      Page(s):
    1377-1380

    We evaluated software maintenance of an open source cloud platform system we developed using an agile software development method. We previously reported on a rapid service launch using the agile software development method in spite of large-scale development. For this study, we analyzed inquiries and the defect removal efficiency of our recently developed software throughout one-year operation. We found that the defect removal efficiency of our recently developed software was 98%. This indicates that we could achieve sufficient quality in spite of large-scale agile development. In term of maintenance process, we could answer all enquiries within three business days and could conduct version-upgrade fast. Thus, we conclude that software maintenance of agile software development is not ineffective.

  • 1/2fs Direct RF Under Sampling Receiver for Multi Channel Satellite Systems

    Daliso BANDA  Mizuki MOTOYOSHI  Tomokazu KOIZUMI  Osamu WADA  Tuan Thanh TA  Suguru KAMEDA  Noriharu SUEMATSU  Tadashi TAKAGI  Kazuo TSUBOUCHI  

     
    PAPER-Active Circuits/Devices/Monolithic Microwave Integrated Circuits

      Vol:
    E98-C No:7
      Page(s):
    669-676

    RF under sampling is more suitable for Satellite receiver systems in comparison to terrestrial systems. In conventional RF under sampling the minimum sampling frequency (fs) should be atleast twice the system bandwidth; therefore for a system with a wide bandwidth, a relatively high fs is necessary. In this paper we propose a direct RF under sampling reception method that halves fs. The proposed f's is achieved by folding in band noise in half. A method of adapting f's for the reception of signals in different channels is also proposed; this ensures that the SNR is not degraded for any channel. To evaluate the proposed technique's performance and compare it to the conventional case a 3 channel, 1 GHz band test receiver and it's key device (i.e. S/H circuit) are developed. Using SNR and EVM as performance indexes, the performance of the proposed technique has been evaluated and compared to that of the conventional technique. The evaluation results show that the proposed technique can achieve the same performance as conventional RF under sampling for all 3 channels, using only half of the sampling frequency of the conventional technique.

  • Approximating the Evolution History of Software from Source Code

    Tetsuya KANDA  Takashi ISHIO  Katsuro INOUE  

     
    PAPER-Software Engineering

      Pubricized:
    2015/03/17
      Vol:
    E98-D No:6
      Page(s):
    1185-1193

    Once a software product has been released, a large number of software products may be derived from an original single product. Management and maintenance of product variants are important, but those are hardly cared because developers do not make efforts for the further maintainability in the initial phase of software development. However, history of products would be lost in typical cases and developers have only source code of products in the worst case. In this paper, we approximate the evolution history of software products using source code of them. Our key idea is that two successive products are the most similar pair of products in evolution history, and have many similar source files. We did an experiment to compare the analysis result with actual evolution history. The result shows 78% (on average) of edges in the extracted trees are consistent with the actual evolution history of the products.

  • A New Approach to Embedded Software Optimization Based on Reverse Engineering

    Nguyen Ngoc BINH  Pham Van HUONG  Bui Ngoc HAI  

     
    PAPER-Computer System

      Pubricized:
    2015/03/17
      Vol:
    E98-D No:6
      Page(s):
    1166-1175

    Optimizing embedded software is a problem having scientific and practical signification. Optimizing embedded software can be done in different phases of the software life cycle under different optimal conditions. Most studies of embedded software optimization are done in forward engineering and these studies have not given an overall model for the optimization problem of embedded software in both forward engineering and reverse engineering. Therefore, in this paper, we propose a new approach to embedded software optimization based on reverse engineering. First, we construct an overall model for the embedded software optimization in both forward engineering and reverse engineering and present a process of embedded software optimization in reverse engineering. The main idea of this approach is that decompiling executable code to source code, converting the source code to models and optimizing embedded software under different levels such as source code and model. Then, the optimal source code is recompiled. To develop this approach, we present two optimization techniques such as optimizing power consumption of assembly programs based on instruction schedule and optimizing performance based on alternating equivalent expressions.

  • A Novel Algorithm for Burst Detection in Wideband Networking Waveform of Software Defined Radio

    Muhammad ZEESHAN  Shoab KHAN  

     
    PAPER-Digital Signal Processing

      Vol:
    E98-A No:6
      Page(s):
    1225-1233

    The correct detection of the start of burst is very important in wideband networking radio operation as it directly affects the Time Division Multiple Access (TDMA) adaptive time slot algorithm. In this paper, we propose a robust Data Aided (DA) algorithm for burst detection in a hybrid CDMA/Adaptive TDMA based wideband networking waveform of a software defined radio. The proposed algorithm is based on a novel differentially modulated training sequence designed by using precoding sequence. The training sequence structure and precoding sequence are exploited in the calculation of proposed timing metric which is normalized by the signal energy. The precoding sequence is adequately designed for the timing metric to have a sharp peak. The algorithm shows excellent performance for multiuser scenario. It is shown through computer simulations that by increasing the active users from 1 to 8, the performance degradation is only about 1∼2dB. The proposed algorithm is compared with other algorithms and found to outperform them even in the presence of multipath fading effects. The proposed algorithm has been implemented on Field Programmable Gate Array (FPGA) platform for high data rate applications and it is shown that the results from hardware are identical to the simulation results.

  • Recovering Traceability Links between Requirements and Source Code Using the Configuration Management Log

    Ryosuke TSUCHIYA  Hironori WASHIZAKI  Yoshiaki FUKAZAWA  Tadahisa KATO  Masumi KAWAKAMI  Kentaro YOSHIMURA  

     
    PAPER-Software Engineering

      Pubricized:
    2015/01/06
      Vol:
    E98-D No:4
      Page(s):
    852-862

    Traceability links between requirements and source code are helpful in software reuse and maintenance tasks. However, manually recovering links in a large group of products requires significant costs and some links may be overlooked. Here, we propose a semi-automatic method to recover traceability links between requirements and source code in the same series of large software products. In order to support differences in representation between requirements and source code, we recover links by using the configuration management log as an intermediary. We refine the links by classifying requirements and code elements in terms of whether they are common to multiple products or specific to one. As a result of applying our method to real products that have 60KLOC, we have recovered valid traceability links within a reasonable amount of time. Automatic parts have taken 13 minutes 36 seconds, and non-automatic parts have taken about 3 hours, with a recall of 76.2% and a precision of 94.1%. Moreover, we recovered some links that were unknown to engineers. By recovering traceability links, software reusability and maintainability will be improved.

  • An Adaptation of Proxy Mobile IPv6 to OpenFlow Architecture over Software Defined Networking

    Seong-Mun KIM  Hyon-Young CHOI  Youn-Hee HAN  Sung-Gi MIN  

     
    PAPER-Network

      Vol:
    E98-B No:4
      Page(s):
    596-606

    In this paper, Proxy Mobile IPv6 (PMIPv6), which is a network-based mobility management protocol, is adapted to the OpenFlow architecture. Mobility-related signaling is generally performed by network entities on behalf of a mobile node, but in standard PMIPv6, the control and data packets are delivered and processed over the same network entities, which prevents the separation of the control and the data planes. In addition, IP tunneling inherent to PMIPv6 imposes excessive overhead for the network entities. In order to adapt PMIPv6 to the OpenFlow architecture, the mobility management function is separated from the PMIPv6 components, and components are reconstructed to take advantage of the offerings of the OpenFlow architecture. The components configure the flow table of the switches located in a path, which comprise the OpenFlow controller. Mobility-related signaling can then be performed at the dedicated secure channel, and all of the data packets can be sent normally in accordance with the flow table of the OpenFlow switches. Consequently, the proposed scheme eliminates IP tunneling when user traffic is forwarded and separates the data and the control planes. The performance analysis revealed that the proposed scheme can outperform PMIPv6 in terms of the signaling cost, packet delivery cost, and handover latency.

  • An Original Entry Point Detection Method with Candidate-Sorting for More Effective Generic Unpacking

    Ryoichi ISAWA  Daisuke INOUE  Koji NAKAO  

     
    PAPER-Information Network

      Pubricized:
    2015/01/06
      Vol:
    E98-D No:4
      Page(s):
    883-893

    Many malware programs emerging from the Internet are compressed and/or encrypted by a wide variety of packers to deter code analysis, thus making it necessary to perform unpacking first. To do this task efficiently, Guo et al. proposed a generic unpacking system named Justin that provides original entry point (OEP) candidates. Justin executes a packed program, and then it extracts written-and-executed points caused by the decryption of the original binary until it determines the OEP has appeared, taking those points as candidates. However, for several types of packers, the system can provide comparatively large sets of candidates or fail to capture the OEP. For more effective generic unpacking, this paper presents a novel OEP detection method featuring two mechanisms. One identifies the decrypting routine by tracking relations between writing instructions and written areas. This is based on the fact that the decrypting routine is the generator for the original binary. In case our method fails to detect the OEP, the other mechanism sorts candidates based on the most likely candidate so that analysts can reach the correct one quickly. With experiments using a dataset of 753 samples packed by 25 packers, we confirm that our method can be more effective than Justin's heuristics, in terms of detecting OEPs and reducing candidates. After that, we also propose a method combining our method with one of Justin's heuristics.

  • Method Verb Recommendation Using Association Rule Mining in a Set of Existing Projects

    Yuki KASHIWABARA  Takashi ISHIO  Hideaki HATA  Katsuro INOUE  

     
    PAPER-Software Engineering

      Pubricized:
    2014/12/16
      Vol:
    E98-D No:3
      Page(s):
    627-636

    It is well-known that program readability is important for maintenance tasks. Method names are important identifiers for program readability because they are used for understanding the behavior of methods without reading a part of the program. Although developers can create a method name by arbitrarily choosing a verb and objects, the names are expected to represent the behavior consistently. However, it is not easy for developers to choose verbs and objects consistently since each developer may have a different notion of a suitable lexicon for method names. In this paper, we propose a technique to recommend candidate verbs for a method name so that developers can use various verbs consistently. We recommend candidate verbs likely to be used as a part of a method name, using association rules extracted from existing methods. To evaluate our technique, we have extracted rules from 445 open source projects written in Java and confirmed the accuracy of our approach by applying the extracted rules to several open source applications. As a result, we found that 84.9% of the considered methods in four projects are recommended the existing verb. Moreover, we found that 73.2% of the actual renamed methods in six projects are recommended the correct verb.

  • A Scenario-Based Reliability Analysis Approach for Component-Based Software

    Chunyan HOU  Chen CHEN  Jinsong WANG  Kai SHI  

     
    PAPER-Software Engineering

      Pubricized:
    2014/12/04
      Vol:
    E98-D No:3
      Page(s):
    617-626

    With the rise of component-based software development, its reliability has attracted much attention from both academic and industry communities. Component-based software development focuses on architecture design, and thus it is important for reliability analysis to emphasize software architecture. Existing approaches to architecture-based software reliability analysis don't model the usage profile explicitly, and they ignore the difference between the testing profile and the practical profile of components, which limits their applicability and accuracy. In response to these issues, a new reliability modeling and prediction approach is introduced. The approach considers reliability-related architecture factors by explicitly modeling the system usage profile, and transforms the testing profile into the practical usage profile of components by representing the profile with input sub-domains. Finally, the evaluation experiment shows the potential of the approach.

  • Comparative Analysis of Modularity Metrics for Evaluating Evolutionary Software

    Ki-Seong LEE  Chan-Gun LEE  

     
    LETTER-Software Engineering

      Vol:
    E98-D No:2
      Page(s):
    439-443

    Modularity is an effective evaluation approach for understanding the structural quality of evolutionary software. However, there are many diverse ways to measure it. In this paper, we analyze and compare various modularity metrics that have been studied in different domains to assess their applicability to evolutionary software analysis. Through extensive experiments with artificial DSMs and open-source software, we find that the correlations of those metrics are generally high despite their differences. However, our experiments show that a certain metric can be more sensitive to particular modular factors, hence applying of comprehensive modularity metrics must be taken into consideration.

  • Low-Power Loop Parallelization onto CGRA Utilizing Variable Dual VDD

    Bing XU  Shouyi YIN  Leibo LIU  Shaojun WEI  

     
    PAPER-Architecture

      Pubricized:
    2014/11/19
      Vol:
    E98-D No:2
      Page(s):
    243-251

    Coarse Grained Reconfigurable Architectures (CGRAs) are promising platform based on its high-performance and low cost. Researchers have developed efficient compilers for mapping compute-intensive applications on CGRA using modulo scheduling. In order to generate loop kernel, every stage of kernel are forced to have the same execution time which is determined by the critical PE. Hence non-critical PEs can decrease the supply voltage according to its slack time. The variable Dual-VDD CGRA incorporates this feature to reduce power consumption. Previous work mainly focuses on calculating a global optimal VDDL using overall optimization method that does not fully exploit the flexibility of architecture. In this brief, we adopt variable optimal VDDL in each stage of kernel concerning their pattern respectively instead of the fixed simulated global optimal VDDL. Experiment shows our proposed heuristic approach could reduce the power by 27.6% on average without decreasing performance. The compilation time is also acceptable.

  • Software-Hardware-Cooperative Protocol Processor for Extendable 10G-EPON MAC Chip

    Naoki MIURA  Akihiko MIYAZAKI  Junichi KATO  Nobuyuki TANAKA  Satoshi SHIGEMATSU  Masami URANO  Mamoru NAKANISHI  Tsugumichi SHIBATA  

     
    PAPER-Electronic Circuits

      Vol:
    E98-C No:1
      Page(s):
    45-52

    A 10-gigabit Ethernet passive optical network (10G-EPON) is promising for the next generation of access networks. A protocol processor for 10G-EPON needs to not only achieve 10-Gbps throughput but also to have protocol extendibility for various potential services. However, the conventional protocol processor does not have the ability to install additional protocols after chip fabrication, due to its hardware-based architecture. This paper presents a software-hardware cooperative protocol processor for 10G-EPON that provides the protocol extendibility. To achieve the software-hardware cooperation, the protocol processor newly employs a software-hardware partitioning technique driven by the timing requirements of 10G-EPON and a software-hardware interface circuit with event FIFO to absorb performance difference between software and hardware. The fabricated chip with this protocol processor properly works cooperatively and is able to accept newly standardized protocols. This protocol processor enables network operators to install additional service protocols adaptively for their own services.

  • Software-Defined Data Plane Enhancing SDN and NFV Open Access

    Akihiro NAKAO  

     
    INVITED PAPER

      Vol:
    E98-B No:1
      Page(s):
    12-19

    In this paper, we posit that extension of SDN to support deeply and flexibly programmable, software-defined data plane significantly enhance SDN and NFV and their interaction in terms of (1) enhanced interaction between applications and networks, (2) optimization of network functions, and (3) rapid development of new network protocols. All of these benefits are expected to contribute to improving the quality of diversifying communication networks and services. We identify three major technical challenges for enabling software-defined data plane as (1) ease of programming, (2) reasonable and predictable performance and (3) isolation among multiple concurrent logics. We also promote application-driving thinking towards defining software defined data-plane. We briefly introduce our project FLARE and its related technologies and review four use cases of flexible and deeply programmable data plane.

  • An Adaptive Multiple-Fault Injection Attack on Microcontrollers and a Countermeasure

    Sho ENDO  Naofumi HOMMA  Yu-ichi HAYASHI  Junko TAKAHASHI  Hitoshi FUJI  Takafumi AOKI  

     
    PAPER-Foundation

      Vol:
    E98-A No:1
      Page(s):
    171-181

    This paper proposes a multiple-fault injection attack based on adaptive control of fault injection timing in embedded microcontrollers. The proposed method can be conducted under the black-box condition that the detailed cryptographic software running on the target device is not known to attackers. In addition, the proposed method is non-invasive, without the depackaging required in previous works, since such adaptive fault injection is performed by precisely generating a clock glitch. We first describe the proposed method which injects two kinds of faults to obtain a faulty output available for differential fault analysis while avoiding a conditional branch in a typical recalculation-based countermeasure. We then show that the faulty output can be obtained by the proposed method without using information from the detailed instruction sequence. In particular, the validity of the proposed method is demonstrated through experiments on Advanced Encryption Standard (AES) software with a recalculation-based countermeasure on 8-bit and 32-bit microcontrollers. We also present a countermeasure resistant to the proposed method.

  • Evaluation Method for Access-Driven Cache Attacks Using Correlation Coefficient

    Junko TAKAHASHI  Toshinori FUKUNAGA  Kazumaro AOKI  Hitoshi FUJI  

     
    PAPER-Foundation

      Vol:
    E98-A No:1
      Page(s):
    192-202

    This paper proposes a new accurate evaluation method for examining the resistance of cryptographic implementations against access-driven cache attacks (CAs). We show that a mathematical correlation method between the sets of measured access time and the ideal data, which depend on the guessed key, can be utilized to evaluate quantitatively the correct key in access-driven CAs. We show the effectiveness of the proposed method using the access time measured in noisy environments. We also estimate the number of key candidates based on mathematical proof while considering memory allocation. Furthermore, based on the proposed method, we analyze quantitatively how the correlation values change with the number of plaintexts for a successful attack.

121-140hit(508hit)