The search functionality is under construction.

Author Search Result

[Author] Hiroki NAKAYAMA(5hit)

1-5hit
  • A Flow Aggregation Method under Allowable Delay Limitation in SDN

    Takuya KOSUGIYAMA  Kazuki TANABE  Hiroki NAKAYAMA  Tsunemasa HAYASHI  Katsunori YAMAOKA  

     
    PAPER-Network

      Pubricized:
    2017/09/14
      Vol:
    E101-B No:3
      Page(s):
    795-804

    Software-Defined Networking (SDN) can be applied for managing application flows dynamically by a logically centralized SDN controller and SDN switches. Because one SDN switch can support just a few thousand forwarding rule installations per second, it is a barrier to dynamic and scalable application flow management. For this reason, it is essential to reduce the number of application flows if they are to be successfully managed. Nowadays, since much attention has been paid to developing a network service that reduces application delay, the allowable delay of application flows has become an important factor. However, there has been no work on minimizing the number of flows while satisfying end-to-end delay of flows. In this paper, we propose a method that can aggregate flows and minimize the number flows in a network while ensuring all flows satisfy their allowable delay in accordance with QoS or SLA. Since the problem is classified as NP-hard, we propose a heuristic algorithm. We compared the aggregation effect of the proposed method, simple aggregation method and optimal solution by simulation. In addition, we clarify the characteristics of the proposed method by performing simulations with various parameter settings. The results show that the proposed method decreases the number of rules than comparative aggregation method and has very shorter computational time than optimal solution.

  • vEPC Optimal Resource Assignment Method for Accommodating M2M Communications

    Kazuki TANABE  Hiroki NAKAYAMA  Tsunemasa HAYASHI  Katsunori YAMAOKA  

     
    PAPER

      Pubricized:
    2017/09/19
      Vol:
    E101-B No:3
      Page(s):
    637-647

    The 5G mobile network environment has been studied and developed, and the concept of a vEPC (Virtualized Evolved Packet Core) has been introduced as a framework for Network Functions Virtualization (NFV). Machine-to-Machine (M2M) communications in 5G networks require much faster response than are possible in 4G networks. However, if both the control plane (C-plane) and the data plane (D-plane) functions of the EPC are migrated into a single vEPC server, M2M devices and other user equipments (UEs) share the same resources. To accommodate delay-sensitive M2M sessions in vEPC networks, not only signaling performance on the C-plane but also packet processing performance on the D-plane must be optimized. In this paper, we propose a method for optimizing resource assignment of C-plane and D-plane Virtualized Network Functions (VNFs) in a vEPC server, called the vEPC-ORA method. We distinguish the communications of M2M devices and smartphones and model the vEPC server by using queueing theory. Numerical analysis of optimal resource assignment shows that our proposed method minimizes the blocking rates of M2M sessions and smartphone sessions. We also confirmed that the mean packet processing time is kept within the allowable delay for each communication type, as long as the vEPC server has enough VM resources. Moreover, we study a resource granularity effect on the optimal resource assignment. Numerical analysis under a fixed number of hardware resources of MME and S/P-GW is done for various resource granularities of the vEPC server. The evaluation results of numerical analyses showed that the vEPC-ORA method derives the optimal resource assignment in practical calculation times.

  • A Software Approach of Controlling the CPU Resource Assignment in Network Virtualization

    Shin MURAMATSU  Ryota KAWASHIMA  Shoichi SAITO  Hiroshi MATSUO  Hiroki NAKAYAMA  Tsunemasa HAYASHI  

     
    PAPER

      Vol:
    E98-B No:11
      Page(s):
    2171-2179

    Many public cloud datacenters have adopted the Edge-Overlay model which supports virtual switch-based network virtualization using IP tunneling. However, software-implemented virtual switches can cause performance degradation because the packet processing load can concentrate on a particular CPU core. As a result, such load concentration decreases and destabilizes the performance of virtual networks. Although multi-queue functions like Receive Side Scaling (RSS) can distribute the load onto multiple CPU cores, they still have performance problems such as IRQ core collision between priority flows as well as competitive resource use between host and guest machines for received packet processing. In this paper, we propose Virtual Switch Extension (VSE) that adaptively determines CPU core assignment for SoftIRQ to prevent performance degradation. VSE supports two types of SoftIRQ core selection mechanisms, on-the-fly or predetermined. In the on-the-fly mode, VSE selects a SoftIRQ core based on current CPU load to exploit low-loaded CPU resources. In the predetermined mode, SoftIRQ cores are assigned in advance to differentiate the performance of priority flows. This paper describes a basic architecture and implementation of VSE and how VSE assigns a SoftIRQ cores. Moreover, we evaluate fundamental throughput of various CPU assignment models in the predetermined mode. Finally, we evaluate the performance of a priority VM in two VM usecases, the client-usecase which is receive-oriented and the router-usecase which performs bi-directional communications. In the client-usecase, the throughput of the priority VM was improved by 31% compared with RSS when the priority VM had one dedicated core. In the router-usecase, the throughput was improved by 29% when three dedicated cores were provided for the VM.

  • Fast Datapath Processing Based on Hop-by-Hop Packet Aggregation for Service Function Chaining Open Access

    Yuki TAGUCHI  Ryota KAWASHIMA  Hiroki NAKAYAMA  Tsunemasa HAYASHI  Hiroshi MATSUO  

     
    PAPER-Information Network

      Pubricized:
    2019/08/22
      Vol:
    E102-D No:11
      Page(s):
    2184-2194

    Many studies have revealed that the performance of software-based Virtual Network Functions (VNFs) is insufficient for mission-critical networks. Scaling-out approaches, such as auto-scaling of VNFs, could handle a huge amount of traffic; however, the exponential traffic growth confronts us the limitations of both expandability of physical resources and complexity of their management. In this paper, we propose a fast datapath processing method called Packet Aggregation Flow (PA-Flow) that is based on hop-by-hop packet aggregation for more efficient Service Function Chaining (SFC). PA-Flow extends a notion of existing intra-node packet aggregation toward network-wide packet aggregation, and we introduce following three novel features. First, packet I/O overheads at intermediate network devices including NFV-nodes are mitigated by reduction of packet amount. Second, aggregated packets are further aggregated as going through the service chain in a hop-by-hop manner. Finally, next-hop aware packet aggregation is realized using OpenFlow-based flow tables. PA-Flow is designed to be available with various VNF forms (e.g. VM/container/baremetal-based) and virtual I/O technologies (e.g. vhost-user/SR-IOV), and its implementation does not bring noticeable delay for aggregation. We conducted two evaluations: (i) a baseline evaluation for understanding fundamental performance characteristics of PA-Flow (ii) a simulation-based SFC evaluation for proving PA-Flow's effect in a realistic environment. The results showed that throughput of short packet forwarding was improved by 4 times. Moreover, the total number of packets was reduced by 93% in a large-scale SFC.

  • Public WLAN Virtualization for Multiple Services

    Kazuhiko KINOSHITA  Kazuki GINNAN  Keita KAWANO  Hiroki NAKAYAMA  Tsunemasa HAYASHI  Takashi WATANABE  

     
    PAPER-Network

      Pubricized:
    2018/10/10
      Vol:
    E102-B No:4
      Page(s):
    832-844

    The recent widespread use of high-performance terminals has resulted in a rapid increase in mobile data traffic. Therefore, public wireless local area networks (WLANs) are being used often to supplement the cellular networks. Capacity improvement through the dense deployment of access points (APs) is being considered. However, the effective throughput degrades significantly when many users connect to a single AP. In this paper, users are classified into guaranteed bit rate (GBR) users and best effort (BE) users, and we propose a network model to provide those services. In the proposed model, physical APs and the bandwidths are assigned to each service class dynamically using a virtual AP configuration and a virtualized backhaul network, for reducing the call-blocking probability of GBR users and improving the satisfaction degree of BE users. Finally, we evaluate the performance of the proposed model through simulation experiments and discuss its feasibility.