1-12hit |
Hideyuki SHIMONISHI Shuji ISHII Lei SUN Yoshihiko KANAUMI
We propose a flexible and scalable architecture for a network controller platform used for OpenFlow. The OpenFlow technology was proposed as a means for researchers, network service creators, and others to easily design, test, and virtually deploy their innovative ideas in a large network infrastructure, which will accelerate research activities on Future Internet architectures. The technology enables the independent evolution of the network control plane and the data plane. Rather than having programmability within each network node, the separated OpenFlow controller provides network control through pluggable software. Our proposed network controller architecture will enable researchers to use their own software to control their own virtual networks. Flexibility and scalability were achieved by designing the network controller as a modularized and distributed system on a cluster of servers. Testing showed that a group of servers can efficiently cooperate to serve as a scalable OpenFlow controller. Testing using the nationwide JGN2plus network demonstrated that high-definition video can be delivered through OpenFlow-based point-to-point and point-to-multipoint paths.
Kazuya SUZUKI Kentaro SONODA Nobuyuki TOMIZAWA Yutaka YAKUWA Terutaka UCHIDA Yuta HIGUCHI Toshio TONOUCHI Hideyuki SHIMONISHI
The paper presents a survey on OpenFlow related technologies that have been proposed as a means for researchers, network service creators, and others to easily design, test, and deploy their innovative ideas in experimental or production networks to accelerate research activities on network technologies. Rather than having programmability within each network node, separated OpenFlow controllers provide network control through pluggable software modules; thus, it is easy to develop new network control functions in executable form and test them in production networks. The emergence of OpenFlow has started various research activities. The paper surveys these activities and their results.
Yudai HONMA Masaki AIDA Hideyuki SHIMONISHI Atsushi IWATA
We present a new multi-path routing methodology, MLB-routing, that is based on the multinomial logit model, which is well known in the random utility field. The key concept of the study is to set multiple paths from the origin to the destination, and distribute packets in accordance with multinomial logit type probability. Since MLB-routing is pure multi-path routing, it reduces the convergence on some links and increases bandwidth utilization in the network. Unlike existing multi-path routing schemes, which pre-set alternate paths, the proposed method can dynamically distribute packets to every possible path and thus is more efficient. Furthermore, it should be mentioned that this methodology can be implemented as either a link-state protocol or a distance-vector protocol. Therefore, it well supports the existing Internet. Simulations show that this methodology raises network utilization and significantly reduces end-to-end delay and jitter.
Yasuhiro YAMASAKI Hideyuki SHIMONISHI Tutomu MURASE
The advent of various quality-sensitive applications has greatly changed the requirements for IP network management and made the monitoring of individual traffic flows more important. Since the processing costs of per-flow quality monitoring are high, especially in high-speed backbone links, packet sampling techniques have been attracting considerable attention. Existing sampling techniques, such as those used in Sampled NetFlow and sFlow, however, focus on the monitoring of traffic volume, and there has been little discussion of the monitoring of such quality indexes as packet loss ratio. In this paper we propose a method for estimating, from sampled packets, packet loss ratios in individual TCP sessions. It detects packet loss events by monitoring duplicate ACK events raised by each TCP receiver. Because sampling reveals only a portion of the actual packet loss, the actual packet loss ratio is estimated statistically. Simulation results show that the proposed method can estimate the TCP packet loss ratio accurately from a 10% sampling of packets.
Tutomu MURASE Hideyuki SHIMONISHI Masayuki MURATA
Overlay networks are expected to be a promising technology for the realization of QoS (Quality of Service) control. Overlay networks have recently attracted considerable attention due to the following advantages: a new service can be developed in a short duration and it can be started with a low cost. The definition and necessity of the overlay network is described, and the classification of various current and future overlay networks, particularly according to the QoS feature, is attempted. In order to realize QoS control, it is considered that routing overlay and session overlay are promising solutions. In particular, session and overlay networks are explained in detail since new TCP protocols for QoS instead of current TCP protocols that control congestion in the Internet can be used within overlay networks. However, many open issues such as scalability still need further research and development although overlay networks have many attractive features and possess the potential to become a platform for the deployment of new services.
Yusuke SAKUMOTO Masaki AIDA Hideyuki SHIMONISHI
In this paper, we propose a novel Autonomous Decentralized Control (ADC) scheme for indirectly controlling a system performance variable of large-scale and wide-area networks. In a large-scale and wide-area network, since it is impractical for any one node to gather full information of the entire network, network control must be realized by inter-node collaboration using information local to each node. Several critical network problems (e.g., resource allocation) are often formulated by a system performance variable that is an amount to quantify system state. We solve such problems by designing an autonomous node action that indirectly controls, via the Markov Chain Monte Carlo method, the probability distribution of a system performance variable by using only local information. Analyses based on statistical mechanics confirm the effectiveness of the proposed node action. Moreover, the proposal is used to implement traffic-aware virtual machine placement control with load balancing in a data center network. Simulations confirm that it can control the system performance variable and is robust against system fluctuations. A comparison against a centralized control scheme verifies the superiority of the proposal.
Hideyuki SHIMONISHI Hiroshi SUZUKI
Weighted Round Robin (WRR) scheduling is an extension of round robin scheduling. Because of its simplicity and bandwidth guarantee, WRR cell scheduling is commonly used in ATM switches. However, since cells in individual queues are sent cyclically, the delay bounds in WRR scheduling grow as the number of queues increases. Thus, static priority scheduling is often used with WRR to improve the delay bounds of real-time queues. In this paper, we show that the burstiness generated in the network is an even greater factor affecting the degradation of delay bounds. In ATM switches with per-class queueing, a number of connections are multiplexed into one class-queue. The multiplexed traffic will have a burstiness even if each connection has no burstiness, and when the multiplexed traffic is separated at the down stream switches, the separated traffic will have a burstiness even if the multiplexed traffic has been shaped in the upstream switches. In this paper, we propose a new WRR scheme, namely, WRR with Save and Borrow (WRR/SB), that helps improving the delay bound performance of WRR by taking into account the burstiness generated in the network. We analyze these cell scheduling methods to discuss their delay characteristics. Through some numerical examples, we show that delay bounds in WRR are mainly dominated by the burstiness of input traffic and, thus WRR/SP, which is a combination of WRR and static priority scheduling, is less effective in improving delay bounds. We show that WRR/SB can provide better delay bounds than WRR and that it can achieve the same target delay bound with a smaller extra bandwidth, while large extra bandwidth must be allocated for WRR.
Ichinoshin MAKI Hideyuki SHIMONISHI Tutomu MURASE Masayuki MURATA
Because of the development of recent broadband access technologies, fair service among users is becoming more important goal. The most promising router mechanisms for providing fair service is per-flow traffic management. However, it is difficult to implement in high-speed core routers because per-flow state management is prohibitively expensive; thus, a large number of flows are aggregated into a small number of queues. This is not an acceptable situation because fairness degrades as the number of flows so aggregated increases. In this paper, we propose a new traffic management scheme called Hierarchically Aggregated Fair Queueing (HAFQ) to provide per-flow fair service. Our scheme can adjust flow aggregation levels according to the queue handling capability of various routers. This means the proposed scheme scales well in high-speed networks. HAFQ improves the fairness among aggregated flows by estimating the number of flows aggregated in a queue and allocating bandwidth to the queue proportionally. In addition, since HAFQ can identify flows having higher arrival rates simultaneously while estimating the number of flows, it enhances the fairness by preferentially dropping their packets. We show that our scheme can provide per-flow fair service through extensive simulation and experiments using a network processor. Since the currently available network processors (Intel IXP1200 in our case) are not high capacity, we also give extensive discussions on the applicability of our scheme to the high-speed core routers.
Hideyuki SHIMONISHI Takayuki HAMA M.Y. SANADIDI Mario GERLA Tutomu MURASE
An overlay traffic control is a way to provide flexible and deployable QoS mechanisms over existing networks, such as the Internet. While most of QoS mechanisms proposed so far require router supports, overlay QoS mechanisms rely on traffic control at transport layer without modifying existing routers in the network. Thus, traffic control algorithms, which are implemented at traffic sources or PEPs (Performance Enhancement Proxies), play a key role in an overlay QoS mechanism. In this paper, we propose an end-to-end prioritization scheme using TCP-Westwood Low-Priority (TCPW-LP), a low-priority traffic control scheme that maximizes the utilization of residual capacity without intrusion on coexisting foreground flows. Simulation and Internet measurement results show that TCPW-LP appropriately provides end-to-end low-priority service without any router supports. Under a wide range of buffer capacity and link error losses, TCPW-LP appropriately defers to foreground flows and better utilizes the residual capacity than other proposed priority schemes or even TCP Reno.
Takuya KUWAHARA Takayuki KURODA Manabu NAKANOYA Yutaka YAKUWA Hideyuki SHIMONISHI
As IT systems, including network systems using SDN/NFV technologies, become large-scaled and complicated, the cost of system management also increases rapidly. Network operators have to maintain their workflow in constructing and consistently updating such complex systems, and thus these management tasks in generating system update plan are desired to be automated. Declarative system update with state space search is a promising approach to enable this automation, however, the current methods is not enough scalable to practical systems. In this paper, we propose a novel heuristic approach to greatly reduce computation time to solve system update procedure for practical systems. Our heuristics accounts for structural bottleneck of the system update and advance search to resolve bottlenecks of current system states. This paper includes the following contributions: (1) formal definition of a novel heuristic function specialized to system update for A* search algorithm, (2) proofs that our heuristic function is consistent, i.e., A* algorithm with our heuristics returns a correct optimal solution and can omit repeatedly expansion of nodes in search spaces, and (3) results of performance evaluation of our heuristics. We evaluate the proposed algorithm in two cases; upgrading running hypervisor and rolling update of running VMs. The results show that computation time to solve system update plan for a system with 100 VMs does not exceed several minutes, whereas the conventional algorithm is only applicable for a very small system.