The search functionality is under construction.

Author Search Result

[Author] Zixiao ZHANG(2hit)

1-2hit
  • Dynamic VNF Scheduling: A Deep Reinforcement Learning Approach

    Zixiao ZHANG  Fujun HE  Eiji OKI  

     
    PAPER-Network

      Pubricized:
    2023/01/10
      Vol:
    E106-B No:7
      Page(s):
    557-570

    This paper introduces a deep reinforcement learning approach to solve the virtual network function scheduling problem in dynamic scenarios. We formulate an integer linear programming model for the problem in static scenarios. In dynamic scenarios, we define the state, action, and reward to form the learning approach. The learning agents are applied with the asynchronous advantage actor-critic algorithm. We assign a master agent and several worker agents to each network function virtualization node in the problem. The worker agents work in parallel to help the master agent make decision. We compare the introduced approach with existing approaches by applying them in simulated environments. The existing approaches include three greedy approaches, a simulated annealing approach, and an integer linear programming approach. The numerical results show that the introduced deep reinforcement learning approach improves the performance by 6-27% in our examined cases.

  • Joint Virtual Network Function Deployment and Scheduling via Heuristics and Deep Reinforcement Learning

    Zixiao ZHANG  Eiji OKI  

     
    PAPER-Network

      Pubricized:
    2023/08/01
      Vol:
    E106-B No:12
      Page(s):
    1424-1440

    This paper introduces heuristic approaches and a deep reinforcement learning approach to solve a joint virtual network function deployment and scheduling problem in a dynamic scenario. We formulate the problem as an optimization problem. Based on the mathematical description of the optimization problem, we introduce three heuristic approaches and a deep reinforcement learning approach to solve the problem. We define an objective to maximize the ratio of delay-satisfied requests while minimizing the average resource cost for a dynamic scenario. Our introduced two greedy approaches are named finish time greedy and computational resource greedy, respectively. In the finish time greedy approach, we make each request be finished as soon as possible despite its resource cost; in the computational resource greedy approach, we make each request occupy as few resources as possible despite its finish time. Our introduced simulated annealing approach generates feasible solutions randomly and converges to an approximate solution. In our learning-based approach, neural networks are trained to make decisions. We use a simulated environment to evaluate the performances of our introduced approaches. Numerical results show that the introduced deep reinforcement learning approach has the best performance in terms of benefit in our examined cases.