The search functionality is under construction.

Keyword Search Result

[Keyword] federated learning(6hit)

1-6hit
  • Federated Learning of Neural ODE Models with Different Iteration Counts Open Access

    Yuto HOSHINO  Hiroki KAWAKAMI  Hiroki MATSUTANI  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2024/02/09
      Vol:
    E107-D No:6
      Page(s):
    781-791

    Federated learning is a distributed machine learning approach in which clients train models locally with their own data and upload them to a server so that their trained results are shared between them without uploading raw data to the server. There are some challenges in federated learning, such as communication size reduction and client heterogeneity. The former can mitigate the communication overheads, and the latter can allow the clients to choose proper models depending on their available compute resources. To address these challenges, in this paper, we utilize Neural ODE based models for federated learning. The proposed flexible federated learning approach can reduce the communication size while aggregating models with different iteration counts or depths. Our contribution is that we experimentally demonstrate that the proposed federated learning can aggregate models with different iteration counts or depths. It is compared with a different federated learning approach in terms of the accuracy. Furthermore, we show that our approach can reduce communication size by up to 89.4% compared with a baseline ResNet model using CIFAR-10 dataset.

  • Federated Deep Reinforcement Learning for Multimedia Task Offloading and Resource Allocation in MEC Networks Open Access

    Rongqi ZHANG  Chunyun PAN  Yafei WANG  Yuanyuan YAO  Xuehua LI  

     
    PAPER-Network

      Vol:
    E107-B No:6
      Page(s):
    446-457

    With maturation of 5G technology in recent years, multimedia services such as live video streaming and online games on the Internet have flourished. These multimedia services frequently require low latency, which pose a significant challenge to compute the high latency requirements multimedia tasks. Mobile edge computing (MEC), is considered a key technology solution to address the above challenges. It offloads computation-intensive tasks to edge servers by sinking mobile nodes, which reduces task execution latency and relieves computing pressure on multimedia devices. In order to use MEC paradigm reasonably and efficiently, resource allocation has become a new challenge. In this paper, we focus on the multimedia tasks which need to be uploaded and processed in the network. We set the optimization problem with the goal of minimizing the latency and energy consumption required to perform tasks in multimedia devices. To solve the complex and non-convex problem, we formulate the optimization problem as a distributed deep reinforcement learning (DRL) problem and propose a federated Dueling deep Q-network (DDQN) based multimedia task offloading and resource allocation algorithm (FDRL-DDQN). In the algorithm, DRL is trained on the local device, while federated learning (FL) is responsible for aggregating and updating the parameters from the trained local models. Further, in order to solve the not identically and independently distributed (non-IID) data problem of multimedia devices, we develop a method for selecting participating federated devices. The simulation results show that the FDRL-DDQN algorithm can reduce the total cost by 31.3% compared to the DQN algorithm when the task data is 1000 kbit, and the maximum reduction can be 35.3% compared to the traditional baseline algorithm.

  • Frameworks for Privacy-Preserving Federated Learning

    Le Trieu PHONG  Tran Thi PHUONG  Lihua WANG  Seiichi OZAWA  

     
    INVITED PAPER

      Pubricized:
    2023/09/25
      Vol:
    E107-D No:1
      Page(s):
    2-12

    In this paper, we explore privacy-preserving techniques in federated learning, including those can be used with both neural networks and decision trees. We begin by identifying how information can be leaked in federated learning, after which we present methods to address this issue by introducing two privacy-preserving frameworks that encompass many existing privacy-preserving federated learning (PPFL) systems. Through experiments with publicly available financial, medical, and Internet of Things datasets, we demonstrate the effectiveness of privacy-preserving federated learning and its potential to develop highly accurate, secure, and privacy-preserving machine learning systems in real-world scenarios. The findings highlight the importance of considering privacy in the design and implementation of federated learning systems and suggest that privacy-preserving techniques are essential in enabling the development of effective and practical machine learning systems.

  • Physical Status Representation in Multiple Administrative Optical Networks by Federated Unsupervised Learning

    Takahito TANIMURA  Riu HIRAI  Nobuhiko KIKUCHI  

     
    PAPER

      Pubricized:
    2023/08/01
      Vol:
    E106-B No:11
      Page(s):
    1084-1092

    We present our data-collection and deep neural network (DNN)-training scheme for extracting the optical status from signals received by digital coherent optical receivers in fiber-optic networks. The DNN is trained with unlabeled datasets across multiple administrative network domains by combining federated learning and unsupervised learning. The scheme allows network administrators to train a common DNN-based encoder that extracts optical status in their networks without revealing their private datasets. An early-stage proof of concept was numerically demonstrated by simulation by estimating the optical signal-to-noise ratio and modulation format with 64-GBd 16QAM and quadrature phase-shift keying signals.

  • Communication-Efficient Federated Indoor Localization with Layerwise Swapping Training-FedAvg

    Jinjie LIANG  Zhenyu LIU  Zhiheng ZHOU  Yan XU  

     
    PAPER-Mobile Information Network and Personal Communications

      Pubricized:
    2022/05/11
      Vol:
    E105-A No:11
      Page(s):
    1493-1502

    Federated learning is a promising strategy for indoor localization that can reduce the labor cost of constructing a fingerprint dataset in a distributed training manner without privacy disclosure. However, the traffic generated during the whole training process of federated learning is a burden on the up-and-down link, which leads to huge energy consumption for mobile devices. Moreover, the non-independent and identically distributed (Non-IID) problem impairs the global localization performance during the federated learning. This paper proposes a communication-efficient FedAvg method for federated indoor localization which is improved by the layerwise asynchronous aggregation strategy and layerwise swapping training strategy. Energy efficiency can be improved by performing asynchronous aggregation between the model layers to reduce the traffic cost in the training process. Moreover, the impact of the Non-IID problem on the localization performance can be mitigated by performing swapping training on the deep layers. Extensive experimental results show that the proposed methods reduce communication traffic and improve energy efficiency significantly while mitigating the impact of the Non-IID problem on the precision of localization.

  • Layer-Based Communication-Efficient Federated Learning with Privacy Preservation

    Zhuotao LIAN  Weizheng WANG  Huakun HUANG  Chunhua SU  

     
    PAPER

      Pubricized:
    2021/09/28
      Vol:
    E105-D No:2
      Page(s):
    256-263

    In recent years, federated learning has attracted more and more attention as it could collaboratively train a global model without gathering the users' raw data. It has brought many challenges. In this paper, we proposed layer-based federated learning system with privacy preservation. We successfully reduced the communication cost by selecting several layers of the model to upload for global averaging and enhanced the privacy protection by applying local differential privacy. We evaluated our system in non independently and identically distributed scenario on three datasets. Compared with existing works, our solution achieved better performance in both model accuracy and training time.