The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] source(799hit)

1-20hit(799hit)

  • Joint Optimization of Task Offloading and Resource Allocation for UAV-Assisted Edge Computing: A Stackelberg Bilayer Game Approach Open Access

    Peng WANG  Guifen CHEN  Zhiyao SUN  

     
    PAPER-Information Network

      Pubricized:
    2024/05/21
      Vol:
    E107-D No:9
      Page(s):
    1174-1181

    Unmanned Aerial Vehicle (UAV)-assisted Mobile Edge Computing (MEC) can provide mobile users (MU) with additional computing services and a wide range of connectivity. This paper investigates the joint optimization strategy of task offloading and resource allocation for UAV-assisted MEC systems in complex scenarios with the goal of reducing the total system cost, consisting of task execution latency and energy consumption. We adopt a game theoretic approach to model the interaction process between the MEC server and the MU Stackelberg bilayer game model. Then, the original problem with complex multi-constraints is transformed into a duality problem using the Lagrangian duality method. Furthermore, we prove that the modeled Stackelberg bilayer game has a unique Nash equilibrium solution. In order to obtain an approximate optimal solution to the proposed problem, we propose a two-stage alternating iteration (TASR) algorithm based on the subgradient method and the marginal revenue optimization method. We evaluate the effective performance of the proposed algorithm through detailed simulation experiments. The simulation results show that the proposed algorithm is superior and robust compared to other benchmark methods and can effectively reduce the task execution latency and total system cost in different scenarios.

  • Cloud-Edge-End Collaborative Multi-Service Resource Management for IoT-Based Distribution Grid Open Access

    Feng WANG  Xiangyu WEN  Lisheng LI  Yan WEN  Shidong ZHANG  Yang LIU  

     
    PAPER-Communications Environment and Ethics

      Pubricized:
    2024/05/13
      Vol:
    E107-A No:9
      Page(s):
    1542-1555

    The rapid advancement of cloud-edge-end collaboration offers a feasible solution to realize low-delay and low-energy-consumption data processing for internet of things (IoT)-based smart distribution grid. The major concern of cloud-edge-end collaboration lies on resource management. However, the joint optimization of heterogeneous resources involves multiple timescales, and the optimization decisions of different timescales are intertwined. In addition, burst electromagnetic interference will affect the channel environment of the distribution grid, leading to inaccuracies in optimization decisions, which can result in negative influences such as slow convergence and strong fluctuations. Hence, we propose a cloud-edge-end collaborative multi-timescale multi-service resource management algorithm. Large-timescale device scheduling is optimized by sliding window pricing matching, which enables accurate matching estimation and effective conflict elimination. Small-timescale compression level selection and power control are jointly optimized by disturbance-robust upper confidence bound (UCB), which perceives the presence of electromagnetic interference and adjusts exploration tendency for convergence improvement. Simulation outcomes illustrate the excellent performance of the proposed algorithm.

  • MDX-Mixer: Music Demixing by Leveraging Source Signals Separated by Existing Demixing Models Open Access

    Tomoyasu NAKANO  Masataka GOTO  

     
    PAPER-Music Information Processing

      Pubricized:
    2024/04/05
      Vol:
    E107-D No:8
      Page(s):
    1079-1088

    This paper presents MDX-Mixer, which improves music demixing (MDX) performance by leveraging source signals separated by multiple existing MDX models. Deep-learning-based MDX models have improved their separation performances year by year for four kinds of sound sources: “vocals,” “drums,” “bass,” and “other”. Our research question is whether mixing (i.e., weighted sum) the signals separated by state-of-the-art MDX models can obtain either the best of everything or higher separation performance. Previously, in singing voice separation and MDX, there have been studies in which separated signals of the same sound source are mixed with each other using time-invariant or time-varying positive mixing weights. In contrast to those, this study is novel in that it allows for negative weights as well and performs time-varying mixing using all of the separated source signals and the music acoustic signal before separation. The time-varying weights are estimated by modeling the music acoustic signals and their separated signals by dividing them into short segments. In this paper we propose two new systems: one that estimates time-invariant weights using 1×1 convolution, and one that estimates time-varying weights by applying the MLP-Mixer layer proposed in the computer vision field to each segment. The latter model is called MDX-Mixer. Their performances were evaluated based on the source-to-distortion ratio (SDR) using the well-known MUSDB18-HQ dataset. The results show that the MDX-Mixer achieved higher SDR than the separated signals given by three state-of-the-art MDX models.

  • A Joint Coverage Constrained Task Offloading and Resource Allocation Method in MEC Open Access

    Daxiu ZHANG  Xianwei LI  Bo WEI  Yukun SHI  

     
    PAPER-Mobile Information Network and Personal Communications

      Vol:
    E107-A No:8
      Page(s):
    1277-1285

    With the increase of the number of Mobile User Equipments (MUEs), numerous tasks that with high requirements of resources are generated. However, the MUEs have limited computational resources, computing power and storage space. In this paper, a joint coverage constrained task offloading and resource allocation method based on deep reinforcement learning is proposed. The aim is offloading the tasks that cannot be processed locally to the edge servers to alleviate the conflict between the resource constraints of MUEs and the high performance task processing. The studied problem considers the dynamic variability and complexity of the system model, coverage, offloading decisions, communication relationships and resource constraints. An entropy weight method is used to optimize the resource allocation process and balance the energy consumption and execution time. The results of the study show that the number of tasks and MUEs affects the execution time and energy consumption of the task offloading and resource allocation processes in the interest of the service provider, and enhances the user experience.

  • Analytical Model of Maximum Operating Frequency of Class-D ZVS Inverter with Linearized Parasitic Capacitance and any Duty Ratio Open Access

    Yi XIONG  Senanayake THILAK  Yu YONEZAWA  Jun IMAOKA  Masayoshi YAMAMOTO  

     
    PAPER-Circuit Theory

      Pubricized:
    2023/12/05
      Vol:
    E107-A No:8
      Page(s):
    1115-1126

    This paper proposes an analytical model of maximum operating frequency of class-D zero-voltage-switching (ZVS) inverter. The model includes linearized drain-source parasitic capacitance and any duty ratio. The nonlinear drain-source parasitic capacitance is equally linearized through a charge-related equation. The model expresses the relationship among frequency, shunt capacitance, duty ratio, load impedance, output current phase, and DC input voltage under the ZVS condition. The analytical result shows that the maximum operating frequency under the ZVS condition can be obtained when the duty ratio, the output current phase, and the DC input voltage are set to optimal values. A 650 V/30 A SiC-MOSFET is utilized for both simulated and experimental verification, resulting in good consistency.

  • Improved PBFT-Based High Security and Large Throughput Data Resource Sharing for Distribution Power Grid Open Access

    Zhimin SHAO  Chunxiu LIU  Cong WANG  Longtan LI  Yimin LIU  Zaiyan ZHOU  

     
    PAPER-Systems and Control

      Pubricized:
    2024/01/31
      Vol:
    E107-A No:8
      Page(s):
    1085-1097

    Data resource sharing can guarantee the reliable and safe operation of distribution power grid. However, it faces the challenges of low security and high delay in the sharing process. Consortium blockchain can ensure the security and efficiency of data resource sharing, but it still faces problems such as arbitrary master node selection and high consensus delay. In this paper, we propose an improved practical Byzantine fault tolerance (PBFT) consensus algorithm based on intelligent consensus node selection to realize high-security and real-time data resource sharing for distribution power grid. Firstly, a blockchain-based data resource sharing model is constructed to realize secure data resource storage by combining the consortium blockchain and interplanetary file system (IPFS). Then, the improved PBFT consensus algorithm is proposed to optimize the consensus node selection based on the upper confidence bound of node performance. It prevents Byzantine nodes from participating in the consensus process, reduces the consensus delay, and improves the security of data resource sharing. The simulation results verify the effectiveness of the proposed algorithm.

  • Conflict Management Method Based on a New Belief Divergence in Evidence Theory Open Access

    Zhu YIN  Xiaojian MA  Hang WANG  

     
    PAPER-Office Information Systems, e-Business Modeling

      Pubricized:
    2024/03/01
      Vol:
    E107-D No:7
      Page(s):
    857-868

    Highly conflicting evidence that may lead to the counter-intuitive results is one of the challenges for information fusion in Dempster-Shafer evidence theory. To deal with this issue, evidence conflict is investigated based on belief divergence measuring the discrepancy between evidence. In this paper, the pignistic probability transform belief χ2 divergence, named as BBχ2 divergence, is proposed. By introducing the pignistic probability transform, the proposed BBχ2 divergence can accurately quantify the difference between evidence with the consideration of multi-element sets. Compared with a few belief divergences, the novel divergence has more precision. Based on this advantageous divergence, a new multi-source information fusion method is devised. The proposed method considers both credibility weights and information volume weights to determine the overall weight of each evidence. Eventually, the proposed method is applied in target recognition and fault diagnosis, in which comparative analysis indicates that the proposed method can realize the highest accuracy for managing evidence conflict.

  • Dataset of Functionally Equivalent Java Methods and Its Application to Evaluating Clone Detection Tools Open Access

    Yoshiki HIGO  

     
    PAPER-Software System

      Pubricized:
    2024/02/21
      Vol:
    E107-D No:6
      Page(s):
    751-760

    Modern high-level programming languages have a wide variety of grammar and can implement the required functionality in different ways. The authors believe that a large amount of code that implements the same functionality in different ways exists even in open source software where the source code is publicly available, and that by collecting such code, a useful data set can be constructed for various studies in software engineering. In this study, we construct a dataset of pairs of Java methods that have the same functionality but different structures from approximately 314 million lines of source code. To construct this dataset, the authors used an automated test generation technique, EvoSuite. Test cases generated by automated test generation techniques have the property that the test cases always succeed. In constructing the dataset, using this property, test cases generated from two methods were executed against each other to automatically determine whether the behavior of the two methods is the same to some extent. Pairs of methods for which all test cases succeeded in cross-running test cases are manually investigated to be functionally equivalent. This paper also reports the results of an accuracy evaluation of code clone detection tools using the constructed dataset. The purpose of this evaluation is assessing how accurately code clone detection tools could find the functionally equivalent methods, not assessing the accuracy of detecting ordinary clones. The constructed dataset is available at github (https://github.com/YoshikiHigo/FEMPDataset).

  • Federated Deep Reinforcement Learning for Multimedia Task Offloading and Resource Allocation in MEC Networks Open Access

    Rongqi ZHANG  Chunyun PAN  Yafei WANG  Yuanyuan YAO  Xuehua LI  

     
    PAPER-Network

      Vol:
    E107-B No:6
      Page(s):
    446-457

    With maturation of 5G technology in recent years, multimedia services such as live video streaming and online games on the Internet have flourished. These multimedia services frequently require low latency, which pose a significant challenge to compute the high latency requirements multimedia tasks. Mobile edge computing (MEC), is considered a key technology solution to address the above challenges. It offloads computation-intensive tasks to edge servers by sinking mobile nodes, which reduces task execution latency and relieves computing pressure on multimedia devices. In order to use MEC paradigm reasonably and efficiently, resource allocation has become a new challenge. In this paper, we focus on the multimedia tasks which need to be uploaded and processed in the network. We set the optimization problem with the goal of minimizing the latency and energy consumption required to perform tasks in multimedia devices. To solve the complex and non-convex problem, we formulate the optimization problem as a distributed deep reinforcement learning (DRL) problem and propose a federated Dueling deep Q-network (DDQN) based multimedia task offloading and resource allocation algorithm (FDRL-DDQN). In the algorithm, DRL is trained on the local device, while federated learning (FL) is responsible for aggregating and updating the parameters from the trained local models. Further, in order to solve the not identically and independently distributed (non-IID) data problem of multimedia devices, we develop a method for selecting participating federated devices. The simulation results show that the FDRL-DDQN algorithm can reduce the total cost by 31.3% compared to the DQN algorithm when the task data is 1000 kbit, and the maximum reduction can be 35.3% compared to the traditional baseline algorithm.

  • Performance Comparison of the Two Reconstruction Methods for Stabilizer-Based Quantum Secret Sharing

    Shogo CHIWAKI  Ryutaroh MATSUMOTO  

     
    LETTER-Quantum Information Theory

      Pubricized:
    2023/09/20
      Vol:
    E107-A No:3
      Page(s):
    526-529

    Stabilizer-based quantum secret sharing has two methods to reconstruct a quantum secret: The erasure correcting procedure and the unitary procedure. It is known that the unitary procedure has a smaller circuit width. On the other hand, it is unknown which method has smaller depth and fewer circuit gates. In this letter, it is shown that the unitary procedure has smaller depth and fewer circuit gates than the erasure correcting procedure which follows a standard framework performing measurements and unitary operators according to the measurements outcomes, when the circuits are designed for quantum secret sharing using the [[5, 1, 3]] binary stabilizer code. The evaluation can be reversed if one discovers a better circuit for the erasure correcting procedure which does not follow the standard framework.

  • An Efficient Bayes Coding Algorithm for Changing Context Tree Model

    Koshi SHIMADA  Shota SAITO  Toshiyasu MATSUSHIMA  

     
    PAPER-Source Coding and Data Compression

      Pubricized:
    2023/08/24
      Vol:
    E107-A No:3
      Page(s):
    448-457

    The context tree model has the property that the occurrence probability of symbols is determined from a finite past sequence and is a broader class of sources that includes i.i.d. or Markov sources. This paper proposes a non-stationary source with context tree models that change from interval to interval. The Bayes code for this source requires weighting of the posterior probabilities of the context tree models and change points, so the computational complexity of it usually increases to exponential order. Therefore, the challenge is how to reduce the computational complexity. In this paper, we propose a special class of prior probability distribution of context tree models and change points and develop an efficient Bayes coding algorithm by combining two existing Bayes coding algorithms. The algorithm minimizes the Bayes risk function of the proposed source in this paper, and the computational complexity of the proposed algorithm is polynomial order. We investigate the behavior and performance of the proposed algorithm by conducting experiments.

  • Properties of k-Bit Delay Decodable Codes

    Kengo HASHIMOTO  Ken-ichi IWATA  

     
    PAPER-Source Coding and Data Compression

      Pubricized:
    2023/09/07
      Vol:
    E107-A No:3
      Page(s):
    417-447

    The class of k-bit delay decodable codes, source codes allowing decoding delay of at most k bits for k≥0, can attain a shorter average codeword length than Huffman codes. This paper discusses the general properties of the class of k-bit delay decodable codes with a finite number of code tables and proves two theorems which enable us to limit the scope of codes to be considered when discussing optimal k-bit delay decodable codes.

  • Proof of Achievability Part of Rate-Distortion Theorem without Random Coding

    Mikihiko NISHIARA  Yuki ITO  

     
    PAPER-Source Coding and Data Compression

      Pubricized:
    2023/10/10
      Vol:
    E107-A No:3
      Page(s):
    404-408

    The achievability part of the rate-distortion theorem is proved by showing existence of good codes. For i.i.d. sources, two methods showing existence are known; random coding and non-random coding. For general sources, however, no proof in which good codes are constructed with non-random coding is found. In this paper, with a non-random method of code construction, we prove the achievability part of the rate-distortion theorem for general sources. Moreover, we also prove a stochastic variation of the rate-distortion theorem with the same method.

  • A Fundamental Limit of Variable-Length Compression with Worst-Case Criteria in Terms of Side Information

    Sho HIGUCHI  Yuta SAKAI  

     
    PAPER-Source Coding and Data Compression

      Pubricized:
    2023/07/03
      Vol:
    E107-A No:3
      Page(s):
    384-392

    In this study, we consider the data compression with side information available at both the encoder and the decoder. The information source is assigned to a variable-length code that does not have to satisfy the prefix-free constraints. We define several classes of codes whose codeword lengths and error probabilities satisfy worse-case criteria in terms of side-information. As a main result, we investigate the exact first-order asymptotics with second-order bounds scaled as Θ(√n) as blocklength n increases under the regime of nonvanishing error probabilities. To get this result, we also derive its one-shot bounds by employing the cutoff operation.

  • Performance of Collaborative MIMO Reception with User Grouping Schemes

    Eiku ANDO  Yukitoshi SANADA  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2023/10/23
      Vol:
    E107-B No:1
      Page(s):
    253-261

    This paper proposes user equipment (UE) grouping schemes and evaluates the performance of a scheduling scheme for each formed group in collaborative multiple-input multiple-output (MIMO) reception. In previous research, the criterion for UE grouping and the effects of group scheduling has never been presented. In the UE grouping scheme, two criteria, the base station (BS)-oriented one and the UE-oriented one, are presented. The BS-oriented full search scheme achieves ideal performance though it requires knowledge of the relative positions of all UEs. Therefore, the UE-oriented local search scheme is also proposed. As the scheduling scheme, proportional fairness scheduling is used in resource allocation for each formed group. When the number of total UEs increases, the difference in the number of UEs among groups enlarges. Numerical results obtained through computer simulation show that the throughput per user increases and the fairness among users decreases when the number of UEs in a cell increases in the proposed schemes compared to those of the conventional scheme.

  • Backhaul Prioritized Point-to-Multi-Point Wireless Transmission Using Orbital Angular Momentum Multiplexing

    Tomoya KAGEYAMA  Jun MASHINO  Doohwan LEE  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2023/09/21
      Vol:
    E107-B No:1
      Page(s):
    232-243

    Orbital angular momentum (OAM) multiplexing technology is being investigated for high-capacity point-to-point (PtP) wireless transmission toward beyond 5G systems. OAM multiplexing is a spatial multiplexing technique that utilizes the twisting of electromagnetic waves. Its advantage is that it reduces the computational complexity of the signal processing on spatial multiplexing. Meanwhile point-to-multi point (PtMP) wireless transmission, such as integrated access and backhaul (IAB) will be expected to simultaneously accommodates a high-capacity prioritized backhaul-link and access-links. In this paper, we study the extension of OAM multiplexing transmission from PtP to PtMP to meet the above requirements. We propose a backhaul prioritized resource control algorithm that maximizes the received signal-to-interference and noise ratio (SINR) of the access-links while maintaining the backhaul-link. The proposed algorithm features adaptive mode selection that takes into account the difference in the received power of each OAM mode depending on the user equipment position and the guaranteed power allocation of the backhaul capacity. We then evaluate the performance of the proposed method through computer simulation. The results show that throughput of the access-links improved compared with the conventional multi-beam multi-user multi-input multi-output (MIMO) techniques while maintaining the throughput of the backhaul-link above the required value with minimal feedback information.

  • Deep Neural Networks Based End-to-End DOA Estimation System Open Access

    Daniel Akira ANDO  Yuya KASE  Toshihiko NISHIMURA  Takanori SATO  Takeo OHGANE  Yasutaka OGAWA  Junichiro HAGIWARA  

     
    PAPER

      Pubricized:
    2023/09/11
      Vol:
    E106-B No:12
      Page(s):
    1350-1362

    Direction of arrival (DOA) estimation is an antenna array signal processing technique used in, for instance, radar and sonar systems, source localization, and channel state information retrieval. As new applications and use cases appear with the development of next generation mobile communications systems, DOA estimation performance must be continually increased in order to support the nonstop growing demand for wireless technologies. In previous works, we verified that a deep neural network (DNN) trained offline is a strong candidate tool with the promise of achieving great on-grid DOA estimation performance, even compared to traditional algorithms. In this paper, we propose new techniques for further DOA estimation accuracy enhancement incorporating signal-to-noise ratio (SNR) prediction and an end-to-end DOA estimation system, which consists of three components: source number estimator, DOA angular spectrum grid estimator, and DOA detector. Here, we expand the performance of the DOA detector and angular spectrum estimator, and present a new solution for source number estimation based on DNN with very simple design. The proposed DNN system applied with said enhancement techniques has shown great estimation performance regarding the success rate metric for the case of two radio wave sources although not fully satisfactory results are obtained for the case of three sources.

  • Single UAV-Based Wave Source Localization in NLOS Environments Open Access

    Shinichi MURATA  Takahiro MATSUDA  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2023/08/01
      Vol:
    E106-B No:12
      Page(s):
    1491-1500

    To localize an unknown wave source in non-line-of-sight environments, a wave source localization scheme using multiple unmanned-aerial-vehicles (UAVs) is proposed. In this scheme, each UAV estimates the direction-of-arrivals (DoAs) of received signals and the wave source is localized from the estimated DoAs by means of maximum likelihood estimation. In this study, by extending the concept of this scheme, we propose a novel wave source localization scheme using a single UAV. In the proposed scheme, the UAV moves on the path comprising multiple measurement points and the wave source is sequentially localized from DoA distributions estimated at these measurement points. At each measurement point, with a moving path planning algorithm, the UAV determines the next measurement point from the estimated DoA distributions and measurement points that the UAV has already visited. We consider two moving path planning algorithms, and validate the proposed scheme through simulation experiments.

  • Backup Resource Allocation Model with Probabilistic Protection Considering Service Delay

    Shinya HORIMOTO  Fujun HE  Eiji OKI  

     
    PAPER-Network

      Pubricized:
    2023/03/24
      Vol:
    E106-B No:9
      Page(s):
    798-816

    This paper proposes a backup resource allocation model for virtual network functions (VNFs) to minimize the total allocated computing capacity for backup with considering the service delay. If failures occur to primary hosts, the VNFs in failed hosts are recovered by backup hosts whose allocation is pre-determined. We introduce probabilistic protection, where the probability that the protection by a backup host fails is limited within a given value; it allows backup resource sharing to reduce the total allocated computing capacity. The previous work does not consider the service delay constraint in the backup resource allocation problem. The proposed model considers that the probability that the service delay, which consists of networking delay between hosts and processing delay in each VNF, exceeds its threshold is constrained within a given value. We introduce a basic algorithm to solve our formulated delay-constraint optimization problem. In a problem with the size that cannot be solved within an acceptable computation time limit by the basic algorithm, we develop a simulated annealing algorithm incorporating Yen's algorithm to handle the delay constraint heuristically. We observe that both algorithms in the proposed model reduce the total allocated computing capacity by up to 56.3% compared to a baseline; the simulated annealing algorithm can get feasible solutions in problems where the basic algorithm cannot.

  • Single-Power-Supply Six-Transistor CMOS SRAM Enabling Low-Voltage Writing, Low-Voltage Reading, and Low Standby Power Consumption Open Access

    Tadayoshi ENOMOTO  Nobuaki KOBAYASHI  

     
    PAPER-Electronic Circuits

      Pubricized:
    2023/03/16
      Vol:
    E106-C No:9
      Page(s):
    466-476

    We developed a self-controllable voltage level (SVL) circuit and applied this circuit to a single-power-supply, six-transistor complementary metal-oxide-semiconductor static random-access memory (SRAM) to not only improve both write and read performances but also to achieve low standby power and data retention (holding) capability. The SVL circuit comprises only three MOSFETs (i.e., pull-up, pull-down and bypass MOSFETs). The SVL circuit is able to adaptively generate both optimal memory cell voltages and word line voltages depending on which mode of operation (i.e., write, read or hold operation) was used. The write margin (VWM) and read margin (VRM) of the developed (dvlp) SRAM at a supply voltage (VDD) of 1V were 0.470 and 0.1923V, respectively. These values were 1.309 and 2.093 times VWM and VRM of the conventional (conv) SRAM, respectively. At a large threshold voltage (Vt) variability (=+6σ), the minimum power supply voltage (VMin) for the write operation of the conv SRAM was 0.37V, whereas it decreased to 0.22V for the dvlp SRAM. VMin for the read operation of the conv SRAM was 1.05V when the Vt variability (=-6σ) was large, but the dvlp SRAM lowered it to 0.41V. These results show that the SVL circuit expands the operating voltage range for both write and read operations to lower voltages. The dvlp SRAM reduces the standby power consumption (PST) while retaining data. The measured PST of the 2k-bit, 90-nm dvlp SRAM was only 0.957µW at VDD=1.0V, which was 9.46% of PST of the conv SRAM (10.12µW). The Si area overhead of the SVL circuits was only 1.383% of the dvlp SRAM.

1-20hit(799hit)