The search functionality is under construction.

Keyword Search Result

[Keyword] latency(105hit)

1-20hit(105hit)

  • RAN Slicing with Inter-Cell Interference Control and Link Adaptation for Reliable Wireless Communications Open Access

    Yoshinori TANAKA  Takashi DATEKI  

     
    PAPER-Terrestrial Wireless Communication/Broadcasting Technologies

      Vol:
    E107-B No:7
      Page(s):
    513-528

    Efficient multiplexing of ultra-reliable and low-latency communications (URLLC) and enhanced mobile broadband (eMBB) traffic, as well as ensuring the various reliability requirements of these traffic types in 5G wireless communications, is becoming increasingly important, particularly for vertical services. Interference management techniques, such as coordinated inter-cell scheduling, can enhance reliability in dense cell deployments. However, tight inter-cell coordination necessitates frequent information exchange between cells, which limits implementation. This paper introduces a novel RAN slicing framework based on centralized frequency-domain interference control per slice and link adaptation optimized for URLLC. The proposed framework does not require tight inter-cell coordination but can fulfill the requirements of both the decoding error probability and the delay violation probability of each packet flow. These controls are based on a power-law estimation of the lower tail distribution of a measured data set with a smaller number of discrete samples. As design guidelines, we derived a theoretical minimum radio resource size of a slice to guarantee the delay violation probability requirement. Simulation results demonstrate that the proposed RAN slicing framework can achieve the reliability targets of the URLLC slice while improving the spectrum efficiency of the eMBB slice in a well-balanced manner compared to other evaluated benchmarks.

  • Traffic Reduction for Speculative Video Transmission in Cloud Gaming Systems Open Access

    Takumasa ISHIOKA  Tatsuya FUKUI  Toshihito FUJIWARA  Satoshi NARIKAWA  Takuya FUJIHASHI  Shunsuke SARUWATARI  Takashi WATANABE  

     
    PAPER-Network

      Vol:
    E107-B No:5
      Page(s):
    408-418

    Cloud gaming systems allow users to play games that require high-performance computational capability on their mobile devices at any location. However, playing games through cloud gaming systems increases the Round-Trip Time (RTT) due to increased network delay. To simulate a local gaming experience for cloud users, we must minimize RTTs, which include network delays. The speculative video transmission pre-generates and encodes video frames corresponding to all possible user inputs and sends them to the user before the user’s input. The speculative video transmission mitigates the network, whereas a simple solution significantly increases the video traffic. This paper proposes tile-wise delta detection for traffic reduction of speculative video transmission. More specifically, the proposed method determines a reference video frame from the generated video frames and divides the reference video frame into multiple tiles. We calculate the similarity between each tile of the reference video frame and other video frames based on a hash function. Based on calculated similarity, we determine redundant tiles and do not transmit them to reduce traffic volume in minimal processing time without implementing a high compression ratio video compression technique. Evaluations using commercial games showed that the proposed method reduced 40-50% in traffic volume when the SSIM index was around 0.98 in certain genres, compared with the speculative video transmission method. Furthermore, to evaluate the feasibility of the proposed method, we investigated the effectiveness of network delay reduction with existing computational capability and the requirements in the future. As a result, we found that the proposed scheme may mitigate network delay by one to two frames, even with existing computational capability under limited conditions.

  • Resource Allocation for Mobile Edge Computing System Considering User Mobility with Deep Reinforcement Learning

    Kairi TOKUDA  Takehiro SATO  Eiji OKI  

     
    PAPER-Network

      Pubricized:
    2023/10/06
      Vol:
    E107-B No:1
      Page(s):
    173-184

    Mobile edge computing (MEC) is a key technology for providing services that require low latency by migrating cloud functions to the network edge. The potential low quality of the wireless channel should be noted when mobile users with limited computing resources offload tasks to an MEC server. To improve the transmission reliability, it is necessary to perform resource allocation in an MEC server, taking into account the current channel quality and the resource contention. There are several works that take a deep reinforcement learning (DRL) approach to address such resource allocation. However, these approaches consider a fixed number of users offloading their tasks, and do not assume a situation where the number of users varies due to user mobility. This paper proposes Deep reinforcement learning model for MEC Resource Allocation with Dummy (DMRA-D), an online learning model that addresses the resource allocation in an MEC server under the situation where the number of users varies. By adopting dummy state/action, DMRA-D keeps the state/action representation. Therefore, DMRA-D can continue to learn one model regardless of variation in the number of users during the operation. Numerical results show that DMRA-D improves the success rate of task submission while continuing learning under the situation where the number of users varies.

  • Chunk Grouping Method to Estimate Available Bandwidth for Adaptive Bitrate Live Streaming

    Daichi HATTORI  Masaki BANDAI  

     
    PAPER-Network

      Pubricized:
    2023/07/24
      Vol:
    E106-B No:11
      Page(s):
    1133-1142

    The Common Media Application Format (CMAF) is a standard for adaptive bitrate live streaming. The CMAF adapts chunk encoding and enables low-latency live streaming. However, conventional bandwidth estimation for adaptive bitrate streaming underestimates bandwidth because download time is affected not only by network bandwidth but also by the idle times between chunks in the same segment. Inaccurate bandwidth estimation decreases the quality of experience of the streaming client. In this paper, we propose a chunk-grouping method to estimate the available bandwidth for adaptive bitrate live streaming. In the proposed method, by delaying HTTP request transmission and bandwidth estimation using grouped chunks, the client estimates the available bandwidth accurately due to there being no idle times in the grouped chunks. In addition, we extend the proposed method to dynamically change the number of grouping chunks according to buffer length during downloading of the previous segment. We evaluate the proposed methods under various network conditions in order to confirm the effectiveness of the proposed methods.

  • Switch-Based Quorum Coordination for Low Tail Latency in Replicated Storage

    Gyuyeong KIM  

     
    LETTER-Information Network

      Pubricized:
    2023/08/22
      Vol:
    E106-D No:11
      Page(s):
    1922-1925

    Modern distributed storage requires microsecond-scale tail latency, but the current coordinator-based quorum coordination causes a burdensome latency overhead. This paper presents Archon, a new quorum coordination architecture that supports low tail latency for microsecond-scale replicated storage. The key idea of Archon is to perform the quorum coordination in the network switch by leveraging the flexibility and capability of emerging programmable switch ASICs. Our in-network quorum coordination is based on the observation that the modern programmable switch provides nanosecond-scale processing delay and high flexibility simultaneously. To realize the idea, we design a custom switch data plane. We implement a Archon prototype on an Intel Tofino switch and conduct a series of testbed experiments. Our experimental results show that Archon can provide lower tail latency than the coordinator-based solution.

  • Optimizing Edge-Cloud Cooperation for Machine Learning Accuracy Considering Transmission Latency and Bandwidth Congestion Open Access

    Kengo TAJIRI  Ryoichi KAWAHARA  Yoichi MATSUO  

     
    PAPER-Network Management/Operation

      Pubricized:
    2023/03/24
      Vol:
    E106-B No:9
      Page(s):
    827-836

    Machine learning (ML) has been used for various tasks in network operations in recent years. However, since the scale of networks has grown and the amount of data generated has increased, it has been increasingly difficult for network operators to conduct their tasks with a single server using ML. Thus, ML with edge-cloud cooperation has been attracting attention for efficiently processing and analyzing a large amount of data. In the edge-cloud cooperation setting, although transmission latency, bandwidth congestion, and accuracy of tasks using ML depend on the load balance of processing data with edge servers and a cloud server in edge-cloud cooperation, the relationship is too complex to estimate. In this paper, we focus on monitoring anomalous traffic as an example of ML tasks for network operations and formulate transmission latency, bandwidth congestion, and the accuracy of the task with edge-cloud cooperation considering the ratio of the amount of data preprocessed in edge servers to that in a cloud server. Moreover, we formulate an optimization problem under constraints for transmission latency and bandwidth congestion to select the proper ratio by using our formulation. By solving our optimization problem, the optimal load balance between edge servers and a cloud server can be selected, and the accuracy of anomalous traffic monitoring can be estimated. Our formulation and optimization framework can be used for other ML tasks by considering the generating distribution of data and the type of an ML model. In accordance with our formulation, we simulated the optimal load balance of edge-cloud cooperation in a topology that mimicked a Japanese network and conducted an anomalous traffic detection experiment by using real traffic data to compare the estimated accuracy based on our formulation and the actual accuracy based on the experiment.

  • LFWS: Long-Operation First Warp Scheduling Algorithm to Effectively Hide the Latency for GPUs

    Song LIU  Jie MA  Chenyu ZHAO  Xinhe WAN  Weiguo WU  

     
    PAPER-Algorithms and Data Structures

      Pubricized:
    2023/02/10
      Vol:
    E106-A No:8
      Page(s):
    1043-1050

    GPUs have become the dominant computing units to meet the need of high performance in various computational fields. But the long operation latency causes the underutilization of on-chip computing resources, resulting in performance degradation when running parallel tasks on GPUs. A good warp scheduling strategy is an effective solution to hide latency and improve resource utilization. However, most current warp scheduling algorithms on GPUs ignore the ability of long operations to hide latency. In this paper, we propose a long-operation-first warp scheduling algorithm, LFWS, for GPU platforms. The LFWS filters warps in the ready state to a ready queue and updates the queue in time according to changes in the status of the warp. The LFWS divides the warps in the ready queue into long and short operation groups based on the type of operations in their instruction buffers, and it gives higher priority to the long-operating warp in the ready queue. This can effectively use the long operations to hide some of the latency from each other and enhance the system's ability to hide the latency. To verify the effectiveness of the LFWS, we implement the LFWS algorithm on a simulation platform GPGPU-Sim. Experiments are conducted over various CUDA applications to evaluate the performance of LFWS algorithm, compared with other five warp scheduling algorithms. The results show that the LFWS algorithm achieves an average performance improvement of 8.01% and 5.09%, respectively, over three traditional and two novel warp scheduling algorithms, effectively improving computational resource utilization on GPU.

  • EMRNet: Efficient Modulation Recognition Networks for Continuous-Wave Radar Signals

    Kuiyu CHEN  Jingyi ZHANG  Shuning ZHANG  Si CHEN  Yue MA  

     
    BRIEF PAPER-Electronic Instrumentation and Control

      Pubricized:
    2023/03/24
      Vol:
    E106-C No:8
      Page(s):
    450-453

    Automatic modulation recognition(AMR) of radar signals is a currently active area, especially in electronic reconnaissance, where systems need to quickly identify the intercepted signal and formulate corresponding interference measures on computationally limited platforms. However, previous methods generally have high computational complexity and considerable network parameters, making the system unable to detect the signal timely in resource-constrained environments. This letter firstly proposes an efficient modulation recognition network(EMRNet) with tiny and low latency models to match the requirements for mobile reconnaissance equipments. One-dimensional residual depthwise separable convolutions block(1D-RDSB) with an adaptive size of receptive fields is developed in EMRNet to replace the traditional convolution block. With 1D-RDSB, EMRNet achieves a high classification accuracy and dramatically reduces computation cost and network paraments. The experiment results show that EMRNet can achieve higher precision than existing 2D-CNN methods, while the computational cost and parament amount of EMRNet are reduced by about 13.93× and 80.88×, respectively.

  • Protection of Latency-Strict Stations on WLAN Systems Using CTS-to-STA Frames

    Kenichi KAWAMURA  Shouta NAKAYAMA  Keisuke WAKAO  Takatsune MORIYAMA  Yasushi TAKATORI  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2022/11/28
      Vol:
    E106-B No:6
      Page(s):
    518-527

    Low-latency and highly reliable communication on wireless LAN (WLAN) is difficult due to interference from the surroundings. To overcome this problem, we have developed a scheme called Clear to Send-to-Station (CTS-STA) frame transmission control that enables stable latency communication in environments with strong interference from surrounding WLAN systems. This scheme uses the basic functions of WLAN standards and is effective for both the latest and legacy standard devices. It operates when latency-strict transmission is required for an STA and there is interference from surrounding WLAN devices while minimizing the control signal overhead. Experimental evaluations with prototype systems demonstrate the effectiveness of the proposed scheme.

  • A Low-Latency 4K HEVC Multi-Channel Encoding System with Content-Aware Bitrate Control for Live Streaming

    Daisuke KOBAYASHI  Ken NAKAMURA  Masaki KITAHARA  Tatsuya OSAWA  Yuya OMORI  Takayuki ONISHI  Hiroe IWASAKI  

     
    PAPER-Image Processing and Video Processing

      Pubricized:
    2022/09/30
      Vol:
    E106-D No:1
      Page(s):
    46-57

    This paper describes a novel low-latency 4K 60 fps HEVC (high efficiency video coding)/H.265 multi-channel encoding system with content-aware bitrate control for live streaming. Adaptive bitrate (ABR) streaming techniques, such as MPEG-DASH (dynamic adaptive streaming over HTTP) and HLS (HTTP live streaming), spread widely on Internet video streaming. Live content has increased with the expansion of streaming services, which has led to demands for traffic reduction and low latency. To reduce network traffic, we propose content-aware dynamic and seamless bitrate control that supports multi-channel real-time encoding for ABR, including 4K 60 fps video. Our method further supports chunked packaging transfer to provide low-latency streaming. We adopt a hybrid architecture consisting of hardware and software processing. The system consists of multiple 4K HEVC encoder LSIs that each LSI can encode 4K 60 fps or up to high-definition (HD) ×4 videos efficiently with the proposed bitrate control method. The software takes the packaging process according to the various streaming protocol. Experimental results indicate that our method reduces encoding bitrates obtained with constant bitrate encoding by as much as 56.7%, and the streaming latency over MPEG-DASH is 1.77 seconds.

  • Boosting the Performance of Interconnection Networks by Selective Data Compression

    Naoya NIWA  Hideharu AMANO  Michihiro KOIBUCHI  

     
    PAPER

      Pubricized:
    2022/07/12
      Vol:
    E105-D No:12
      Page(s):
    2057-2065

    This study presents a selective data-compression interconnection network to boost its performance. Data compression virtually increases the effective network bandwidth. One drawback of data compression is a long latency to perform (de-)compression operation at a compute node. In terms of the communication latency, we explore the trade-off between the compression latency overhead and the reduced injection latency by shortening the packet length by compression algorithms. As a result, we present to selectively apply a compression technique to a packet. We perform a compression operation to long packets and it is also taken when network congestion is detected at a source compute node. Through a cycle-accurate network simulation, the selective compression method using the above compression algorithms improves by up to 39% the network throughput with a moderate increase in the communication latency of short packets.

  • User-Centric Design of Millimeter Wave Communications for Beyond 5G and 6G Open Access

    Koji ISHIBASHI  Takanori HARA  Sota UCHIMURA  Tetsuya IYE  Yoshimi FUJII  Takahide MURAKAMI  Hiroyuki SHINBO  

     
    INVITED PAPER

      Pubricized:
    2022/07/13
      Vol:
    E105-B No:10
      Page(s):
    1117-1129

    In this paper, we propose new radio access network (RAN) architecture for reliable millimeter-wave (mmWave) communications, which has the flexibility to meet users' diverse and fluctuating requirements in terms of communication quality. This architecture is composed of multiple radio units (RUs) connected to a common distributed unit (DU) via fronthaul links to virtually enlarge its coverage. We further present grant-free non-orthogonal multiple access (GF-NOMA) for low-latency uplink communications with a massive number of users and robust coordinated multi-point (CoMP) transmission using blockage prediction for uplink/downlink communications with a high data rate and a guaranteed minimum data rate as the technical pillars of the proposed RAN. The numerical results indicate that our proposed architecture can meet completely different user requirements and realize a user-centric design of the RAN for beyond 5G/6G.

  • Energy-Efficient KBP: Kernel Enhancements for Low-Latency and Energy-Efficient Networking Open Access

    Kei FUJIMOTO  Ko NATORI  Masashi KANEKO  Akinori SHIRAGA  

     
    PAPER-Network

      Pubricized:
    2022/03/14
      Vol:
    E105-B No:9
      Page(s):
    1039-1052

    Real-time applications are becoming more and more popular, and due to the demand for more compact and portable user devices, offloading terminal processes to edge servers is being considered. Moreover, it is necessary to process packets with low latency on edge servers, which are often virtualized for operability. When trying to achieve low-latency networking, the increase in server power consumption due to performance tuning and busy polling for fast packet receiving becomes a problem. Thus, we design and implement a low-latency and energy-efficient networking system, energy-efficient kernel busy poll (EE-KBP), which meets four requirements: (A) low latency in the order of microseconds for packet forwarding in a virtual server, (B) lower power consumption than existing solutions, (C) no need for application modification, and (D) no need for software redevelopment with each kernel security update. EE-KBP sets a polling thread in a Linux kernel that receives packets with low latency in polling mode while packets are arriving, and when no packets are arriving, it sleeps and lowers the CPU operating frequency. Evaluations indicate that EE-KBP achieves microsecond-order low-latency networking under most traffic conditions, and 1.4× to 3.1× higher throughput with lower power consumption than NAPI used in a Linux kernel.

  • Adiabatic Quantum-Flux-Parametron with Delay-Line Clocking Using Square Excitation Currents

    Taiki YAMAE  Naoki TAKEUCHI  Nobuyuki YOSHIKAWA  

     
    PAPER

      Pubricized:
    2022/01/19
      Vol:
    E105-C No:6
      Page(s):
    277-282

    The adiabatic quantum-flux-parametron (AQFP) is an energy-efficient superconductor logic device. In a previous study, we proposed a low-latency clocking scheme called delay-line clocking, and several low-latency AQFP logic gates have been demonstrated. In delay-line clocking, the latency between adjacent excitation phases is determined by the propagation delay of excitation currents, and thus the rising time of excitation currents should be sufficiently small; otherwise, an AQFP gate can switch before the previous gate is fully excited. This means that delay-line clocking needs high clock frequencies, because typical excitation currents are sinusoidal and the rising time depends on the frequency. However, AQFP circuits need to be tested in a wide frequency range experimentally. Hence, in the present study, we investigate AQFP circuits adopting delay-line clocking with square excitation currents to apply delay-line clocking in a low frequency range. Square excitation currents have shorter rising time than sinusoidal excitation currents and thus enable low frequency operation. We demonstrate an AQFP buffer chain with delay-line clocking using square excitation currents, in which the latency is approximately 20ps per gate, and confirm that the operating margin for the buffer chain is kept sufficiently wide at clock frequencies below 1GHz, whereas in the sinusoidal case the operating margin shrinks below 500MHz. These results indicate that AQFP circuits adopting delay-line clocking can operate in a low frequency range by using square excitation currents.

  • KBP: Kernel Enhancements for Low-Latency Networking for Virtual Machine and Container without Application Customization Open Access

    Kei FUJIMOTO  Masashi KANEKO  Kenichi MATSUI  Masayuki AKUTSU  

     
    PAPER-Network

      Pubricized:
    2021/10/26
      Vol:
    E105-B No:5
      Page(s):
    522-532

    Packet processing on commodity hardware is a cost-efficient and flexible alternative to specialized networking hardware. However, virtualizing dedicated networking hardware as a virtual machine (VM) or a container on a commodity server results in performance problems, such as longer latency and lower throughput. This paper focuses on obtaining a low-latency networking system in a VM and a container. We reveal mechanisms that cause millisecond-scale networking delays in a VM through a series of experiments. To eliminate such delays, we design and implement a low-latency networking system, kernel busy poll (KBP), which achieves three goals: (1) microsecond-scale tail delays and higher throughput than conventional solutions are achieved in a VM and a container; (2) application customization is not required, so applications can use the POSIX sockets application program interface; and (3) KBP software does not need to be developed for every Linux kernel security update. KBP can be applied to both a VM configuration and a container configuration. Evaluation results indicate that KBP achieves microsecond-scale tail delays in both a VM and a container. In the VM configuration, KBP reduces maximum round-trip latency by more than 98% and increases the throughput by up to three times compared with existing NAPI and Open vSwitch with the Data Plane Development Kit (OvS-DPDK). In the container configuration, KBP reduces maximum round-trip latency by 21% to 96% and increases the throughput by up to 1.28 times compared with NAPI.

  • Metric-Combining Multiuser Detection Using Replica Cancellation with RTS and Enhanced CTS for High-Reliable and Low-Latency Wireless Communications

    Hideya SO  Kazuhiko FUKAWA  Hayato SOYA  Yuyuan CHANG  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2021/06/01
      Vol:
    E104-B No:11
      Page(s):
    1441-1453

    In unlicensed spectrum, wireless communications employing carrier sense multiple access with collision avoidance (CSMA/CA) suffer from longer transmission delay time as the number of user terminals (UTs) increases, because packet collisions are more likely to occur. To cope with this problem, this paper proposes a new multiuser detection (MUD) scheme that uses both request-to-send (RTS) and enhanced clear-to-send (eCTS) for high-reliable and low-latency wireless communications. As in conventional MUD scheme, the metric-combining MUD (MC-MUD) calculates log likelihood functions called metrics and accumulates the metrics for the maximum likelihood detection (MLD). To avoid increasing the number of states for MLD, MC-MUD forces the relevant UTs to retransmit their packets until all the collided packets are correctly detected, which requires a kind of central control and reduces the system throughput. To overcome these drawbacks, the proposed scheme, which is referred to as cancelling MC-MUD (CMC-MUD), deletes replicas of some of the collided packets from the received signals, once the packets are correctly detected during the retransmission. This cancellation enables new UTs to transmit their packets and then performs MLD without increasing the number of states, which improves the system throughput without increasing the complexity. In addition, the proposed scheme adopts RTS and eCTS. One UT that suffers from packet collision transmits RTS before the retransmission. Then, the corresponding access point (AP) transmits eCTS including addresses of the other UTs, which have experienced the same packet collision. To reproduce the same packet collision, these other UTs transmit their packets once they receive the eCTS. Computer simulations under one AP conditions evaluate an average carrier-to-interference ratio (CIR) range in which the proposed scheme is effective, and clarify that the transmission delay time of the proposed scheme is shorter than that of the conventional schemes. In two APs environments that can cause the hidden terminal problem, it is demonstrated that the proposed scheme achieves shorter transmission delay times than the conventional scheme with RTS and conventional CTS.

  • Field Evaluation of 5G Low Latency and High Reliability Vehicle-to-Vehicle Direct Communication for Application to Truck Platooning

    Manabu MIKAMI  Koichi SERIZAWA  Kohei MOTO  Hitoshi YOSHINO  

     
    PAPER

      Pubricized:
    2021/03/17
      Vol:
    E104-B No:9
      Page(s):
    1026-1034

    Fifth generation mobile communication system (5G) mobile operators need to explore new use cases and/or applications together with vertical industries, the industries which are potential users of 5G, in order to fully exploit the new 5G capabilities in terms of its application. Vehicular communications for platooning are considered to be one of new use cases of 5G whose low-latency and ultra-reliability are required. This paper presents our field evaluations on latency and reliability performance of 5G V2V Direct communication towards application to truck platooning. The authors build a field experimental environment, for V2X communications of truck platooning, with actual large-size trucks and a prototype system employing 5G New Radio (NR) technologies, and performed some field experiments in rural areas. In this paper, we introduce the 5G NR-V2X prototype system. Its most distinctive feature is that the prototype system is equipped with V2V Direct communication radio interface (i.e., sidelink), in addition to the traditional radio interfaces between BS and UE (i.e., downlink and uplink). Then, we present the field evaluation results of radio propagation environment results and over-the-air transmission performance of latency and reliability characteristics on the V2V Direct communication of the prototype in real public express highway environment including tunnel area as well as tunnel outside area, in order to assess 5G NR-V2X system applying to truck platooning. The radio propagation and the latency performance evaluation results clarify that the latency performance is degraded due to Hybrid Automatic Repeat reQuest (HARQ) retransmission at the outside of tunnel more possibly than the inside of tunnel, since larger path loss values can be observed at the outside of tunnel than the inside of tunnel, in V2V Direct communications of truck platooning. The over-the-air latency and reliability evaluation results confirm that it is important to set an appropriate maximum number of HARQ retransmissions since there is a trade-off problem in order to realize low latency and high reliability simultaneously.

  • A Novel Multi-AP Diversity for Highly Reliable Transmissions in Wireless LANs

    Toshihisa NABETANI  Masahiro SEKIYA  

     
    PAPER-Terrestrial Wireless Communication/Broadcasting Technologies

      Pubricized:
    2021/01/08
      Vol:
    E104-B No:7
      Page(s):
    913-921

    With the development of the IEEE 802.11 standard for wireless LANs, there has been an enormous increase in the usage of wireless LANs in factories, plants, and other industrial environments. In industrial applications, wireless LAN systems require high reliability for stable real-time communications. In this paper, we propose a multi-access-point (AP) diversity method that contributes to the realization of robust data transmissions toward realization of ultra-reliable low-latency communications (URLLC) in wireless LANs. The proposed method can obtain a diversity effect of multipaths with independent transmission errors and collisions without modification of the IEEE 802.11 standard or increasing overhead of communication resources. We evaluate the effects of the proposed method by numerical analysis, develop a prototype to demonstrate its feasibility, and perform experiments using the prototype in a factory wireless environment. These numerical evaluations and experiments show that the proposed method increases reliability and decreases transmission delay.

  • Design Method of Variable-Latency Circuit with Tunable Approximate Completion-Detection Mechanism

    Yuta UKON  Shimpei SATO  Atsushi TAKAHASHI  

     
    PAPER

      Pubricized:
    2020/12/21
      Vol:
    E104-C No:7
      Page(s):
    309-318

    Advanced information-processing services such as computer vision require a high-performance digital circuit to perform high-load processing at high speed. To achieve high-speed processing, several image-processing applications use an approximate computing technique to reduce idle time of the circuit. However, it is difficult to design the high-speed image-processing circuit while controlling the error rate so as not to degrade service quality, and this technique is used for only a few applications. In this paper, we propose a method that achieves high-speed processing effectively in which processing time for each task is changed by roughly detecting its completion. Using this method, a high-speed processing circuit with a low error rate can be designed. The error rate is controllable, and a circuit design method to minimize the error rate is also presented in this paper. To confirm the effectiveness of our proposal, a ripple-carry adder (RCA), 2-dimensional discrete cosine transform (2D-DCT) circuit, and histogram of oriented gradients (HOG) feature calculation circuit are evaluated. Effective clock periods of these circuits obtained by our method with around 1% error rate are improved about 64%, 6%, and 12%, respectively, compared with circuits without error. Furthermore, the impact of the miscalculation on a video monitoring service using an object detection application is investigated. As a result, more than 99% of detection points required to be obtained are detected, and it is confirmed the miscalculation hardly degrades the service quality.

  • Traffic Reduction Technologies and Data Aggregation Control to Minimize Latency in IoT Systems Open Access

    Hideaki YOSHINO  Kenko OTA  Takefumi HIRAGURI  

     
    INVITED PAPER

      Pubricized:
    2021/02/04
      Vol:
    E104-B No:7
      Page(s):
    706-715

    The spread of the Internet of Things (IoT) has led to the generation of large amounts of data, requiring massive communication, computing, and storage resources. Cloud computing plays an important role in realizing most IoT applications classified as massive machine type communication and cyber-physical control applications in vertical domains. To handle the increasing amount of IoT data, it is important to reduce the traffic concentrated in the cloud by distributing the computing and storage resources to the network edge side and to suppress the latency of the IoT applications. In this paper, we first present a recent literature review on fog/edge computing and data aggregation as representative traffic reduction technologies for efficiently utilizing communication, computing, and storage resources in IoT systems, and then focus on data aggregation control minimizing the latency in an IoT gateway. We then present a unified modeling for statistical and nonstatistical data aggregation and analyze its latency. We analytically derive the Laplace-Stieltjes transform and average of the stationary distribution of the latency and approximate the average latency; we subsequently apply it to an adaptive aggregation number control for the time-variant data arrival. The transient traffic characteristics, that is, the absorption of traffic fluctuations realizing a stable optimal latency, were clarified through a simulation with a time-variant Poisson input and non-Poisson inputs, such as a Beta input, which is a typical IoT traffic model.

1-20hit(105hit)