The search functionality is under construction.

IEICE TRANSACTIONS on Fundamentals

  • Impact Factor

    0.48

  • Eigenfactor

    0.003

  • article influence

    0.1

  • Cite Score

    1.1

Advance publication (published online immediately after acceptance)

Volume E102-A No.12  (Publication Date:2019/12/01)

    Special Section on Information Theory and Its Applications
  • FOREWORD Open Access

    Noboru KUNIHIRO  Yasutada OOHAMA  

     
    FOREWORD

      Page(s):
    1590-1590
  • An Improvement of Non-Binary Single b-Burst of Insertion/Deletion Correcting Code

    Toyohiko SAEKI  Takayuki NOZAKI  

     
    PAPER-Coding Theory

      Page(s):
    1591-1599

    This paper constructs non-binary codes correcting a single b-burst of insertions or deletions with large cardinalities. This paper also provides insertion and deletion correcting algorithms of the constructed codes and evaluates a lower bound of the cardinalities of the constructed codes. Moreover, we evaluate a non-asymptotic upper bound on the cardinalities of arbitrary codes which correct a single b-burst of insertions or deletions.

  • Fast Serial Iterative Decoding Algorithm for Zigzag Decodable Fountain Codes by Efficient Scheduling

    Yoshihiro MURAYAMA  Takayuki NOZAKI  

     
    PAPER-Erasure Correction

      Page(s):
    1600-1610

    Fountain codes are erasure correcting codes realizing reliable communication systems for the multicast on the Internet. The zigzag decodable fountain (ZDF) codes are one of generalization of the Raptor codes, i.e., applying shift operation to generate the output packets. The ZDF codes are decoded by a two-stage iterative decoding algorithm, which combines the packet-wise peeling algorithm and the bit-wise peeling algorithm. By the bit-wise peeling algorithm and shift operation, ZDF codes outperform Raptor codes under iterative decoding in terms of decoding erasure rates and overheads. However, the bit-wise peeling algorithm spends long decoding time. This paper proposes fast bit-wise decoding algorithms for the ZDF codes. Simulation results show that the proposed algorithm drastically reduces the decoding time compared with the previous algorithm.

  • Shifted Coded Slotted ALOHA: A Graph-Based Random Access with Shift Operation

    Tomokazu EMOTO  Takayuki NOZAKI  

     
    PAPER-Erasure Correction

      Page(s):
    1611-1621

    A random access scheme is a fundamental scenario in which the users transmit through a shared channel and cannot coordinate with each other. Recently, successive interference cancellation (SIC) is introduced into the random access scheme. The SIC decodes the transmitted packets using collided packets. The coded slotted ALOHA (CSA) is a random access scheme using the SIC. The CSA encodes each packet by a local code prior to transmission. It is known that the CSA achieves excellent throughput. On the other hand, it is reported that shift operation improves the decoding performance for packet-oriented erasure correcting coding systems. In this paper, we propose a protocol which applies the shift operation to the CSA. Numerical simulations show that the proposed protocol achieves better throughput and packet loss rate than the CSA. Moreover, we analyze the asymptotic behavior of the throughput and the decoding erasure rate for the proposed protocol by the density evolution.

  • Packet-Oriented Erasure Correcting Codes by Bit-Level Shift Operation and Exclusive OR

    Yuta HANAKI  Takayuki NOZAKI  

     
    PAPER-Erasure Correction

      Page(s):
    1622-1630

    This paper constructs packet-oriented erasure correcting codes and their systematic forms for the distributed storage systems. The proposed codes are encoded by exclusive OR and bit-level shift operation. By the shift operation, the encoded packets are slightly longer than the source packets. This paper evaluates the extra length of the encoded packets, called overhead, and shows that the proposed codes have smaller overheads than the zigzag decodable codes, which are existing codes using bit-level shift operation and exclusive OR.

  • Achievable Rate Regions for Source Coding with Delayed Partial Side Information Open Access

    Tetsunao MATSUTA  Tomohiko UYEMATSU  

     
    PAPER-Shannon Theory

      Page(s):
    1631-1641

    In this paper, we consider a source coding with side information partially used at the decoder through a codeword. We assume that there exists a relative delay (or gap) of the correlation between the source sequence and side information. We also assume that the delay is unknown but the maximum of possible delays is known to two encoders and the decoder, where we allow the maximum of delays to change by the block length. In this source coding, we give an inner bound and an outer bound on the achievable rate region, where the achievable rate region is the set of rate pairs of encoders such that the decoding error probability vanishes as the block length tends to infinity. Furthermore, we clarify that the inner bound coincides with the outer bound when the maximum of delays for the block length converges to a constant.

  • Variable-Length Intrinsic Randomness on Two Performance Criteria Based on Variational Distance

    Jun YOSHIZAWA  Shota SAITO  Toshiyasu MATSUSHIMA  

     
    PAPER-Shannon Theory

      Page(s):
    1642-1650

    This paper investigates the problem of variable-length intrinsic randomness for a general source. For this problem, we can consider two performance criteria based on the variational distance: the maximum and average variational distances. For the problem of variable-length intrinsic randomness with the maximum variational distance, we derive a general formula of the average length of uniform random numbers. Further, we derive the upper and lower bounds of the general formula and the formula for a stationary memoryless source. For the problem of variable-length intrinsic randomness with the average variational distance, we also derive a general formula of the average length of uniform random numbers.

  • Hadamard-Type Matrices on Finite Fields and Complete Complementary Codes

    Tetsuya KOJIMA  

     
    PAPER-Sequences

      Page(s):
    1651-1658

    Hadamard matrix is defined as a square matrix where any components are -1 or +1, and where any pairs of rows are mutually orthogonal. In this work, we consider the similar matrix on finite field GF(p) where p is an odd prime. In such a matrix, every component is one of the integers on GF(p){0}, that is, {1,2,...,p-1}. Any additions and multiplications should be executed under modulo p. In this paper, a method to generate such matrices is proposed. In addition, the paper includes the applications to generate n-shift orthogonal sequences and complete complementary codes. The generated complete complementary code is a family of multi-valued sequences on GF(p){0}, where the number of sequence sets, the number of sequences in each sequence set and the sequence length depend on the various divisors of p-1. Such complete complementary codes with various parameters have not been proposed in previous studies.

  • Algebraic Group Structure of the Random Number Generator: Theoretical Analysis of NTU Sequence(s)

    Yuta KODERA  Md. Arshad ALI  Takeru MIYAZAKI  Takuya KUSAKA  Yasuyuki NOGAMI  Satoshi UEHARA  Robert H. MORELOS-ZARAGOZA  

     
    PAPER-Sequences

      Page(s):
    1659-1667

    An algebraic group is an essential mathematical structure for current communication systems and information security technologies. Further, as a widely used technology underlying such systems, pseudorandom number generators have become an indispensable part of their construction. This paper focuses on a theoretical analysis for a series of pseudorandom sequences generated by a trace function and the Legendre symbol over an odd characteristic field. As a consequence, the authors give a theoretical proof that ensures a set of subsequences forms a group with a specific binary operation.

  • A New Combiner for Key Encapsulation Mechanisms

    Goichiro HANAOKA  Takahiro MATSUDA  Jacob C. N. SCHULDT  

     
    PAPER-Cryptography

      Page(s):
    1668-1675

    Key encapsulation mechanism (KEM) combiners, recently formalized by Giacon, Heuer, and Poettering (PKC'18), enable hedging against insecure KEMs or weak parameter choices by combining ingredient KEMs into a single KEM that remains secure assuming just one of the underlying ingredient KEMs is secure. This seems particularly relevant when considering quantum-resistant KEMs which are often based on arguably less well-understood hardness assumptions and parameter choices. We propose a new simple KEM combiner based on a one-time secure message authentication code (MAC) and two-time correlated input secure hash. Instantiating the correlated input secure hash with a t-wise independent hash for an appropriate value of t, yields a KEM combiner based on a strictly weaker additional primitive than the standard model construction of Giaon et al. and furthermore removes the need to do n full passes over the encapsulation, where n is the number of ingredient KEMs, which Giacon et al. highlight as a disadvantage of their scheme. However, unlike Giacon et al., our construction requires the public key of the combined KEM to include a hash key, and furthermore requires a MAC tag to be added to the encapsulation of the combined KEM.

  • Speeding Up Revocable Group Signature with Compact Revocation List Using Vector Commitments

    Yasuyuki SEITA  Toru NAKANISHI  

     
    PAPER-Cryptography

      Page(s):
    1676-1687

    In ID-based user authentications, a privacy problem can occur, since the service provider (SP) can accumulate the user's access history from the user ID. As a solution to that problem, group signatures have been researched. One of important issues in the group signatures is the user revocation. Previously, an efficient revocable scheme with signing/verification of constant complexity was proposed by Libert et al. In this scheme, users are managed by a binary tree, and a list of data for revoked users, called a revocation list (RL), is used for revocation. However, the scheme suffers from the large RL. Recently, an extended scheme has been proposed by Sadiah and Nakanishi, where the RL size is reduced by compressing RL. On the other hand, there is a problem that some overhead occurs in the authentication as a price for reducing the size of RL. In this paper, we propose an extended scheme where the authentication is speeded up by reducing the number of Groth-Sahai (GS) proofs. Furthermore, we implemented it on a PC to show the effectiveness. The verification time is about 30% shorter than that of the previous scheme by Sadiah and Nakanishi.

  • Ternary Convolutional Codes with Optimum Distance Spectrum

    Shungo MIYAGI  Motohiko ISAKA  

     
    LETTER-Coding Theory

      Page(s):
    1688-1690

    This letter presents ternary convolutional codes and their punctured codes with optimum distance spectrum.

  • Optimal Balanced Almost 8-QAM Sequences with Three-Level Autocorrelation

    Fanxin ZENG  Xiping HE  Guixin XUAN  Zhenyu ZHANG  Yanni PENG  Linjie QIAN  Li YAN  

     
    LETTER-Sequences

      Page(s):
    1691-1696

    Based on the number of cyclotomy of order eight, a class of balanced almost 8-QAM sequences with odd prime periods is presented. The resultant sequences have low two-level nontrivial autocorrelation values, and their distribution is determined. Furthermore, the smallest possible absolute sidelobes (SPASs) of autocorrelation functions of balanced almost 8-QAM sequences are derived. Compared with the obtained SPASs, some of the proposed sequences is optimal or suboptimal.

  • 16-QAM Sequences with Good Periodic Autocorrelation Function

    Fanxin ZENG  Yue ZENG  Lisheng ZHANG  Xiping HE  Guixin XUAN  Zhenyu ZHANG  Yanni PENG  Linjie QIAN  Li YAN  

     
    LETTER-Sequences

      Page(s):
    1697-1700

    Sequences that attain the smallest possible absolute sidelobes (SPASs) of periodic autocorrelation function (PACF) play fairly important roles in synchronization of communication systems, Large scale integrated circuit testing, and so on. This letter presents an approach to construct 16-QAM sequences of even periods, based on the known quaternary sequences. A relationship between the PACFs of 16-QAM and quaternary sequences is established, by which when quaternary sequences that attain the SPASs of PACF are employed, the proposed 16-QAM sequences have good PACF.

  • Special Section on VLSI Design and CAD Algorithms
  • FOREWORD Open Access

    Toshiyuki SHIBUYA  

     
    FOREWORD

      Page(s):
    1701-1701
  • Passage of Faulty Nodes: A Novel Approach for Fault-Tolerant Routing on NoCs

    Yota KUROKAWA  Masaru FUKUSHI  

     
    PAPER

      Page(s):
    1702-1710

    This paper addresses the problem of developing an efficient fault-tolerant routing method for 2D mesh Network-on-Chips (NoCs) to realize dependable and high performance many core systems. Existing fault-tolerant routing methods have two critical problems of high communication latency and low node utilization. Unlike almost all existing methods where packets always detour faulty nodes, we propose a novel and unique approach that packets can pass through faulty nodes. For this approach, we enhance the common NoC architecture by adding switches and links around each node and propose a fault-tolerant routing method with no virtual channels based on the well-known simple XY routing method. Simulation results show that the proposed method reduces average communication latency by about 97.1% compared with the existing method, without sacrificing fault-free nodes.

  • Selective Use of Stitch-Induced Via for V0 Mask Reduction: Standard Cell Design and Placement Optimization

    Daijoon HYUN  Younggwang JUNG  Youngsoo SHIN  

     
    PAPER

      Page(s):
    1711-1719

    Multiple patterning lithography allows fine patterns beyond lithography limit, but it suffers from a large process cost. In this paper, we address a method to reduce the number of V0 masks; it consists of two sub-problems. First, stitch-induced via (SIV) is introduced to reduce the number of V0 masks. It involves the redesign of standard cells to replace some vias in V0 layer with SIVs, such that the remaining vias can be assigned to the reduced masks. Since SIV formation requires metal stitches in different masks, SIV replacement and metal mask assignment should be solved simultaneously. This sub-problem is formulated as integer linear programming (ILP). In the second sub-problem, inter-row via conflict aware detailed placement is addressed. Single row placement optimization is performed for each row to remove metal and inter-row via conflicts, while minimizing cell displacements. Since it is time consuming to consider many cell operations at once, we apply a few operations iteratively, where different operations are applied to each iteration and to each cell depending on whether the cell has a conflict in the previous iteration. Remaining conflicts are then removed by mapping conflict cells to white spaces. To this end, we minimize the number of cells to move and maximize the number of large white spaces before mapping. Experimental results demonstrate that the cell placement with two V0 masks is completed by proposed methods, with 7 times speedup and 21% reduction in total cell displacement, compared to conventional detailed placement.

  • Density Optimization for Analog Layout Based on Transistor-Array

    Chao GENG  Bo LIU  Shigetoshi NAKATAKE  

     
    PAPER

      Page(s):
    1720-1730

    In integrated circuit design of advanced technology nodes, layout density uniformity significantly influences the manufacturability due to the CMP variability. In analog design, especially, designers are suffering from passing the density checking since there are few useful tools. To tackle this issue, we focus a transistor-array(TA)-style analog layout, and propose a density optimization algorithm consistent with complicated design rules. Based on TA-style, we introduce a density-aware layout format to explicitly control the layout pattern density, and provide the mathematical optimization approach. Hence, a design flow incorporating our density optimization can drastically reduce the design time with fewer iterations. In a design case of an OPAMP layout in a 65nm CMOS process, the result demonstrates that the proposed approach achieves more than 48× speed-up compared with conventional manual layout, meanwhile it shows a good circuit performance in the post-layout simulation.

  • Authenticated-Encrypted Analog-to-Digital Conversion Based on Non-Linearity and Redundancy Transformation

    Vinod V. GADDE  Makoto IKEDA  

     
    PAPER

      Page(s):
    1731-1740

    We have proposed a generic architecture that can integrate the aspects of confidentiality and integrity into the A/D conversion framework. A conceptual account of the development of the proposed architecture is presented. Using the principle of this architecture we have presented a CMOS circuit design to facilitate a fully integrated Authenticated-Encrypted ADC (AE-ADC). We have implemented and demonstrated a partial 8-bit ADC Analog Front End of this proposed circuit in 0.18µm CMOS with an ENOB of 7.64 bits.

  • On-Chip Cache Architecture Exploiting Hybrid Memory Structures for Near-Threshold Computing

    Hongjie XU  Jun SHIOMI  Tohru ISHIHARA  Hidetoshi ONODERA  

     
    PAPER

      Page(s):
    1741-1750

    This paper focuses on power-area trade-off axis to memory systems. Compared with the power-performance-area trade-off application on the traditional high performance cache, this paper focuses on the edge processing environment which is becoming more and more important in the Internet of Things (IoT) era. A new power-oriented trade-off is proposed for on-chip cache architecture. As a case study, this paper exploits a good energy efficiency of Standard-Cell Memory (SCM) operating in a near-threshold voltage region and a good area efficiency of Static Random Access Memory (SRAM). A hybrid 2-level on-chip cache structure is first introduced as a replacement of 6T-SRAM cache as L0 cache to save the energy consumption. This paper proposes a method for finding the best capacity combination for SCM and SRAM, which minimizes the energy consumption of the hybrid cache under a specific cache area constraint. The simulation result using a 65-nm process technology shows that up to 80% energy consumption is reduced without increasing the die area by replacing the conventional SRAM instruction cache with the hybrid 2-level cache. The result shows that energy consumption can be reduced if the area constraint for the proposed hybrid cache system is less than the area which is equivalent to a 8kB SRAM. If the target operating frequency is less than 100MHz, energy reduction can be achieved, which implies that the proposed cache system is suitable for low-power systems where a moderate processing speed is required.

  • Methods for Reducing Power and Area of BDD-Based Optical Logic Circuits

    Ryosuke MATSUO  Jun SHIOMI  Tohru ISHIHARA  Hidetoshi ONODERA  Akihiko SHINYA  Masaya NOTOMI  

     
    PAPER

      Page(s):
    1751-1759

    Optical circuits using nanophotonic devices attract significant interest due to its ultra-high speed operation. As a consequence, the synthesis methods for the optical circuits also attract increasing attention. However, existing methods for synthesizing optical circuits mostly rely on straight-forward mappings from established data structures such as Binary Decision Diagram (BDD). The strategy of simply mapping a BDD to an optical circuit sometimes results in an explosion of size and involves significant power losses in branches and optical devices. To address these issues, this paper proposes a method for reducing the size of BDD-based optical logic circuits exploiting wavelength division multiplexing (WDM). The paper also proposes a method for reducing the number of branches in a BDD-based circuit, which reduces the power dissipation in laser sources. Experimental results obtained using a partial product accumulation circuit used in a 4-bit parallel multiplier demonstrates significant advantages of our method over existing approaches in terms of area and power consumption.

  • A Low Area Overhead Design Method for High-Performance General-Synchronous Circuits with Speculative Execution

    Shimpei SATO  Eijiro SASSA  Yuta UKON  Atsushi TAKAHASHI  

     
    PAPER

      Page(s):
    1760-1769

    In order to obtain high-performance circuits in advanced technology nodes, design methodology has to take the existence of large delay variations into account. Clock scheduling and speculative execution have overheads to realize them, but have potential to improve the performance by averaging the imbalance of maximum delay among paths and by utilizing valid data available earlier than worst-case scenarios, respectively. In this paper, we propose a high-performance digital circuit design method with speculative executions with less overhead by utilizing clock scheduling with delay insertions effectively. The necessity of speculations that cause overheads is effectively reduced by clock scheduling with delay insertion. Experiments show that a generated circuit achieves 26% performance improvement with 1.3% area overhead compared to a circuit without clock scheduling and without speculative execution.

  • Signal Selection Methods for Debugging Gate-Level Sequential Circuits

    Yusuke KIMURA  Amir Masoud GHAREHBAGHI  Masahiro FUJITA  

     
    PAPER

      Page(s):
    1770-1780

    This paper introduces methods to modify a buggy sequential gate-level circuit to conform to the specification. In order to preserve the optimization efforts, the modifications should be as small as possible. Assuming that the locations to be modified are given, our proposed method finds an appropriate set of fan-in signals for the patch function of those locations by iteratively calculating the state correspondence between the specification and the buggy circuit and applying a method for debugging combinational circuits. The experiments are conducted on ITC99 benchmark circuits, and it is shown that our proposed method can work when there are at most 30,000 corresponding reachable state pairs between two circuits. Moreover, a heuristic method using the information of data-path FFs is proposed, which can find a correct set of fan-ins for all the benchmark circuits within practical time.

  • Design of Low-Cost Approximate Multipliers Based on Probability-Driven Inexact Compressors

    Yi GUO  Heming SUN  Ping LEI  Shinji KIMURA  

     
    PAPER

      Page(s):
    1781-1791

    Approximate computing has emerged as a promising approach for error-tolerant applications to improve hardware performance at the cost of some loss of accuracy. Multiplication is a key arithmetic operation in these applications. In this paper, we propose a low-cost approximate multiplier design by employing new probability-driven inexact compressors. This compressor design is introduced to reduce the height of partial product matrix into two rows, based on the probability distribution of the sum result of partial products. To compensate the accuracy loss of the multiplier, a grouped error recovery scheme is proposed and achieves different levels of accuracy. In terms of mean relative error distance (MRED), the accuracy losses of the proposed multipliers are from 1.07% to 7.86%. Compared with the Wallace multiplier using 40nm process, the most accurate variant of the proposed multipliers can reduce power by 59.75% and area by 42.47%. The critical path delay reduction is larger than 12.78%. The proposed multiplier design has a better accuracy-performance trade-off than other designs with comparable accuracy. In addition, the efficiency of the proposed multiplier design is assessed in an image processing application.

  • High Performance Application Specific Stream Architecture for Hardware Acceleration of HOG-SVM on FPGA

    Piyumal RANAWAKA  Mongkol EKPANYAPONG  Adriano TAVARES  Mathew DAILEY  Krit ATHIKULWONGSE  Vitor SILVA  

     
    PAPER

      Page(s):
    1792-1803

    Conventional sequential processing on software with a general purpose CPU has become significantly insufficient for certain heavy computations due to the high demand of processing power to deliver adequate throughput and performance. Due to many reasons a high degree of interest could be noted for high performance real time video processing on embedded systems. However, embedded processing platforms with limited performance could least cater the processing demand of several such intensive computations in computer vision domain. Therefore, hardware acceleration could be noted as an ideal solution where process intensive computations could be accelerated using application specific hardware integrated with a general purpose CPU. In this research we have focused on building a parallelized high performance application specific architecture for such a hardware accelerator for HOG-SVM computation implemented on Zynq 7000 FPGA. Histogram of Oriented Gradients (HOG) technique combined with a Support Vector Machine (SVM) based classifier is versatile and extremely popular in computer vision domain in contrast to high demand for processing power. Due to the popularity and versatility, various previous research have attempted on obtaining adequate throughput on HOG-SVM. This research with a high throughput of 240FPS on single scale on VGA frames of size 640x480 out performs the best case performance on a single scale of previous research by approximately a factor of 3-4. Further it's an approximately 15x speed up over the GPU accelerated software version with the same accuracy. This research has explored the possibility of using a novel architecture based on deep pipelining, parallel processing and BRAM structures for achieving high performance on the HOG-SVM computation. Further the above developed (video processing unit) VPU which acts as a hardware accelerator will be integrated as a co-processing peripheral to a host CPU using a novel custom accelerator structure with on chip buses in a System-On-Chip (SoC) fashion. This could be used to offload the heavy video stream processing redundant computations to the VPU whereas the processing power of the CPU could be preserved for running light weight applications. This research mainly focuses on the architectural techniques used to achieve higher performance on the hardware accelerator and on the novel accelerator structure used to integrate the accelerator with the host CPU.

  • Reconfigurable 3D Sound Processor and Its Automatic Design Environment Using High-Level Synthesis

    Saya OHIRA  Naoki TSUCHIYA  Tetsuya MATSUMURA  

     
    PAPER

      Page(s):
    1804-1812

    We propose a three-dimensional (3D) sound processor architecture that includes super-directional modulation intellectual property (IP) and 3D sound processing IP and for consumer applications. In addition, we also propose an automatic design environment for 3D sound processing IP. This processor can generate realistic small sound fields in arbitrary spaces using ultrasound. In particular, in the 3D sound processing IP, in order to reproduce 3D audio, it is necessary to reproduce the personal frequency characteristics of complex head related transfer functions. For this reason, we have constructed an automatic design environment with high reconfigurability. This automatic design environment is based on high-level synthesis, and it is possible to automatically generate a C-based algorithm simulator and automatically synthesize the IP hardware by inputting a parameter description file for filter design. This automatic design environment can reduce the design period to approximately 1/5 as compared with conventional manual design. Applying the automatic design environment, a 3D sound processing IP was designed experimentally. The designed IP can be sufficiently applied to consumer applications from the viewpoints of hardware amount and power consumption.

  • Implementation and Area Optimization of LUT6 Based Convolution Structure on FPGA

    Huangtao WU  Wenjin HUANG  Rui CHEN  Yihua HUANG  

     
    LETTER

      Page(s):
    1813-1815

    To implement the parallel acceleration of convolution operation of Convolutional Neural Networks (CNNs) on field programmable gate array (FPGA), large quantities of the logic resources will be consumed, expecially DSP cores. Many previous researches fail to make a well balance between DSP and LUT6. For better resource efficiency, a typical convolution structure is implemented with LUT6s in this paper. Besides, a novel convolution structure is proposed to further reduce the LUT6 resource consumption by modifying the typical convolution structure. The equations to evaluate the LUT6 resource consumptions of both structures are presented and validated. The theoretical evaluation and experimental results show that the novel structure can save 3.5-8% of LUT6s compared with the typical structure.

  • Special Section on Smart Multimedia & Communication Systems
  • FOREWORD Open Access

    Hirokazu TANAKA  

     
    FOREWORD

      Page(s):
    1816-1816
  • Detecting Surface Defects of Wind Tubine Blades Using an Alexnet Deep Learning Algorithm Open Access

    Xiao-Yi ZHAO  Chao-Yi DONG  Peng ZHOU  Mei-Jia ZHU  Jing-Wen REN  Xiao-Yan CHEN  

     
    PAPER-Machine Learning

      Page(s):
    1817-1824

    The paper employed an Alexnet, which is a deep learning framework, to automatically diagnose the damages of wind power generator blade surfaces. The original images of wind power generator blade surfaces were captured by machine visions of a 4-rotor UAV (unmanned aerial vehicle). Firstly, an 8-layer Alexnet, totally including 21 functional sub-layers, is constructed and parameterized. Secondly, the Alexnet was trained with 10000 images and then was tested by 6-turn 350 images. Finally, the statistic of network tests shows that the average accuracy of damage diagnosis by Alexnet is about 99.001%. We also trained and tested a traditional BP (Back Propagation) neural network, which have 20-neuron input layer, 5-neuron hidden layer, and 1-neuron output layer, with the same image data. The average accuracy of damage diagnosis of BP neural network is 19.424% lower than that of Alexnet. The point shows that it is feasible to apply the UAV image acquisition and the deep learning classifier to diagnose the damages of wind turbine blades in service automatically.

  • Precoder and Postcoder Design for Wireless Video Streaming with Overloaded Multiuser MIMO-OFDM Systems

    Koji TASHIRO  Masayuki KUROSAKI  Hiroshi OCHI  

     
    PAPER-Digital Signal Processing

      Page(s):
    1825-1833

    Mobile video traffic is expected to increase explosively because of the proliferating number of Wi-Fi terminals. An overloaded multiple-input multiple-output (MIMO) technique allows the receiver to implement smaller number of antennas than the transmitter in exchange for degradation in video quality and a large amount of computational complexity for postcoding at the receiver side. This paper proposes a novel linear precoder for high-quality video streaming in overloaded multiuser MIMO systems, which protects visually significant portions of a video stream. A low complexity postcoder is also proposed, which detects some of data symbols by linear detection and the others by a prevoting vector cancellation (PVC) approach. It is shown from simulation results that the combination use of the proposed precoder and postcoder achieves higher-quality video streaming to multiple users in a wider range of signal-to-noise ratio (SNR) than a conventional unequal error protection scheme. The proposed precoder attains 40dB in peak signal-to-noise ratio even in poor channel conditions such as the SNR of 12dB. In addition, due to the stepwise acquisition of data symbols by means of linear detection and PVC, the proposed postcoder reduces the number of complex additions by 76% and that of multiplications by 64% compared to the conventional PVC.

  • Frequency Efficient Subcarrier Spacing in Multicarrier Backscatter Sensors System Open Access

    Jin MITSUGI  Yuki SATO  Yuusuke KAWAKITA  Haruhisa ICHIKAWA  

     
    PAPER-Digital Signal Processing

      Page(s):
    1834-1841

    Backscatter wireless communications offer advantages such as batteryless operations, small form factor, and radio regulatory exemption sensors. The major challenge ahead of backscatter wireless communications is synchronized multicarrier data collection, which can be realized by rejecting mutual harmonics among backscatters. This paper analyzes the mutual interferences of digitally modulated multicarrier backscatter to find interferences from higher frequency subcarriers to lower frequency subcarriers, which do not take place in analog modulated multicarrier backscatters, is harmful for densely populated subcarriers. This reverse interference distorts the harmonics replica, deteriorating the performance of the existing method, which rejects mutual interference among subcarriers by 5dB processing gain. To solve this problem, this paper analyzes the relationship between subcarrier spacing and reverse interference, and reveals that an alternate channel spacing, with channel separation twice the bandwidth of a subcarrier, can provide reasonably dense subcarrier allocation and can alleviate reverse interference. The idea is examined with prototype sensors in a wired experiment and in an indoor propagation experiment. The results reveal that with alternate channel spacing, the reverse interference practically becomes negligible, and the existing interference rejection method achieves the original processing gain of 5dB with one hundredth packet error rate reduction.

  • Two-Layer Near-Lossless HDR Coding Using Zero-Skip Quantization with Backward Compatibility to JPEG

    Hiroyuki KOBAYASHI  Osamu WATANABE  Hitoshi KIYA  

     
    PAPER-Image

      Page(s):
    1842-1848

    We propose an efficient two-layer near-lossless coding method using an extended histogram packing technique with backward compatibility to the legacy JPEG standard. The JPEG XT, which is the international standard to compress HDR images, adopts a two-layer coding method for backward compatibility to the legacy JPEG standard. However, there are two problems with this two-layer coding method. One is that it does not exhibit better near-lossless performance than other methods for HDR image compression with single-layer structure. The other problem is that the determining the appropriate values of the coding parameters may be required for each input image to achieve good compression performance of near-lossless compression with the two-layer coding method of the JPEG XT. To solve these problems, we focus on a histogram-packing technique that takes into account the histogram sparseness of HDR images. We used zero-skip quantization, which is an extension of the histogram-packing technique proposed for lossless coding, for implementing the proposed near-lossless coding method. The experimental results indicate that the proposed method exhibits not only a better near-lossless compression performance than that of the two-layer coding method of the JPEG XT, but also there are no issue regarding the combination of parameter values without losing backward compatibility to the JPEG standard.

  • Privacy-Preserving Support Vector Machine Computing Using Random Unitary Transformation

    Takahiro MAEKAWA  Ayana KAWAMURA  Takayuki NAKACHI  Hitoshi KIYA  

     
    PAPER-Image

      Page(s):
    1849-1855

    A privacy-preserving support vector machine (SVM) computing scheme is proposed in this paper. Cloud computing has been spreading in many fields. However, the cloud computing has some serious issues for end users, such as the unauthorized use of cloud services, data leaks, and privacy being compromised. Accordingly, we consider privacy-preserving SVM computing. We focus on protecting visual information of images by using a random unitary transformation. Some properties of the protected images are discussed. The proposed scheme enables us not only to protect images, but also to have the same performance as that of unprotected images even when using typical kernel functions such as the linear kernel, radial basis function (RBF) kernel and polynomial kernel. Moreover, it can be directly carried out by using well-known SVM algorithms, without preparing any algorithms specialized for secure SVM computing. In an experiment, the proposed scheme is applied to a face-based authentication algorithm with SVM classifiers to confirm the effectiveness.

  • An Image Fusion Scheme for Single-Shot High Dynamic Range Imaging with Spatially Varying Exposures

    Chihiro GO  Yuma KINOSHITA  Sayaka SHIOTA  Hitoshi KIYA  

     
    PAPER-Image

      Page(s):
    1856-1864

    This paper proposes a novel multi-exposure image fusion (MEF) scheme for single-shot high dynamic range imaging with spatially varying exposures (SVE). Single-shot imaging with SVE enables us not only to produce images without color saturation regions from a single-shot image, but also to avoid ghost artifacts in the producing ones. However, the number of exposures is generally limited to two, and moreover it is difficult to decide the optimum exposure values before the photographing. In the proposed scheme, a scene segmentation method is applied to input multi-exposure images, and then the luminance of the input images is adjusted according to both of the number of scenes and the relationship between exposure values and pixel values. The proposed method with the luminance adjustment allows us to improve the above two issues. In this paper, we focus on dual-ISO imaging as one of single-shot imaging. In an experiment, the proposed scheme is demonstrated to be effective for single-shot high dynamic range imaging with SVE, compared with conventional MEF schemes with exposure compensation.

  • A Hue-Preserving Tone Mapping Scheme Based on Constant-Hue Plane Without Gamut Problem

    Yuma KINOSHITA  Kouki SEO  Artit VISAVAKITCHAROEN  Hitoshi KIYA  

     
    PAPER-Image

      Page(s):
    1865-1871

    We propose a novel hue-preserving tone mapping scheme. Various tone mapping operations have been studied so far, but there are very few works on color distortion caused in image tone mapping. First, LDR images produced from HDR ones by using conventional tone mapping operators (TMOs) are pointed out to have some distortion in hue values due to clipping and rounding quantization processing. Next,we propose a novel method which allows LDR images to have the same maximally saturated color values as those of HDR ones. Generated LDR images by the proposed method have smaller hue degradation than LDR ones generated by conventional TMOs. Moreover, the proposed method is applicable to any TMOs. In an experiment, the proposed method is demonstrated not only to produce images with small hue degradation but also to maintain well-mapped luminance, in terms of three objective metrics: TMQI, hue value in CIEDE2000, and the maximally saturated color on the constant-hue plane in the RGB color space.

  • Adaptive-Partial Template Update with Center-Shifting Recovery for High Frame Rate and Ultra-Low Delay Deformation Matching

    Songlin DU  Yuhao XU  Tingting HU  Takeshi IKENAGA  

     
    PAPER-Image

      Page(s):
    1872-1881

    High frame rate and ultra-low delay matching system plays an important role in various human-machine interactive applications, which demands better performance in matching deformable and out-of-plane rotating objects. Although many algorithms have been proposed for deformation tracking and matching, few of them are suitable for hardware implementation due to complicated operations and large time consumption. This paper proposes a hardware-oriented template update and recovery method for high frame rate and ultra-low delay deformation matching system. In the proposed method, the new template is generated in real time by partially updating the template descriptor and adding new keypoints simultaneously with the matching process in pixels (proposal #1), which avoids the large inter-frame delay. The size and shape of region of interest (ROI) are made flexible and the Hamming threshold used for brute-force matching is adjusted according to pixel position and the flexible ROI (proposal #2), which solves the problem of template drift. The template is recovered by the previous one with a relative center-shifting vector when it is judged as lost via region-wise difference check (proposal #3). Evaluation results indicate that the proposed method successfully achieves the real-time processing of 784fps at the resolution of 640×480 on field-programmable gate array (FPGA), with a delay of 0.808ms/frame, as well as achieves satisfactory deformation matching results in comparison with other general methods.

  • Representative Spatial Selection and Temporal Combination for 60fps Real-Time 3D Tracking of Twelve Volleyball Players on GPU

    Xina CHENG  Yiming ZHAO  Takeshi IKENAGA  

     
    PAPER-Image

      Page(s):
    1882-1890

    Real-time 3D players tracking plays an important role in sports analysis, especially for the live services of sports broadcasting, which have a strict limitation on processing time. For these kinds of applications, 3D trajectories of players contribute to high-level game analysis such as tactic analysis and commercial applications such as TV contents. Thus real-time implementation for 3D players tracking is expected. In order to achieve real-time for 60fps videos with high accuracy, (that means the processing time should be less than 16.67ms per frame), the factors that limit the processing time of target algorithm include: 1) Large image area of each player. 2) Repeated processing of multiple players in multiple views. 3) Complex calculation of observation algorithm. To deal with the above challenges, this paper proposes a representative spatial selection and temporal combination based real-time implementation for multi-view volleyball players tracking on the GPU device. First, the representative spatial pixel selection, which detects the pixels that mostly represent one image region to scale down the image spatially, reduces the number of processing pixels. Second, the representative temporal likelihood combination shares observation calculation by using the temporal correlation between images so that the times of complex calculation is reduced. The experiments are based on videos of the Final and Semi-Final Game of 2014 Japan Inter High School Games of Men's Volleyball in Tokyo Metropolitan Gymnasium. On the GPU device GeForce GTX 1080Ti, the tracking system achieves real-time on 60fps videos and keeps the tracking accuracy higher than 97%.

  • 3D Global and Multi-View Local Features Combination Based Qualitative Action Recognition for Volleyball Game Analysis

    Xina CHENG  Yang LIU  Takeshi IKENAGA  

     
    PAPER-Image

      Page(s):
    1891-1899

    Volleyball video analysis plays important roles in providing data for TV contents and developing strategies. Among all the topics of volleyball analysis, qualitative player action recognition is essential because it potentially provides not only the action that being performed but also the quality, which means how well the action is performed. However, most action recognition researches focus on the discrimination between different actions. The quality of an action, which is helpful for evaluation and training of the player skill, has only received little attention so far. The vital problems in qualitative action recognition include occlusion, small inter-class difference and various kinds of appearance caused by the player change. This paper proposes a 3D global and multi-view local features combination based recognition framework with global team formation feature, ball state feature and abrupt pose features. The above problems are solved by the combination of 3D global features (which hide the unstable and incomplete 2D motion feature caused by occlusion) and the multi-view local features (which get detailed local motion features of body parts in multiple viewpoints). Firstly, the team formation extracts the 3D trajectories from the whole team members rather than a single target player. This proposal focuses more on the entire feature while eliminating the personal effect. Secondly, the ball motion state feature extracts features from the 3D ball trajectory. The ball motion is not affected by the personal appearance, so this proposal ignores the influence of the players appearance and makes it more robust to target player change. At last, the abrupt pose feature consists of two parts: the abrupt hit frame pose (which extracts the contour shape of the player's pose at the hit time) and abrupt pose variation (which extracts the pose variation between the preparation pose and ending pose during the action). These two features make difference of each action quality more distinguishable by focusing on the motion standard and stability between different quality actions. Experiments are conducted on game videos from the Semifinal and Final Game of 2014 Japan Inter High School Games of Men's Volleyball in Tokyo Metropolitan Gymnasium. The experimental results show the accuracy achieves 97.26%, improving 11.33% for action discrimination and 91.76%, and improving 13.72% for action quality evaluation.

  • Video Search Reranking with Relevance Feedback Using Visual and Textual Similarities

    Takamasa FUJII  Soh YOSHIDA  Mitsuji MUNEYASU  

     
    PAPER-Multimedia Environment Technology

      Page(s):
    1900-1909

    In video search reranking, in addition to the well-known semantic gap, the intent gap, which is the gap between the representation of the users' demand and the real search intention, is becoming a major problem restricting the improvement of reranking performance. To address this problem, we propose video search reranking based on a semantic representation by multiple tags. In the proposed method, we use relevance feedback, which the user can interact with by specifying some example videos from the initial search results. We apply the relevance feedback to reduce the gap between the real intent of the users and the video search results. In addition, we focus on the fact that multiple tags are used to represent video contents. By vectorizing multiple tags associated with videos on the basis of the Word2Vec algorithm and calculating the centroid of the tag vector as a collective representation, we can evaluate the semantic similarity between videos by using tag features. We conduct experiments on the YouTube-8M dataset, and the results show that our reranking approach is effective and efficient.

  • Sparse Time-Varying Complex AR (TV-CAR) Speech Analysis Based on Adaptive LASSO

    Keiichi FUNAKI  

     
    LETTER-Speech and Hearing

      Page(s):
    1910-1914

    Linear Prediction (LP) analysis is commonly used in speech processing. LP is based on Auto-Regressive (AR) model and it estimates the AR model parameter from signals with l2-norm optimization. Recently, sparse estimation is paid attention since it can extract significant features from big data. The sparse estimation is realized by l1 or l0-norm optimization or regularization. Sparse LP analysis methods based on l1-norm optimization have been proposed. Since excitation of speech is not white Gaussian, a sparse LP estimation can estimate more accurate parameter than the conventional l2-norm based LP. These are time-invariant and real-valued analysis. We have been studied Time-Varying Complex AR (TV-CAR) analysis for an analytic signal and have evaluated the performance on speech processing. The TV-CAR methods are l2-norm methods. In this paper, we propose the sparse TV-CAR analysis based on adaptive LASSO (Least absolute shrinkage and selection operator) that is l1-norm regularization and evaluate the performance on F0 estimation of speech using IRAPT (Instantaneous RAPT). The experimental results show that the sparse TV-CAR methods perform better for a high level of additive Pink noise.

  • Impulse Noise Removal of Digital Image Considering Local Line Structure

    Shi BAO  Go TANAKA  

     
    LETTER-Image

      Page(s):
    1915-1919

    For the impulse noise removal from a digital image, most of existing methods cannot repair line structures in an input image. In this letter, a method which considers the local line structure is proposed. In order to judge the direction of the line structure, adjacent lines are considered. The effectiveness of the proposed filter is shown by experiments.

  • Image Regularization with Total Variation and Optimized Morphological Gradient Priors

    Shoya OOHARA  Mitsuji MUNEYASU  Soh YOSHIDA  Makoto NAKASHIZUKA  

     
    LETTER-Image

      Page(s):
    1920-1924

    For image restoration, an image prior that is obtained from the morphological gradient has been proposed. In the field of mathematical morphology, the optimization of the structuring element (SE) used for this morphological gradient using a genetic algorithm (GA) has also been proposed. In this paper, we introduce a new image prior that is the sum of the morphological gradients and total variation for an image restoration problem to improve the restoration accuracy. The proposed image prior makes it possible to almost match the fitness to a quantitative evaluation such as the mean square error. It also solves the problem of the artifact due to the unsuitability of the SE for the image. An experiment shows the effectiveness of the proposed image restoration method.

  • Generating Accurate Candidate Windows by Effective Receptive Field

    Baojun ZHAO  Boya ZHAO  Linbo TANG  Baoxian WANG  

     
    LETTER-Image

      Page(s):
    1925-1927

    Towards involving the convolutional neural networks into the object detection field, many computer vision tasks have achieved favorable successes. In order to adapt targets with various scales, deep feature pyramid is widely used, since the traditional object detection methods detect different objects in Gaussian image pyramid. However, due to the mismatching between the anchors and the feature distributions of targets, the accurate detection for targets with various scales is still a challenge. Considering the differences between the theoretical receptive field and effective receptive field, we propose a novel anchor generation method, which takes the effective receptive field as the standard. The proposed method is evaluated on the PASCAL VOC dataset and shows the favorable results.

  • Wireless Power Transfer in the Radiative Near-Field Using a Novel Reconfigurable Holographic Metasurface Aperture Open Access

    Wenyu LUO  

     
    LETTER-Power Transmission

      Page(s):
    1928-1931

    In this letter, we propose a novel wireless power transfer (WPT) scheme in the radiative near-field (Fresnel) region, which based on machine vision and dynamically reconfigurable holographic metasurface aperture capable of focusing power to multiple spots simultaneously without any information feedback. The states of metamaterial elements, formed by tunable meander line resonators, is determined using holographic design principles, in which the interference pattern of reference mode and the desired radiated field pattern leads to the required phase distribution over the surface of the aperture. The three-dimensional position information of mobile point sources is determined by machine visual localization, which can be used to obtain the aperture field. In contrast to the existing research studies, the proposed scheme is not only designed to achieve free multi-focuses, but also with machine vision, low-dimensionality, high transmission efficiency, real-time continuous reconfigurability and so on. The accuracy of the analysis is confirmed using numerical simulation.

  • Regular Section
  • Acoustic Design Support System of Compact Enclosure for Smartphone Using Deep Neural Network

    Kai NAKAMURA  Kenta IWAI  Yoshinobu KAJIKAWA  

     
    PAPER-Engineering Acoustics

      Page(s):
    1932-1939

    In this paper, we propose an automatic design support system for compact acoustic devices such as microspeakers inside smartphones. The proposed design support system outputs the dimensions of compact acoustic devices with the desired acoustic characteristic. This system uses a deep neural network (DNN) to obtain the relationship between the frequency characteristic of the compact acoustic device and its dimensions. The training data are generated by the acoustic finite-difference time-domain (FDTD) method so that many training data can be easily obtained. We demonstrate the effectiveness of the proposed system through some comparisons between desired and designed frequency characteristics.

  • A Novel Three-Point Windowed Interpolation DFT Method for Frequency Measurement of Real Sinusoid Signal

    Kai WANG  Yiting GAO  Lin ZHOU  

     
    PAPER-Digital Signal Processing

      Page(s):
    1940-1945

    The windowed interpolation DFT methods have been utilized to estimate the parameters of a single frequency and multi-frequency signal. Nevertheless, they do not work well for the real-valued sinusoids with closely spaced positive- and negative- frequency. In this paper, we describe a novel three-point windowed interpolation DFT method for frequency measurement of real-valued sinusoid signal. The exact representation of the windowed DFT with maximum sidelobe decay window (MSDW) is constructed. The spectral superposition of positive- and negative-frequency is considered and calculated to improve the estimation performance. The simulation results match with the theoretical values well. In addition, computer simulations demonstrate that the proposed algorithm provides high estimation accuracy and good noise suppression capability.

  • New Sub-Band Adaptive Volterra Filter for Identification of Loudspeaker

    Satoshi KINOSHITA  Yoshinobu KAJIKAWA  

     
    PAPER-Digital Signal Processing

      Page(s):
    1946-1955

    Adaptive Volterra filters (AVFs) are usually used to identify nonlinear systems, such as loudspeaker systems, and ordinary adaptive algorithms can be used to update the filter coefficients of AVFs. However, AVFs require huge computational complexity even if the order of the AVF is constrained to the second order. Improving calculation efficiency is therefore an important issue for the real-time implementation of AVFs. In this paper, we propose a novel sub-band AVF with high calculation efficiency for second-order AVFs. The proposed sub-band AVF consists of four parts: input signal transformation for a single sub-band AVF, tap length determination to improve calculation efficiency, switching the number of sub-bands while maintaining the estimation accuracy, and an automatic search for an appropriate number of sub-bands. The proposed sub-band AVF can improve calculation efficiency for which the dominant nonlinear components are concentrated in any frequency band, such as loudspeakers. A simulation result demonstrates that the proposed sub-band AVF can realize higher estimation accuracy than conventional efficient AVFs.

  • Analysis and Investigation of Frame Invariance and Particle Behavior for Piecewise-Linear Particle Swarm Optimizer

    Tomoyuki SASAKI  Hidehiro NAKANO  

     
    PAPER-Nonlinear Problems

      Page(s):
    1956-1967

    Particle swarm optimization (PSO) is a swarm intelligence algorithm and has good search performance and simplicity in implementation. Because of its properties, PSO has been applied to various optimization problems. However, the search performance of the classical PSO (CPSO) depends on reference frame of solution spaces for each objective function. CPSO is an invariant algorithm through translation and scale changes to reference frame of solution spaces but is a rotationally variant algorithm. As such, the search performance of CPSO is worse in solving rotated problems than in solving non-rotated problems. In the reference frame invariance, the search performance of an optimization algorithm is independent on rotation, translation, or scale changes to reference frame of solution spaces, which is a property of preferred optimization algorithms. In our previous study, piecewise-linear particle swarm optimizer (PPSO) has been proposed, which is effective in solving rotated problems. Because PPSO particles can move in solution spaces freely without depending on the coordinate systems, PPSO algorithm may have rotational invariance. However, theoretical analysis of reference frame invariance of PPSO has not been done. In addition, although behavior of each particle depends on PPSO parameters, good parameter conditions in solving various optimization problems have not been sufficiently clarified. In this paper, we analyze the reference frame invariance of PPSO theoretically, and investigated whether or not PPSO is invariant under reference frame alteration. We clarify that control parameters of PPSO which affect movement of each particle and performance of PPSO through numerical simulations.

  • An Efficient Blacklistable Anonymous Credentials without TTP of Tracing Authority Using Pairing-Based Accumulator

    Yuu AIKOU  Shahidatul SADIAH  Toru NAKANISHI  

     
    PAPER-Cryptography and Information Security

      Page(s):
    1968-1979

    In conventional ID-based user authentications, privacy issues may occur, since users' behavior histories are collected in Service Providers (SPs). Although anonymous authentications such as group signatures have been proposed, these schemes rely on a Trusted Third Party (TTP) capable of tracing misbehaving users. Thus, the privacy is not high, because the TTP of tracing authority can always trace users. Therefore, the anonymous credential system using a blacklist without the TTP of tracing authority has been proposed, where blacklisted anonymous users can be blocked. Recently, an RSA-based blacklistable anonymous credential system with efficiency improvement has been proposed. However, this system still has an efficiency problem: The data size in the authentication is O(K'), where K' is the maximum number of sessions in which the user can conduct. Furthermore, the O(K')-size data causes the user the computational cost of O(K') exponentiations. In this paper, a blacklistable anonymous credential system using a pairing-based accumulator is proposed. In the proposed system, the data size in the authentication is constant for parameters. Although the user's computational cost depends on parameters, the dependent cost is OBL·K) multiplications, instead of exponentiations, where δBL is the number of sessions added to the blacklist after the last authentication of the user, and K is the number of past sessions of the user. The demerit of the proposed system is O(n)-size public key, where n corresponds to the total number of all sessions of all users in the system. But, the user only has to download the public key once.

  • Hardware-Aware Sum-Product Decoding in the Decision Domain Open Access

    Mizuki YAMADA  Keigo TAKEUCHI  Kiyoyuki KOIKE  

     
    PAPER-Coding Theory

      Page(s):
    1980-1987

    We propose hardware-aware sum-product (SP) decoding for low-density parity-check codes. To simplify an implementation using a fixed-point number representation, we transform SP decoding in the logarithm domain to that in the decision domain. A polynomial approximation is proposed to implement an update rule of the proposed SP decoding efficiently. Numerical simulations show that the approximate SP decoding achieves almost the same performance as the exact SP decoding when an appropriate degree in the polynomial approximation is used, that it improves the convergence properties of SP and normalized min-sum decoding in the high signal-to-noise ratio regime, and that it is robust against quantization errors.

  • Natural Gradient Descent of Complex-Valued Neural Networks Invariant under Rotations

    Jun-ichi MUKUNO  Hajime MATSUI  

     
    PAPER-Neural Networks and Bioengineering

      Page(s):
    1988-1996

    The natural gradient descent is an optimization method for real-valued neural networks that was proposed from the viewpoint of information geometry. Here, we present an extension of the natural gradient descent to complex-valued neural networks. Our idea is to use the Hermitian extension of the Fisher information matrix. Moreover, we generalize the projected natural gradient (PRONG), which is a fast natural gradient descent algorithm, to complex-valued neural networks. We also consider the advantage of complex-valued neural networks over real-valued neural networks. A useful property of complex numbers in the complex plane is that the rotation is simply expressed by the multiplication. By focusing on this property, we construct the output function of complex-valued neural networks, which is invariant even if the input is changed to its rotated value. Then, our complex-valued neural network can learn rotated data without data augmentation. Finally, through simulation of online character recognition, we demonstrate the effectiveness of the proposed approach.

  • Distributed Mutually Referenced Equalization

    Yoshiki SUGITANI  Wataru YAMAMOTO  Teruyuki MIYAJIMA  

     
    LETTER-Digital Signal Processing

      Page(s):
    1997-2000

    We propose a distributed blind equalization method for wireless sensor networks, in which a source sends data and each node performs time-domain equalization to estimate the data from a received signal that is affected by inter-symbol interference. The equalization can be performed distributively based on the mutually referenced equalization principle. Even if the nodes in the network are not fully connected to each other, the average consensus technique enables us to perform the equalization of all channels.

  • Spectra Restoration of Bone-Conducted Speech via Attention-Based Contextual Information and Spectro-Temporal Structure Constraint Open Access

    Changyan ZHENG  Tieyong CAO  Jibin YANG  Xiongwei ZHANG  Meng SUN  

     
    LETTER-Digital Signal Processing

      Page(s):
    2001-2007

    Compared with acoustic microphone (AM) speech, bone-conducted microphone (BCM) speech is much immune to background noise, but suffers from severe loss of information due to the characteristics of the human-body transmission channel. In this letter, a new method for the speaker-dependent BCM speech enhancement is proposed, in which we focus our attention on the spectra restoration of the distorted speech. In order to better infer the missing components, an attention-based bidirectional Long Short-Term Memory (AB-BLSTM) is designed to optimize the use of contextual information to model the relationship between the spectra of BCM speech and its corresponding clean AM speech. Meanwhile, a structural error metric, Structural SIMilarity (SSIM) metric, originated from image processing is proposed to be the loss function, which provides the constraint of the spectro-temporal structures in recovering of the spectra. Experiments demonstrate that compared with approaches based on conventional DNN and mean square error (MSE), the proposed method can better recover the missing phonemes and obtain spectra with spectro-temporal structure more similar to the target one, which leads to great improvement on objective metrics.

  • Enhancing Physical Layer Security Performance in Downlink Cellular Networks through Cooperative Users

    Shijie WANG  Yuanyuan GAO  Xiaochen LIU  Guangna ZHANG  Nan SHA  Mingxi GUO  Kui XU  

     
    LETTER-Graphs and Networks

      Page(s):
    2008-2014

    In this paper, we explore how to enhance the physical layer security performance in downlink cellular networks through cooperative jamming technology. Idle user equipments (UE) are used to cooperatively transmit jamming signal to confuse eavesdroppers (Eve). We propose a threshold-based jammer selection scheme to decide which idle UE should participate in the transmission of jamming signal. Threshold conditions are carefully designed to decrease interference to legitimate channel, while maintain the interference to the Eves. Moreover, fewer UE are activated, which is helpful for saving energy consumptions of cooperative UEs. Analytical expressions of the connection and secrecy performances are derived, which are validated through Monte Carlo simulations. Theoretical and simulation results reveal that our proposed scheme can improve connection performance, while approaches the secrecy performance of [12]. Furthermore, only 43% idle UEs of [12] are used for cooperative jamming, which helps to decrease energy consumption of network.

  • Security Performance Analysis of Joint Multi-Relay and Jammer Selection for Physical-Layer Security under Nakagami-m Fading Channel

    Guangna ZHANG  Yuanyuan GAO  Huadong LUO  Nan SHA  Mingxi GUO  Kui XU  

     
    LETTER-Cryptography and Information Security

      Page(s):
    2015-2020

    In this paper, we investigate a novel joint multi-relay and jammer selection (JMRJS) scheme in order to improve the physical layer security of wireless networks. In the JMRJS scheme, all the relays succeeding in source decoding are selected to assist in the source signal transmission and meanwhile, all the remaining relay nodes are employed to act as friendly jammers to disturb the eavesdroppers by broadcasting artificial noise. Based on the more general Nakagami-m fading channel, we analyze the security performance of the JMRJS scheme for protecting the source signal against eavesdropping. The exact closed-form expressions of outage probability (OP) and intercept probability (IP) for the JMRJS scheme over Nakagami-m fading channel are derived. Moreover, we analyze the security-reliability tradeoff (SRT) of this scheme. Simulation results show that as the number of decode-and-forward (DF)relay nodes increases, the SRT of the JMRJS scheme improves notably. And when the transmit power is below a certain value, the SRT of the JMRJS scheme consistently outperforms the joint single-relay and jammer selection (JSRJS) scheme and joint equal-relay and jammer selection (JERJS) scheme respectively. In addition, the SRT of this scheme is always better than that of the multi-relay selection (MRS) scheme.

  • On the Minimum Distance of Some Improper Array Codes

    Haiyang LIU  Lianrong MA  Hao ZHANG  

     
    LETTER-Coding Theory

      Page(s):
    2021-2026

    For an odd prime q and an integer mq, we can construct a regular quasi-cyclic parity-check matrix HI(m,q) that specifies a linear block code CI(m,q), called an improper array code. In this letter, we prove the minimum distance of CI(4,q) is equal to 10 for any q≥11. In addition, we prove the minimum distance of CI(5,q) is upper bounded by 12 for any q≥11 and conjecture the upper bound is tight.

  • Improved Weighted Least Square Phase Estimation for OFDM-Based WLANs

    Xiaoping ZHOU  Bin WU  Kan ZHENG  Zhou WANG  

     
    LETTER-Communication Theory and Signals

      Page(s):
    2027-2030

    In this paper, we propose an improved weighted least square (IWLS) method to estimate and compensate phase variations utilizing pilots, for Orthogonal Frequency Division Multiplexing (OFDM) based very high throughput wireless local area networks (WLANs). The remaining phase is composed of the common phase error (CPE) and the sampling time offset (STO). For IWLS, the CPE maximum likelihood (ML) estimation is proposed to improve the CPE estimation accuracy, while the STO fitting is proposed to enhance the estimation of STO. With these two mechanisms, IWLS can improve phase estimation performance. Simulation results show that, compared to weighted least square (WLS) scheme, a better pocket error rate (PER) is achieved by using the proposed method, but with a comparable complexity.

  • Decentralized Relay Selection for Large-Scale Dynamic UAVs Networks: A Mood-Driven Approach

    Xijian ZHONG  Yan GUO  Ning LI  Shanling LI  Aihong LU  

     
    LETTER-Communication Theory and Signals

      Page(s):
    2031-2036

    In the large-scale multi-UAV systems, the direct link may be invalid for two remote nodes on account of the constrained power or complex communication environment. Idle UAVs may work as relays between the sources and destinations to enhance communication quality. In this letter, we investigate the opportunistic relay selection for the UAVs dynamic network. On account of the time-varying channel states and the variable numbers of sources and relays, relay selection becomes much more difficult. In addition, information exchange among all nodes may bring much cost and it is difficult to implement in practice. Thus, we propose a decentralized relay selection approach based on mood-driven mechanism to combat the dynamic characteristics, aiming to maximize the total capacity of the network without information exchange. With the proposed approach, the sources can make decisions only according to their own current states and update states according to immediate rewards. Numerical results show that the proposed approach has attractive properties.