The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] reduction(403hit)

21-40hit(403hit)

  • Noise-Robust Distorted Born Iterative Method with Prior Estimate for Microwave Ablation Monitoring Open Access

    Yuriko TAKAISHI  Shouhei KIDERA  

     
    BRIEF PAPER-Electromagnetic Theory

      Pubricized:
    2020/10/06
      Vol:
    E104-C No:4
      Page(s):
    148-152

    A noise-robust and accuracy-enhanced microwave imaging algorithm is presented for microwave ablation monitoring of cancer treatment. The ablation impact of dielectric change can be assessed by microwave inverse scattering analysis, where the dimension and dielectric drop of the ablation zone enable safe ablation monitoring. We focus on the distorted Born iterative method (DBIM), which is applicable to highly heterogeneous and contrasted dielectric profiles. As the reconstruction accuracy and convergence speed of DBIM depend largely on the initial estimate of the dielectric profile or noise level, this study exploits a prior estimate of the DBIM for the pre-ablation state to accelerate the convergence speed and introduces the matched-filter-based noise reduction scheme in the DBIM framework. The two-dimensional finite-difference time-domain numerical test with realistic breast phantoms shows that our method significantly enhances the reconstruction accuracy with a lower computational cost.

  • Complexity-Reduced Adaptive PAPR Reduction Method Using Null Space in MIMO Channel for MIMO-OFDM Signals Open Access

    Taku SUZUKI  Mikihito SUZUKI  Yoshihisa KISHIYAMA  Kenichi HIGUCHI  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2020/03/23
      Vol:
    E103-B No:9
      Page(s):
    1019-1029

    This paper proposes a computational complexity-reduced algorithm for an adaptive peak-to-average power ratio (PAPR) reduction method previously developed by members of our research group that uses the null space in a multiple-input multiple-output (MIMO) channel for MIMO-orthogonal frequency division multiplexing (OFDM) signals. The proposed algorithm is an extension of the peak cancellation (PC) signal-based method that has been mainly investigated for per-antenna PAPR reduction. This method adds the PC signal, which is designed so that the out-of-band radiation is removed/reduced, directly to the time-domain transmission signal at each antenna. The proposed method, referred to as PCCNC (PC with channel-null constraint), performs vector-level signal processing in the PC signal generation so that the PC signal is transmitted only to the null space in the MIMO channel. We investigate three methods to control the beamforming (BF) vector in the PC signal, which is a key factor in determining the achievable PAPR performance of the algorithm. Computer simulation results show that the proposed PCCNC achieves approximately the same throughput-vs.-PAPR performance as the previous method while dramatically reducing the required computational cost.

  • A Flexible Overloaded MIMO Receiver with Adaptive Selection of Extended Rotation Matrices

    Satoshi DENNO  Akihiro KITAMOTO  Ryosuke SAWADA  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2020/01/17
      Vol:
    E103-B No:7
      Page(s):
    787-795

    This paper proposes a novel flexible receiver with virtual channels for overloaded multiple-input multiple-output (MIMO) channels. The receiver applies extended rotation matrices proposed in the paper for the flexibility. In addition, adaptive selection of the extended rotation matrices is proposed for further performance improvement. We propose two techniques to reduce the computational complexity of the adaptive selection. As a result, the proposed receiver gives us an option to reduce the complexity with a slight decrease in the transmission performance by changing receiver configuration parameters. A computer simulation reveals that the adaptive selection attains a gain of about 3dB at the BER of 10-3.

  • A Constant-Size Signature Scheme with a Tighter Reduction from the CDH Assumption Open Access

    Kaisei KAJITA  Kazuto OGAWA  Eiichiro FUJISAKI  

     
    PAPER

      Vol:
    E103-A No:1
      Page(s):
    141-149

    We present a constant-size signature scheme under the CDH assumption. It has a tighter security reduction than any other constant-size signature scheme with a security reduction to solving some intractable search problems. Hofheinz, Jager, and Knapp (PKC 2012) presented a constant-size signature scheme under the CDH assumption with a reduction loss of O(q), where q is the number of signing queries. They also proved that the reduction loss of O(q) is optimal in a black-box security proof. To the best of our knowledge, no constant-size signature scheme has been proposed with a tighter reduction (to the hardness of a search problem) than that proposed by Hofheinz et al., even if it is not re-randomizable. We remark that our scheme is not re-randomizable. We achieve the reduction loss of O(q/d), where d is the number of group elements in a public key.

  • Clustering Method for Reduction of Area and Power Consumption on Post-Silicon Delay Tuning

    Kota MUROI  Hayato MASHIKO  Yukihide KOHIRA  

     
    PAPER

      Vol:
    E102-A No:7
      Page(s):
    894-903

    Due to progressing process technology, yield of chips is reduced by timing violation caused by delay variation of gates and wires in fabrication. Recently, post-silicon delay tuning, which inserts programmable delay elements (PDEs) into clock trees before the fabrication and adjusts the delays of the PDEs to recover the timing violation after the fabrication, is promising to improve the yield. Although post-silicon delay tuning improves the yield, it increases circuit area and power consumption since the PDEs are inserted. In this paper, a PDE structure is taken into consideration to reduce the circuit area and the power consumption. Moreover, a delay selection algorithm, and a clustering method, in which some PDEs are merged into a PDE and the PDE is inserted for multiple registers, are proposed to reduce the circuit area and the power consumption. In computational experiments, the proposed method reduced the circuit area and the power consumption in comparison with an existing method.

  • A Fine-Grained Multicasting of Configuration Data for Coarse-Grained Reconfigurable Architectures

    Takuya KOJIMA  Hideharu AMANO  

     
    PAPER-Computer System

      Pubricized:
    2019/04/05
      Vol:
    E102-D No:7
      Page(s):
    1247-1256

    A novel configuration data compression technique for coarse-grained reconfigurable architectures (CGRAs) is proposed. Reducing the size of configuration data of CGRAs shortens the reconfiguration time especially when the communication bandwidth between a CGRA and a host CPU is limited. In addition, it saves energy consumption of configuration cache and controller. The proposed technique is based on a multicast configuration technique called RoMultiC, which reduces the configuration time by multicasting the same data to multiple PEs (Processing Elements) with two bit-maps. Scheduling algorithms for an optimizing the order of multicasting have been proposed. However, the multicasting is possible only if each PE has completely the same configuration. In general, configuration data for CGRAs can be divided into some fields like machine code formats of general perpose CPUs. The proposed scheme confines a part of fields for multicasting so that the possibility of multicasting more PEs can be increased. This paper analyzes algorithms to find a configuration pattern which maximizes the number of multicasted PEs. We implemented the proposed scheme to CMA (Cool Mega Array), a straight forward CGRA as a case study. Experimental results show that the proposed method achieves 40.0% smaller configuration than a previous method for an image processing application at maximum. The exploration of the multicasted grain size reveals the effective grain size for each algorithm. Furthermore, since both a dynamic power consumption of the configuration controller and a configuration time are improved, it achieves 50.1% reduction of the energy consumption for the configuration with a negligible area overhead.

  • 2-D DOA Estimation Based on Sparse Bayesian Learning for L-Shaped Nested Array

    Lu CHEN  Daping BI  Jifei PAN  

     
    PAPER-Fundamental Theories for Communications

      Pubricized:
    2018/10/23
      Vol:
    E102-B No:5
      Page(s):
    992-999

    In sparsity-based optimization problems for two dimensional (2-D) direction-of-arrival (DOA) estimation using L-shaped nested arrays, one of the major issues is computational complexity. A 2-D DOA estimation algorithm is proposed based on reconsitution sparse Bayesian learning (RSBL) and cross covariance matrix decomposition. A single measurement vector (SMV) model is obtained by the difference coarray corresponding to one-dimensional nested array. Through spatial smoothing, the signal measurement vector is transformed into a multiple measurement vector (MMV) matrix. The signal matrix is separated by singular values decomposition (SVD) of the matrix. Using this method, the dimensionality of the sensing matrix and data size can be reduced. The sparse Bayesian learning algorithm is used to estimate one-dimensional angles. By using the one-dimensional angle estimations, the steering vector matrix is reconstructed. The cross covariance matrix of two dimensions is decomposed and transformed. Then the closed expression of the steering vector matrix of another dimension is derived, and the angles are estimated. Automatic pairing can be achieved in two dimensions. Through the proposed algorithm, the 2-D search problem is transformed into a one-dimensional search problem and a matrix transformation problem. Simulations show that the proposed algorithm has better angle estimation accuracy than the traditional two-dimensional direction finding algorithm at low signal-to-noise ratio and few samples.

  • Symmetric Decomposition of Convolution Kernels

    Jun OU  Yujian LI  

     
    LETTER-Biocybernetics, Neurocomputing

      Pubricized:
    2018/10/18
      Vol:
    E102-D No:1
      Page(s):
    219-222

    It is a hot issue that speeding up the network layers and decreasing the network parameters in convolutional neural networks (CNNs). In this paper, we propose a novel method, namely, symmetric decomposition of convolution kernels (SDKs). It symmetrically separates k×k convolution kernels into (k×1 and 1×k) or (1×k and k×1) kernels. We conduct the comparison experiments of the network models designed by SDKs on MNIST and CIFAR-10 datasets. Compared with the corresponding CNNs, we obtain good recognition performance, with 1.1×-1.5× speedup and more than 30% reduction of network parameters. The experimental results indicate our method is useful and effective for CNNs in practice, in terms of speedup performance and reduction of parameters.

  • New Context-Adaptive Arithmetic Coding Scheme for Lossless Bit Rate Reduction of Parametric Stereo in Enhanced aacPlus

    Hee-Suk PANG  Jun-seok LIM  Hyun-Young JIN  

     
    LETTER-Speech and Hearing

      Pubricized:
    2018/09/18
      Vol:
    E101-D No:12
      Page(s):
    3258-3262

    We propose a new context-adaptive arithmetic coding (CAAC) scheme for lossless bit rate reduction of parametric stereo (PS) in enhanced aacPlus. Based on the probability analysis of stereo parameters indexes in PS, we propose a stereo band-dependent CAAC scheme for PS. We also propose a new coding structure of the scheme which is simple but effective. The proposed scheme has normal and memory-reduced versions, which are superior to the original and conventional schemes and guarantees significant bit rate reduction of PS. The proposed scheme can be an alternative to the original PS coding scheme at low bit rate, where coding efficiency is very important.

  • Accelerating a Lloyd-Type k-Means Clustering Algorithm with Summable Lower Bounds in a Lower-Dimensional Space

    Kazuo AOYAMA  Kazumi SAITO  Tetsuo IKEDA  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2018/08/02
      Vol:
    E101-D No:11
      Page(s):
    2773-2783

    This paper presents an efficient acceleration algorithm for Lloyd-type k-means clustering, which is suitable to a large-scale and high-dimensional data set with potentially numerous classes. The algorithm employs a novel projection-based filter (PRJ) to avoid unnecessary distance calculations, resulting in high-speed performance keeping the same results as a standard Lloyd's algorithm. The PRJ exploits a summable lower bound on a squared distance defined in a lower-dimensional space to which data points are projected. The summable lower bound can make the bound tighter dynamically by incremental addition of components in the lower-dimensional space within each iteration although the existing lower bounds used in other acceleration algorithms work only once as a fixed filter. Experimental results on large-scale and high-dimensional real image data sets demonstrate that the proposed algorithm works at high speed and with low memory consumption when large k values are given, compared with the state-of-the-art algorithms.

  • Dynamic Ensemble Selection Based on Rough Set Reduction and Cluster Matching

    Ying-Chun CHEN  Ou LI  Yu SUN  

     
    PAPER-Fundamental Theories for Communications

      Pubricized:
    2018/04/11
      Vol:
    E101-B No:10
      Page(s):
    2196-2202

    Ensemble learning is widely used in the field of sensor network monitoring and target identification. To improve the generalization ability and classification precision of ensemble learning, we first propose an approximate attribute reduction algorithm based on rough sets in this paper. The reduction algorithm uses mutual information to measure attribute importance and introduces a correction coefficient and an approximation parameter. Based on a random sampling strategy, we use the approximate attribute reduction algorithm to implement the multi-modal sample space perturbation. To further reduce the ensemble size and realize a dynamic subset of base classifiers that best matches the test sample, we define a similarity parameter between the test samples and training sample sets that takes the similarity and number of the training samples into consideration. We then propose a k-means clustering-based dynamic ensemble selection algorithm. Simulations show that the multi-modal perturbation method effectively selects important attributes and reduces the influence of noise on the classification results. The classification precision and runtime of experiments demonstrate the effectiveness of the proposed dynamic ensemble selection algorithm.

  • Meeting Tight Security for Multisignatures in the Plain Public Key Model

    Naoto YANAI  

     
    PAPER

      Vol:
    E101-A No:9
      Page(s):
    1484-1493

    Multisignatures are digital signatures for a group consisting of multiple signers where each signer signs common documents via interaction with its co-signers and the data size of the resultant signatures for the group is independent of the number of signers. In this work, we propose a multisignature scheme, whose security can be tightly reduced to the CDH problem in bilinear groups, in the strongest security model where nothing more is required than that each signer has a public key, i.e., the plain public key model. Loosely speaking, our main idea for a tight reduction is to utilize a three-round interaction in a full-domain hash construction. Namely, we surmise that a full-domain hash construction with three-round interaction will become tightly secure under the CDH problem. In addition, we show that the existing scheme by Zhou et al. (ISC 2011) can be improved to a construction with a tight security reduction as an application of our proof framework.

  • Optimization of Body Biasing for Variable Pipelined Coarse-Grained Reconfigurable Architectures

    Takuya KOJIMA  Naoki ANDO  Hayate OKUHARA  Ng. Anh Vu DOAN  Hideharu AMANO  

     
    PAPER-Computer System

      Pubricized:
    2018/03/09
      Vol:
    E101-D No:6
      Page(s):
    1532-1540

    Variable Pipeline Cool Mega Array (VPCMA) is a low power Coarse Grained Reconfigurable Architecture (CGRA) based on the concept of CMA (Cool Mega Array). It provides a pipeline structure in the PE array that can be configured so as to fit target algorithms and required performance. Also, VPCMA uses the Silicon On Thin Buried oxide (SOTB) technology, a type of Fully Depleted Silicon On Insulator (FDSOI), so it is possible to control its body bias voltage to provide a balance between performance and leakage power. In this paper, we study the optimization of the VPCMA body bias while considering simultaneously its variable pipeline structure. Through evaluations, we can observe that it is possible to achieve an average reduction of energy consumption, for the studied applications, of 17.75% and 10.49% when compared to respectively the zero bias (without body bias control) and the uniform (control of the whole PE array) cases, while respecting performance constraints. Besides, it is observed that, with appropriate body bias control, it is possible to extend the possible performance, hence enabling broader trade-off analyzes between consumption and performance. Considering the dynamic power as well as the static power, more appropriate pipeline structure and body bias voltage can be obtained. In addition, when the control of VDD is integrated, higher performance can be achieved with a steady increase of the power. These promising results show that applying an adequate optimization technique for the body bias control while simultaneously considering pipeline structures can not only enable further power reduction than previous methods, but also allow more trade-off analysis possibilities.

  • Energy-Efficient DRAM Selective Refresh Technique with Page Residence in a Memory Hierarchy of Hardware-Managed TLB

    Miseon HAN  Yeoul NA  Dongha JUNG  Hokyoon LEE  Seon WOOK KIM  Youngsun HAN  

     
    PAPER-Integrated Electronics

      Vol:
    E101-C No:3
      Page(s):
    170-182

    A memory controller refreshes DRAM rows periodically in order to prevent DRAM cells from losing data over time. Refreshes consume a large amount of energy, and the problem becomes worse with the future larger DRAM capacity. Previously proposed selective refreshing techniques are either conservative in exploiting the opportunity or expensive in terms of required implementation overhead. In this paper, we propose a novel DRAM selective refresh technique by using page residence in a memory hierarchy of hardware-managed TLB. Our technique maximizes the opportunity to optimize refreshing by activating/deactivating refreshes for DRAM pages when their PTEs are inserted to/evicted from TLB or data caches, while the implementation cost is minimized by slightly extending the existing infrastructure. Our experiment shows that the proposed technique can reduce DRAM refresh power 43.6% on average and EDP 3.5% with small amount of hardware overhead.

  • On the Properties and Applications of Inconsistent Neighborhood in Neighborhood Rough Set Models

    Shujiao LIAO  Qingxin ZHU  Rui LIANG  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2017/12/20
      Vol:
    E101-D No:3
      Page(s):
    709-718

    Rough set theory is an important branch of data mining and granular computing, among which neighborhood rough set is presented to deal with numerical data and hybrid data. In this paper, we propose a new concept called inconsistent neighborhood, which extracts inconsistent objects from a traditional neighborhood. Firstly, a series of interesting properties are obtained for inconsistent neighborhoods. Specially, some properties generate new solutions to compute the quantities in neighborhood rough set. Then, a fast forward attribute reduction algorithm is proposed by applying the obtained properties. Experiments undertaken on twelve UCI datasets show that the proposed algorithm can get the same attribute reduction results as the existing algorithms in neighborhood rough set domain, and it runs much faster than the existing ones. This validates that employing inconsistent neighborhoods is advantageous in the applications of neighborhood rough set. The study would provide a new insight into neighborhood rough set theory.

  • PAPR Reduction Method for Digital Predistortion Linearizer Compensating for Frequency Dependent IMD Components

    Yasunori SUZUKI  Junya OHKAWARA  Shoichi NARAHASHI  

     
    PAPER-Microwaves, Millimeter-Waves

      Vol:
    E101-C No:2
      Page(s):
    118-125

    This paper proposes a method for reducing the peak-to-average power ratio (PAPR) at the output signal of a digital predistortion linearizer (DPDL) that compensates for frequency dependent intermodulation distortion (IMD) components. The proposed method controls the amplitude and phase values of the frequency components corresponding to the transmission bandwidth of the output signal. A DPDL employing the proposed method simultaneously provides IMD component cancellation of out-of-band components and PAPR reduction at the output signal. This paper identifies the amplitude and phase conditions to minimize the PAPR. Experimental results based on a 2-GHz band 1-W class power amplifier show that the proposed method improves the drain efficiency of the power amplifier when degradation is allowed in the error vector magnitude. To the best knowledge of the authors, this is the first PAPR reduction method for DPDL that reduces the PAPR while simultaneously compensating for IMD components.

  • Non-Linear Precoding Scheme Using MMSE Based Successive Inter-User Interference Pre-Cancellation and Perturbation Vector Search for Downlink MU-MIMO Systems

    Kenji HOSHINO  Manabu MIKAMI  Sourabh MAITI  Hitoshi YOSHINO  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2017/08/22
      Vol:
    E101-B No:2
      Page(s):
    451-461

    Non-linear precoding (NLP) scheme for downlink multi-user multiple-input multiple-output (DL-MU-MIMO) transmission has received much attention as a promising technology to achieve high capacity within the limited bandwidths available to radio access systems. In order to minimize the required transmission power for DL-MU-MIMO and achieve high spectrum efficiency, Vector Perturbation (VP) was proposed as an optimal NLP scheme. Unfortunately, the original VP suffers from significant computation complexity in detecting the optimal perturbation vector from an infinite number of the candidates. To reduce the complexity with near transmission performance of VP, several recent studies investigated various efficient NLP schemes based on the concept of Tomlinson-Harashima precoding (THP) that applies successive pre-cancellation of inter-user interference (IUI) and offsets the transmission vector based on a modulo operation. In order to attain transmission performance improvement over the original THP, a previous work proposed Minimum Mean Square Error based THP (MMSE-THP) employing IUI successive pre-cancellation based on MMSE criteria. On the other hand, to improve the transmission performance of MMSE-THP, other previous works proposed Ordered MMSE-THP and Lattice-Reduction-Aided MMSE-THP (LRA MMSE-THP). This paper investigates the further transmission performance improvement of Ordered MMSE-THP and LRA MMSE-THP. This paper starts by proposing an extension of MMSE-THP employing a perturbation vector search (PVS), called PVS MMSE-THP as a novel NLP scheme, where the modulo operation is substituted by PVS and a subtraction operation from the transmit signal vector. Then, it introduces an efficient search algorithm of appropriate perturbation vector based on a depth-first branch-and-bound search for PVS MMSE-THP. Next, it also evaluates the transmission performance of PVS MMSE-THP with the appropriate perturbation vector detected by the efficient search algorithm. Computer simulations quantitatively clarify that PVS MMSE-THP achieves better transmission performance than the conventional NLP schemes. Moreover, it also clarifies that PVS MMSE-THP increases the effect of required transmission power reduction with the number of transmit antennas compared to the conventional NLP schemes.

  • Inter-Terminal Interference Evaluation of Full Duplex MIMO Using Measured Channel

    Yuta KASHINO  Masakuni TSUNEZAWA  Naoki HONMA  Kentaro NISHIMORI  

     
    PAPER-MIMO

      Pubricized:
    2017/08/22
      Vol:
    E101-B No:2
      Page(s):
    434-440

    In-band full-duplex (FD) Multiple-Input and Multiple-Output (MIMO) communication performs uplink and downlink transmission at the same time using the same frequency. In this system, the spectral efficiency is theoretically double that of conventional duplex schemes, such as Time Division Duplex (TDD) and Frequency Division Duplex (FDD). However, this system suffers interference because the uplink and downlink streams coexist within the same channel. Especially at the terminal side, it is quite difficult for the terminal to eliminate the interference signals from other terminals since it has no knowledge about the contents of the interference signals. This paper presents an inter-terminal interference suppression method between the uplink and downlink signals assuming the multi-user environment. This method uses eigen-beamforming at the transmitting terminal to direct the null to the other terminal. Since this beamforming technique reduces the degrees of freedom available, the interference suppression performance and transmitting data-rate have a trade-off relation. This study investigates the system capacity characteristics in multi-user full-duplex MIMO communication using the propagation channel information measured in an actual outdoor experiment and shows that the proposed communication scheme offers higher system capacity than the conventional scheme.

  • Tighter Reductions for Deterministic Identity-Based Signatures

    Naoto YANAI  Toru FUJIWARA  

     
    PAPER

      Vol:
    E101-A No:1
      Page(s):
    64-76

    Deterministic ID-based signatures are digital signatures where secret keys are probabilistically generated by a key generation center while the signatures are generated deterministically. Although the deterministic ID-based signatures are useful for both systematic and cryptographic applications, to the best of our knowledge, there is no scheme with a tight reduction proof. Loosely speaking, since the security is downgraded through dependence on the number of queries by an adversary, a tighter reduction for the security of a scheme is desirable, and this reduction must be as close to the difficulty of its underlying hard problem as possible. In this work, we discuss mathematical features for a tight reduction of deterministic ID-based signatures, and show that the scheme by Selvi et al. (IWSEC 2011) is tightly secure by our new proof framework under a selective security model where a target identity is designated in advance. Our proof technique is versatile, and hence a reduction cost becomes tighter than the original proof even under an adaptive security model. We furthermore improve the scheme by Herranz (The Comp. Jour., 2006) to prove tight security in the same manner as described above. We furthermore construct an aggregate signature scheme with partial aggregation, which is a key application of deterministic ID-based signatures, from the improved scheme.

  • Black-Box Separations on Fiat-Shamir-Type Signatures in the Non-Programmable Random Oracle Model

    Masayuki FUKUMITSU  Shingo HASEGAWA  

     
    PAPER

      Vol:
    E101-A No:1
      Page(s):
    77-87

    In recent years, Fischlin and Fleischhacker showed the impossibility of proving the security of specific types of FS-type signatures, the signatures constructed by the Fiat-Shamir transformation, via a single-instance reduction in the non-programmable random oracle model (NPROM, for short). In this paper, we pose a question whether or not the impossibility of proving the security of any FS-type signature can be shown in the NPROM. For this question, we show that each FS-type signature cannot be proven to be secure via a key-preserving reduction in the NPROM from the security against the impersonation of the underlying identification scheme under the passive attack, as long as the identification scheme is secure against the impersonation under the active attack. We also show the security incompatibility between the security of some FS-type signatures in the NPROM via a single-instance key-preserving reduction and the underlying cryptographic assumptions. By applying this result to the Schnorr signature, one can prove the incompatibility between the security of the Schnorr signature in this situation and the discrete logarithm assumption, whereas Fischlin and Fleischhacker showed that such an incompatibility cannot be proven via a non-key-preserving reduction.

21-40hit(403hit)