The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] PAR(2741hit)

1021-1040hit(2741hit)

  • Software-Based Parallel Cryptographic Solution with Massive-Parallel Memory-Embedded SIMD Matrix Architecture for Data-Storage Systems

    Takeshi KUMAKI  Tetsushi KOIDE  Hans Jurgen MATTAUSCH  Masaharu TAGAMI  Masakatsu ISHIZAKI  

     
    PAPER-Fundamentals of Information Systems

      Vol:
    E94-D No:9
      Page(s):
    1742-1754

    This paper presents a software-based parallel cryptographic solution with a massive-parallel memory-embedded SIMD matrix (MTX) for data-storage systems. MTX can have up to 2,048 2-bit processing elements, which are connected by a flexible switching network, and supports 2-bit 2,048-way bit-serial and word-parallel operations with a single command. Furthermore, a next-generation SIMD matrix called MX-2 has been developed by expanding processing-element capability of MTX from 2-bit to 4-bit processing. These SIMD matrix architectures are verified to be a better alternative for processing repeated-arithmetic and logical-operations in multimedia applications with low power consumption. Moreover, we have proposed combining Content Addressable Memory (CAM) technology with the massive-parallel memory-embedded SIMD matrix architecture to enable fast pipelined table-lookup coding. Since both arithmetic logical operation and table-lookup coding execute extremely fast on these architectures, efficient execution of encryption and decryption algorithms can be realized. Evaluation results of the CAM-less and CAM-enhanced massive-parallel SIMD matrix processor for the example of the Advanced Encryption Standard (AES), which is a widely-used cryptographic algorithm, show that a throughput of up to 2.19 Gbps becomes possible. This means that several standard data-storage transfer specifications, such as SD, CF (Compact Flash), USB (Universal Serial Bus) and SATA (Serial Advanced Technology Attachment) can be covered. Consequently, the massive-parallel SIMD matrix architecture is very suitable for private information protection in several data-storage media. A further advantage of the software based solution is the flexible update possibility of the implemented-cryptographic algorithm to a safer future algorithm. The massive-parallel memory-embedded SIMD matrix architecture (MTX and MX-2) is therefore a promising solution for integrated realization of real-time cryptographic algorithms with low power dissipation and small Si-area consumption.

  • Nonbinary Quasi-Cyclic LDPC Cycle Codes with Low-Density Systematic Quasi-Cyclic Generator Matrices

    Yang YANG  Chao CHEN  Jianjun MU  Jing WANG  Rong SUN  Xinmei WANG  

     
    LETTER-Fundamental Theories for Communications

      Vol:
    E94-B No:9
      Page(s):
    2620-2623

    In this letter, we propose an appealing class of nonbinary quasi-cyclic low-density parity-check (QC-LDPC) cycle codes. The parity-check matrix is carefully designed such that the corresponding generator matrix has some nice properties: 1) systematic, 2) quasi-cyclic, and 3) sparse, which allows a parallel encoding with low complexity. Simulation results show that the performance of the proposed encoding-aware LDPC codes is comparable to that of the progressive-edge-growth (PEG) constructed nonbinary LDPC cycle codes.

  • Decoupled Location Parameter Estimation of Near-Field Sources with Symmetric ULA

    Bum-Soo KWON  Tae-Jin JUNG  Kyun-Kyung LEE  

     
    LETTER-Antennas and Propagation

      Vol:
    E94-B No:9
      Page(s):
    2646-2649

    A novel algorithm is presented for near-field source localization with a symmetric uniform linear array (ULA) consisting of an even number of sensors. Based on element reordering of a symmetric ULA, the steering vector is factorised with respect to the range-independent bearing parameters and range-relevant 2-D location parameters, which allows the range-independent bearing estimation with rank-reduction idea. With the estimated bearing, the range estimation for each source is then obtained by defining the 1-D MUSIC spectrum. Simulation results are presented to validate the performance of the proposed algorithm.

  • Break Arcs Driven by Transverse Magnetic Field in a DC48 V/6-24 A Resistive Circuit

    Toru SUGIURA  Junya SEKIKAWA  Takayoshi KUBONO  

     
    PAPER

      Vol:
    E94-C No:9
      Page(s):
    1381-1387

    Silver electrical contacts are separated to generate break arcs in a DC48 V/6-24 A resistive circuit. The transverse magnetic field formed by a permanent magnet is applied to the break arcs. A series of experiments are carried out for two different experimental conditions. One condition is a constant contact separating speed while the magnetic flux density is changed to investigate the shortening effect of the arc duration. Another condition is a constant magnetic flux density while the contact separating speed is changed to investigate the changes in the arc duration and the contact gap when the break arc is extinguished. As a result, with constant separating speed, it is confirmed that the duration of break arcs is shortened by the transverse magnetic field and the break arcs are extinguished when the arc length reaches a certain value L. Under the condition of constant transverse magnetic field, (i) the arc duration is shortened by increasing the separation speed; (ii) the contact gap when the break arc is extinguished is almost constant when the separating speed v is sufficiently faster than 5 mm/s.

  • High-Speed FPGA Implementation of the SHA-1 Hash Function

    Je-Hoon LEE  Sang-Choon KIM  Young-Jun SONG  

     
    LETTER-Cryptography and Information Security

      Vol:
    E94-A No:9
      Page(s):
    1873-1876

    This paper presents a high-speed SHA-1 implementation. Unlike the conventional unfolding transformation, the proposed unfolding transformation technique makes the combined hash operation blocks to have almost the same delay overhead regardless of the unfolding factor. It can achieve high throughput of SHA-1 implementation by avoiding the performance degradation caused by the first hash computation. We demonstrate the proposed SHA-1 architecture on a FPGA chip. From the experimental results, the SHA-1 architecture with unfolding factor 5 shows 1.17 Gbps. The proposed SHA-1 architecture can achieve about 31% performance improvements compared to its counterparts. Thus, the proposed SHA-1 can be applicable for the security of the high-speed but compact mobile appliances.

  • Cross Low-Dimension Pursuit for Sparse Signal Recovery from Incomplete Measurements Based on Permuted Block Diagonal Matrix

    Zaixing HE  Takahiro OGAWA  Miki HASEYAMA  

     
    PAPER-Digital Signal Processing

      Vol:
    E94-A No:9
      Page(s):
    1793-1803

    In this paper, a novel algorithm, Cross Low-dimension Pursuit, based on a new structured sparse matrix, Permuted Block Diagonal (PBD) matrix, is proposed in order to recover sparse signals from incomplete linear measurements. The main idea of the proposed method is using the PBD matrix to convert a high-dimension sparse recovery problem into two (or more) groups of highly low-dimension problems and crossly recover the entries of the original signal from them in an iterative way. By sampling a sufficiently sparse signal with a PBD matrix, the proposed algorithm can recover it efficiently. It has the following advantages over conventional algorithms: (1) low complexity, i.e., the algorithm has linear complexity, which is much lower than that of existing algorithms including greedy algorithms such as Orthogonal Matching Pursuit and (2) high recovery ability, i.e., the proposed algorithm can recover much less sparse signals than even 1-norm minimization algorithms. Moreover, we demonstrate both theoretically and empirically that the proposed algorithm can reliably recover a sparse signal from highly incomplete measurements.

  • DCF-Based Cooperative MAC Protocol Employing Fuzzy Logic Partner Selection Scheme

    Verotiana H. RABARIJAONA  Akeo MASUDA  Shigeru SHIMAMOTO  

     
    PAPER-Wireless Communication Technologies

      Vol:
    E94-B No:9
      Page(s):
    2610-2619

    We propose FuzzyCoop, a cooperative MAC layer protocol employing a fuzzy logic partner selection algorithm. The protocol is based on the Distributed Coordination Function (DCF) protocol used in the IEEE 802.11 standard. There are three inputs to the fuzzy system: the Signal to Noise Ratio (SNR), the error ratio between two neighbors and the time the most recent packet was received from a neighbor. The fuzzy output is the partnership probability of a neighboring terminal. Besides, we introduce a cooperation incentive to the stations by providing them with the right to transmit their own data without contention after a successful cooperation. The protocol is evaluated through extensive simulations in different scenarios and is compared to the DCF protocol and a previously proposed cooperative protocol. Simulation results show that FuzzyCoop improves the performances of a wireless network and provides a more robust partner selection scheme.

  • Global Selection vs Local Ordering of Color SIFT Independent Components for Object/Scene Classification

    Dan-ni AI  Xian-hua HAN  Guifang DUAN  Xiang RUAN  Yen-wei CHEN  

     
    PAPER-Pattern Recognition

      Vol:
    E94-D No:9
      Page(s):
    1800-1808

    This paper addresses the problem of ordering the color SIFT descriptors in the independent component analysis for image classification. Component ordering is of great importance for image classification, since it is the foundation of feature selection. To select distinctive and compact independent components (IC) of the color SIFT descriptors, we propose two ordering approaches based on local variation, named as the localization-based IC ordering and the sparseness-based IC ordering. We evaluate the performance of proposed methods, the conventional IC selection method (global variation based components selection) and original color SIFT descriptors on object and scene databases, and obtain the following two main results. First, the proposed methods are able to obtain acceptable classification results in comparison with original color SIFT descriptors. Second, the highest classification rate can be obtained by using the global selection method in the scene database, while the local ordering methods give the best performance for the object database.

  • 2-D Frequency Estimation of Multiple Damped Sinusoids Using Subspace and Projection Separation Approaches

    Longting HUANG  Yuntao WU  Hing Cheung SO  Yanduo ZHANG  

     
    LETTER-Digital Signal Processing

      Vol:
    E94-A No:9
      Page(s):
    1842-1846

    In this paper, a new method for 2-D frequency estimation of multiple damped sinusoids in additive white Gaussian noise is proposed. The key idea is to combine the subspace-based technique and projection separation approach. The frequency parameters in the first dimension are estimated by the MUSIC-based method, and then a set of projection separation matrices are constructed by the estimated frequency parameters. In doing so, the frequency parameters in the second dimension can be separated by the constructed projection separation matrix. Finally, each frequency parameter in the second dimension is estimated by multiple 1-D MUSIC-based methods. The estimated frequency parameters in two dimensions are automatically paired. Computer simulations are included to compare the proposed algorithm with several existing methods.

  • A Novel Framework for Spectrum Sensing in Cognitive Radio Networks

    Navid TAFAGHODI KHAJAVI  Siavash SADEGHI IVRIGH  Seyed Mohammad-Sajad SADOUGH  

     
    PAPER-Wireless Communication Technologies

      Vol:
    E94-B No:9
      Page(s):
    2600-2609

    Cognitive radio (CR) is a key solution for the problem of inefficient usage of spectral resources. Spectrum sensing in each CR aims at detecting whether a preassigned spectrum band is occupied by a primary user or not. Conventional techniques do not allow the CR to communicate with its own base station during the spectrum sensing process. So, only a part of the frame can be used for cognitive data transmission. In this paper, we introduce a new spectrum sensing framework that combines a blind source separation technique with conventional spectrum sensing techniques. In this way, the cognitive transmitter can continue to transmit during spectrum sensing, if it was in operation in the previous frame. Moreover, the accuracy is improved since the decision made by the spectrum unit in each frame depends on the decision made in the previous frame. We use Markov chain tools to model the behavior of our spectrum sensing proposal and to derive the parameters that characterize its performance. Numerical results are provided to confirm the superiority of the proposed technique compared to conventional spectrum sensing techniques.

  • Estimating Environment-Independent Parameters on Basis of Comparative Assessment of User's Subjectivity

    Yasuhiro SATO  Shin'ichiro KITANO  Shingo ATA  Ikuo OKA  

     
    PAPER-Network

      Vol:
    E94-B No:8
      Page(s):
    2265-2273

    We analyze subjective assessments by comparative evaluations of bulk data transmission by using two psychological methods (the method of successive categories and the constant method). From the results of the first experiment, the thresholds at which participants downloading a data file began to feel dissatisfaction with service degradation are 15.7 Mbps and 11.6 Mbps obtained by the two different methods when the throughput without the service degradation is approximately 22 Mbps. In the second experiment, we investigate the threshold of user satisfaction for various network environments. The threshold is 63% of the throughput of the user's usual network environment. Moreover, from the viewpoint of download time, users feel dissatisfaction with the quality when download time becomes 1.5 times longer. These values can be used to more effectively allocate network resources and thereby achieve higher service quality.

  • Adaptive Bare Bones Particle Swarm Inspired by Cloud Model

    Junqi ZHANG  Lina NI  Jing YAO  Wei WANG  Zheng TANG  

     
    PAPER-Fundamentals of Information Systems

      Vol:
    E94-D No:8
      Page(s):
    1527-1538

    Kennedy has proposed the bare bones particle swarm (BBPS) by the elimination of the velocity formula and its replacement by the Gaussian sampling strategy without parameter tuning. However, a delicate balance between exploitation and exploration is the key to the success of an optimizer. This paper firstly analyzes the sampling distribution in BBPS, based on which we propose an adaptive BBPS inspired by the cloud model (ACM-BBPS). The cloud model adaptively produces a different standard deviation of the Gaussian sampling for each particle according to the evolutionary state in the swarm, which provides an adaptive balance between exploitation and exploration on different objective functions. Meanwhile, the diversity of the swarms is further enhanced by the randomness of the cloud model itself. Experimental results show that the proposed ACM-BBPS achieves faster convergence speed and more accurate solutions than five other contenders on twenty-five unimodal, basic multimodal, extended multimodal and hybrid composition benchmark functions. The diversity enhancement by the randomness in the cloud model itself is also illustrated.

  • Nonparametric Regression Method Based on Orthogonalization and Thresholding

    Katsuyuki HAGIWARA  

     
    PAPER-Artificial Intelligence, Data Mining

      Vol:
    E94-D No:8
      Page(s):
    1610-1619

    In this paper, we consider a nonparametric regression problem using a learning machine defined by a weighted sum of fixed basis functions, where the number of basis functions, or equivalently, the number of weights, is equal to the number of training data. For the learning machine, we propose a training scheme that is based on orthogonalization and thresholding. On the basis of the scheme, vectors of basis function outputs are orthogonalized and coefficients of the orthogonalized vectors are estimated instead of weights. The coefficient is set to zero if it is less than a predetermined threshold level assigned component-wise to each coefficient. We then obtain the resulting weight vector by transforming the thresholded coefficients. In this training scheme, we propose asymptotically reasonable threshold levels to distinguish contributed components from unnecessary ones. To see how this works in a simple case, we derive an upper bound for the generalization error of the training scheme with the given threshold levels. It tells us that an increase in the generalization error is of O(log n/n) when there is a sparse representation of a target function in an orthogonal domain. In implementing the training scheme, eigen-decomposition or the Gram–Schmidt procedure is employed for orthogonalization, and the corresponding training methods are referred to as OHTED and OHTGS. Furthermore, modified versions of OHTED and OHTGS, called OHTED2 and OHTGS2 respectively, are proposed for reduced estimation bias. On real benchmark datasets, OHTED2 and OHTGS2 are found to exhibit relatively good generalization performance. In addition, OHTGS2 is found to be obtain a sparse representation of a target function in terms of the basis functions.

  • Partial Derivative Guidance for Weak Classifier Mining in Pedestrian Detection

    Chang LIU  Guijin WANG  Chunxiao LIU  Xinggang LIN  

     
    LETTER-Image Recognition, Computer Vision

      Vol:
    E94-D No:8
      Page(s):
    1721-1724

    Boosting over weak classifiers is widely used in pedestrian detection. As the number of weak classifiers is large, researchers always use a sampling method over weak classifiers before training. The sampling makes the boosting process harder to reach the fixed target. In this paper, we propose a partial derivative guidance for weak classifier mining method which can be used in conjunction with a boosting algorithm. Using weak classifier mining method makes the sampling less degraded in the performance. It has the same effect as testing more weak classifiers while using acceptable time. Experiments demonstrate that our algorithm can process quicker than [1] algorithm in both training and testing, without any performance decrease. The proposed algorithms is easily extending to any other boosting algorithms using a window-scanning style and HOG-like features.

  • Design of an 8-nsec 72-bit-Parallel-Search Content-Addressable Memory Using a Phase-Change Device

    Satoru HANZAWA  Takahiro HANYU  

     
    PAPER-Integrated Electronics

      Vol:
    E94-C No:8
      Page(s):
    1302-1310

    This paper presents a content-addressable memory (CAM) using a phase-change device. A hierarchical match-line structure and a one-hot-spot block code are indispensable to suppress the resistance ratio of the phase-change device and the area overhead of match detectors. As a result, an 8-nsec 72-bit-parallel-search CAM is implemented using a phase-change-device/MOS-hybrid circuitry, where high and low resistances are higher than 2.3 MΩ and lower than 97 kΩ, respectively, while maintaining one-day retention.

  • Multi-Stage Decoding Scheme with Post-Processing for LDPC Codes to Lower the Error Floors

    Beomkyu SHIN  Hosung PARK  Jong-Seon NO  Habong CHUNG  

     
    LETTER-Fundamental Theories for Communications

      Vol:
    E94-B No:8
      Page(s):
    2375-2377

    In this letter, we propose a multi-stage decoding scheme with post-processing for low-density parity-check (LDPC) codes, which remedies the rapid performance degradation in the high signal-to-noise ratio (SNR) range known as error floor. In the proposed scheme, the unsuccessfully decoded words of the previous decoding stage are re-decoded by manipulating the received log-likelihood ratios (LLRs) of the properly selected variable nodes. Two effective criteria for selecting the probably erroneous variable nodes are also presented. Numerical results show that the proposed scheme can correct most of the unsuccessfully decoded words of the first stage having oscillatory behavior, which are regarded as a main cause of the error floor.

  • Performance Evaluation of TCP/IP over SCTP/IP Parallel Networking with Parallel Route Transmission

    Shota YAMADA  Daisuke ISHII  Satoru OKAMOTO  Naoaki YAMANAKA  

     
    LETTER-Network

      Vol:
    E94-B No:8
      Page(s):
    2394-2396

    In this paper, Transmission Control Protocol/Internet Protocol (TCP/IP) over Stream Control Transmission Protocol (SCTP)/IP parallel transmission system is proposed to realize large TCP/IP throughput. The proposed system enables SCTP/IP connection between switches by protocol stacking. The proposed system is implemented on a software switch to evaluate its performance. The evaluation result indicates that proposed system can achieve 90% throughput compared with serial transmission when the delay difference among parallel routes is 20 msec.

  • Image Inpainting Based on Adaptive Total Variation Model

    Zhaolin LU  Jiansheng QIAN  Leida LI  

     
    LETTER-Image

      Vol:
    E94-A No:7
      Page(s):
    1608-1612

    In this letter, a novel adaptive total variation (ATV) model is proposed for image inpainting. The classical TV model is a partial differential equation (PDE)-based technique. While the TV model can preserve the image edges well, it has some drawbacks, such as staircase effect in the inpainted image and slow convergence rate. By analyzing the diffusion mechanism of TV model and introducing a new edge detection operator named difference curvature, we propose a novel ATV inpainting model. The proposed ATV model can diffuse the image information smoothly and quickly, namely, this model not only eliminates the staircase effect but also accelerates the convergence rate. Experimental results demonstrate the effectiveness of the proposed scheme.

  • A Fast Divide-and-Conquer Algorithm for Indexing Human Genome Sequences

    Woong-Kee LOH  Yang-Sae MOON  Wookey LEE  

     
    PAPER-Fundamentals of Information Systems

      Vol:
    E94-D No:7
      Page(s):
    1369-1377

    Since the release of human genome sequences, one of the most important research issues is about indexing the genome sequences, and the suffix tree is most widely adopted for that purpose. The traditional suffix tree construction algorithms suffer from severe performance degradation due to the memory bottleneck problem. The recent disk-based algorithms also provide limited performance improvement due to random disk accesses. Moreover, they do not fully utilize the recent CPUs with multiple cores. In this paper, we propose a fast algorithm based on `divide-and-conquer' strategy for indexing the human genome sequences. Our algorithm nearly eliminates random disk accesses by accessing the disk in the unit of contiguous chunks. In addition, our algorithm fully utilizes the multi-core CPUs by dividing the genome sequences into multiple partitions and then assigning each partition to a different core for parallel processing. Experimental results show that our algorithm outperforms the previous fastest DIGEST algorithm by up to 10.5 times.

  • NUFFT- & GPU-Based Fast Imaging of Vegetation

    Amedeo CAPOZZOLI  Claudio CURCIO  Antonio DI VICO  Angelo LISENO  

     
    PAPER-Sensing

      Vol:
    E94-B No:7
      Page(s):
    2092-2103

    We develop an effective algorithm, based on the filtered backprojection (FBP) approach, for the imaging of vegetation. Under the FBP scheme, the reconstruction amounts at a non-trivial Fourier inversion, since the data are Fourier samples arranged on a non-Cartesian grid. The computational issue is efficiently tackled by Non-Uniform Fast Fourier Transforms (NUFFTs), whose complexity grows asymptotically as that of a standard FFT. Furthermore, significant speed-ups, as compared to fast CPU implementations, are obtained by a parallel versions of the NUFFT algorithm, purposely designed to be run on Graphic Processing Units (GPUs) by using the CUDA language. The performance of the parallel algorithm has been assessed in comparison to a CPU-multicore accelerated, Matlab implementation of the same routine, to other CPU-multicore accelerated implementations based on standard FFT and employing linear, cubic, spline and sinc interpolations and to a different, parallel algorithm exploiting a parallel linear interpolation stage. The proposed approach has resulted the most computationally convenient. Furthermore, an indoor, polarimetric experimental setup is developed, capable to isolate and introduce, one at a time, different non-idealities of a real acquisition, as the sources (wind, rain) of temporal decorrelation. Experimental far-field polarimetric measurements on a thuja plicata (western redcedar) tree point out the performance of the set up algorithm, its robustness against data truncation and temporal decorrelation as well as the possibility of discriminating scatterers with different features within the investigated scene.

1021-1040hit(2741hit)