The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] TE(21534hit)

5301-5320hit(21534hit)

  • Exposure-Resilient One-Round Tripartite Key Exchange without Random Oracles

    Koutarou SUZUKI  Kazuki YONEYAMA  

     
    PAPER

      Vol:
    E97-A No:6
      Page(s):
    1345-1355

    This paper studies Tripartite Key Exchange (3KE) which is a special case of Group Key Exchange. Though general one-round GKE satisfying advanced security properties such as forward secrecy and maximal-exposure-resilience (MEX-resilience) is not known, it can be efficiently constructed with the help of pairings in the 3KE case. In this paper, we introduce the first one-round 3KE which is MEX-resilient in the standard model, though existing one-round 3KE schemes are proved in the random oracle model (ROM), or not MEX-resilient. Each party broadcasts 4 group elements, and executes 14 pairing operations. Complexity is only three or four times larger in computation and communication than the existing most efficient MEX-resilient 3KE scheme in the ROM; thus, our protocol is adequately practical.

  • Voice Conversion Based on Speaker-Dependent Restricted Boltzmann Machines

    Toru NAKASHIKA  Tetsuya TAKIGUCHI  Yasuo ARIKI  

     
    PAPER-Voice Conversion and Speech Enhancement

      Vol:
    E97-D No:6
      Page(s):
    1403-1410

    This paper presents a voice conversion technique using speaker-dependent Restricted Boltzmann Machines (RBM) to build high-order eigen spaces of source/target speakers, where it is easier to convert the source speech to the target speech than in the traditional cepstrum space. We build a deep conversion architecture that concatenates the two speaker-dependent RBMs with neural networks, expecting that they automatically discover abstractions to express the original input features. Under this concept, if we train the RBMs using only the speech of an individual speaker that includes various phonemes while keeping the speaker individuality unchanged, it can be considered that there are fewer phonemes and relatively more speaker individuality in the output features of the hidden layer than original acoustic features. Training the RBMs for a source speaker and a target speaker, we can then connect and convert the speaker individuality abstractions using Neural Networks (NN). The converted abstraction of the source speaker is then back-propagated into the acoustic space (e.g., MFCC) using the RBM of the target speaker. We conducted speaker-voice conversion experiments and confirmed the efficacy of our method with respect to subjective and objective criteria, comparing it with the conventional Gaussian Mixture Model-based method and an ordinary NN.

  • Twin Domination Problems in Round Digraphs

    Tamaki NAKAJIMA  Yuuki TANAKA  Toru ARAKI  

     
    PAPER

      Vol:
    E97-A No:6
      Page(s):
    1192-1199

    A twin dominating set of a digraph D is a subset S of vertices if, for every vertex u ∉ S, there are vertices x,y ∈ S such that ux and yu are arcs of D. A digraph D is round if the vertices can be labeled as v0,v1,...,vn-1 so that, for each vertex vi, the out-neighbors of vi appear consecutively following vi and the in-neighbors of vi appear consecutively preceding vi. In this paper, we give polynomial time algorithms for finding a minimum weight twin dominating set and a minimum weight total twin dominating set for a weighted round digraph. Then we show that there is a polynomial time algorithm for deciding whether a locally semicomplete digraph has an independent twin dominating set. The class of locally semicomplete digraphs contains round digraphs as a special case.

  • Efficient Parallel Interference Cancellation MIMO Detector for Software Defined Radio on GPUs

    Rongchun LI  Yong DOU  Jie ZHOU  Chen CHEN  

     
    PAPER-Digital Signal Processing

      Vol:
    E97-A No:6
      Page(s):
    1388-1395

    The parallel interference cancellation (PIC) multiple input multiple output (MIMO) detection algorithm has bit error ratio (BER) performance comparable to the maximum likelihood (ML) algorithm but with complexity close to the simple linear detection algorithm such as zero forcing (ZF), minimum mean squared error (MMSE), and successive interference cancellation (SIC), etc. However, the throughput of PIC MIMO detector on central processing unit (CPU) cannot meet the requirement of wireless protocols. In order to reach the throughput required by the standards, the graphics processing unit (GPU) is exploited in this paper as the modem processor to accelerate the processing procedure of PIC MIMO detector. The parallelism of PIC algorithm is analyzed and the two-stage PIC detection is carefully developed to efficiently match the multi-core architecture. Several optimization methods are employed to enhance the throughput, such as the memory optimization and asynchronous data transfer. The experiment shows that our MIMO detector has excellent BER performance and the peak throughput is 337.84 Mega bits per second (Mbps), about 7x to 16x faster than that of CPU implementation with SSE2 optimization methods. The implemented MIMO detector has better computing throughput than recent GPU-based implementations.

  • A Note on Cooperating Systems of One-Way Alternating Finite Automata with Only Universal States

    Tatsuya FUJIMOTO  Tsunehiro YOSHINAGA  Makoto SAKAMOTO  

     
    LETTER

      Vol:
    E97-A No:6
      Page(s):
    1375-1377

    A cooperating system of finite automata (CS-FA) has more than one finite automata (FA's) and an input tape. These FA's operate independently on the input tape and can communicate with each other on the same cell of the input tape. For each k ≥ 1, let L[CS-1DFA(k)] (L[CS-1UFA(k)]) be the class of sets accepted by CS-FA's with k one-way deterministic finite automata (alternating finite automata with only universal states). We show that L[CS-1DFA(k+1)] - L[CS-1UFA(k)] ≠ ∅ and L[CS-1UFA(2)] - ∪1≤k<∞L[CS-1DFA(k)] ≠ ∅.

  • Theoretical Comparison of Root Computations in Finite Fields

    Ryuichi HARASAWA  Yutaka SUEYOSHI  Aichi KUDO  

     
    LETTER

      Vol:
    E97-A No:6
      Page(s):
    1378-1381

    In the paper [4], the authors generalized the Cipolla-Lehmer method [2][5] for computing square roots in finite fields to the case of r-th roots with r prime, and compared it with the Adleman-Manders-Miller method [1] from the experimental point of view. In this paper, we compare these two methods from the theoretical point of view.

  • Integration of Spectral Feature Extraction and Modeling for HMM-Based Speech Synthesis

    Kazuhiro NAKAMURA  Kei HASHIMOTO  Yoshihiko NANKAKU  Keiichi TOKUDA  

     
    PAPER-HMM-based Speech Synthesis

      Vol:
    E97-D No:6
      Page(s):
    1438-1448

    This paper proposes a novel approach for integrating spectral feature extraction and acoustic modeling in hidden Markov model (HMM) based speech synthesis. The statistical modeling process of speech waveforms is typically divided into two component modules: the frame-by-frame feature extraction module and the acoustic modeling module. In the feature extraction module, the statistical mel-cepstral analysis technique has been used and the objective function is the likelihood of mel-cepstral coefficients for given speech waveforms. In the acoustic modeling module, the objective function is the likelihood of model parameters for given mel-cepstral coefficients. It is important to improve the performance of each component module for achieving higher quality synthesized speech. However, the final objective of speech synthesis systems is to generate natural speech waveforms from given texts, and the improvement of each component module does not always lead to the improvement of the quality of synthesized speech. Therefore, ideally all objective functions should be optimized based on an integrated criterion which well represents subjective speech quality of human perception. In this paper, we propose an approach to model speech waveforms directly and optimize the final objective function. Experimental results show that the proposed method outperformed the conventional methods in objective and subjective measures.

  • An Adaptive Base Plane Filtering Algorithm for Inter-plane Estimation of RGB Images in HEVC RExt

    Jangwon CHOI  Yoonsik CHOE  

     
    LETTER-Image Processing and Video Processing

      Vol:
    E97-D No:6
      Page(s):
    1686-1689

    This letter proposes an adaptive base plane filtering algorithm for the inter-plane estimation of RGB images in HEVC RExt. Because most high-frequency components of RGB images have low inter-plane correlation, our proposed scheme adaptively removes the high-frequency components of the base plane in order to enhance the inter-plane estimation accuracy. The experimental results show that the proposed scheme provides average BD rate gains of 0.6%, 1.0%, and 1.2% in the G, B, and R planes, respectively, with slightly decreased complexity, as compared to the previous inter-plane filtering method.

  • A 10-bit CMOS Digital-to-Analog Converter with Compact Size for Display Applications

    Mungyu KIM  Hoon-Ju CHUNG  Young-Chan JANG  

     
    PAPER

      Vol:
    E97-C No:6
      Page(s):
    519-525

    A 10-bit digital-to-analog converter (DAC) with a small area is proposed for data-driver integrated circuits of active-matrix liquid crystal display systems. The 10-bit DAC consists of a 7-bit resistor string, a 7-bit two-step decoder, a 2-bit logarithmic time interpolator, and a buffer amplifier. The proposed logarithmic time interpolation is achieved by controlling the charging time of a first-order low-pass filter composed of a resistor and a capacitor. The 7-bit two-step decoder that follows the 7-bit resistor string outputs an analog signal of the stepped wave with two voltage levels using the additional 1-bit digital code for the logarithmic time interpolation. The proposed 10-bit DAC is implemented using a 0.35-µm CMOS process and its supply voltage is scalable from 3.3V to 5.0V. The area of the proposed 10-bit logarithmic time interpolation DAC occupies 57% of that of the conventional 10-bit resistor-string DAC. The DNL and INL of the implemented 10-bit DAC are +0.29/-0.30 and +0.47/-0.36 LSB, respectively.

  • Effects of Voluntary Movements on Audio-Tactile Temporal Order Judgment

    Atsuhiro NISHI  Masanori YOKOYAMA  Ken-ichiro OGAWA  Taiki OGATA  Takayuki NOZAWA  Yoshihiro MIYAKE  

     
    PAPER-Office Information Systems, e-Business Modeling

      Vol:
    E97-D No:6
      Page(s):
    1567-1573

    The present study aims to investigate the effect of voluntary movements on human temporal perception in multisensory integration. We therefore performed temporal order judgment (TOJ) tasks in audio-tactile integration under three conditions: no movement, involuntary movement, and voluntary movement. It is known that the point of subjective simultaneity (PSS) under the no movement condition, that is, normal TOJ tasks, appears when a tactile stimulus is presented before an auditory stimulus. Our experiment showed that involuntary and voluntary movements shift the PSS to a value that reduces the interval between the presentations of auditory and tactile stimuli. Here, the shift of the PSS under the voluntary movement condition was greater than that under the involuntary movement condition. Remarkably, the PSS under the voluntary movement condition appears when an auditory stimulus slightly precedes a tactile stimulus. In addition, a just noticeable difference (JND) under the voluntary movement condition was smaller than those under the other two conditions. These results reveal that voluntary movements alternate the temporal integration of audio-tactile stimuli. In particular, our results suggest that voluntary movements reverse the temporal perception order of auditory and tactile stimuli and improve the temporal resolution of temporal perception. We discuss the functional mechanism of shifting the PSS under the no movement condition with voluntary movements in audio-tactile integration.

  • Quantizer Design Optimized for Distributed Estimation

    Yoon Hak KIM  

     
    LETTER-Fundamentals of Information Systems

      Vol:
    E97-D No:6
      Page(s):
    1639-1643

    We consider the problem of optimizing the quantizer design for distributed estimation systems where all nodes located at different sites collect measurements and transmit quantized data to a fusion node, which then produces an estimate of the parameter of interest. For this problem, the goal is to minimize the amount of information that the nodes have to transmit in order to attain a certain application accuracy. We propose an iterative quantizer design algorithm that seeks to find a non-regular mapping between quantization partitions and their codewords so as to minimize global distortion such as the estimation error. We apply the proposed algorithm to a system where an acoustic amplitude sensor model is employed at each node for source localization. Our experiments demonstrate that a significant performance gain can be achieved by our technique as compared with standard typical designs and even with distributed novel designs recently published.

  • Design of Small CRPA Arrays for Dual-Band GPS Applications

    Gangil BYUN  Seung Mo SEO  Ikmo PARK  Hosung CHOO  

     
    PAPER-Antennas and Propagation

      Vol:
    E97-B No:6
      Page(s):
    1130-1138

    This paper proposes the design of small CRPA arrays for dual-band Global Positioning System (GPS) applications. The array consists of five elements and is mounted on a circular ground platform with a diameter of 15-cm. Each antenna element has a coupled feed structure and consists of a feed patch and two radiating patches for dual-band operation. An external chip coupler is utilized for a broad circular polarization (CP) bandwidth, and its measured characteristics are taken into account in our simulation for more accurate performance estimation. Detailed parameters are optimized by using a genetic algorithm (GA) in conjunction with the FEKO EM simulator. The optimized antenna is fabricated on a ceramic substrate, and its performance is measured in a full anechoic chamber. Furthermore, a field test is also conducted to verify the signal-to-noise ratio (SNR) for real GPS satellite signals. The results prove that the proposed array is suitable for use in GPS CRPA applications.

  • Illumination Normalization-Based Face Detection under Varying Illumination

    Min YAO  Hiroshi NAGAHASHI  Kota AOKI  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E97-D No:6
      Page(s):
    1590-1598

    A number of well-known learning-based face detectors can achieve extraordinary performance in controlled environments. But face detection under varying illumination is still challenging. Possible solutions to this illumination problem could be creating illumination invariant features or utilizing skin color information. However, the features and skin colors are not sufficiently reliable under difficult lighting conditions. Another possible solution is to do illumination normalization (e.g., Histogram Equalization (HE)) prior to executing face detectors. However, applications of normalization to face detection have not been widely studied in the literature. This paper applies and evaluates various existing normalization methods under the framework of combining the illumination normalization and two learning-based face detectors (Haar-like face detector and LBP face detector). These methods were initially proposed for different purposes (face recognition or image quality enhancement), but some of them significantly improve the original face detectors and lead to better performance than HE according to the results of the comparative experiments on two databases. Meanwhile, we propose a new normalization method called segmentation-based half histogram stretching and truncation (SH) for face detection under varying illumination. It first employs Otsu method to segment the histogram (intensities) of the input image into several spans and then does the redistribution on the segmented spans. In this way, the non-uniform illumination can be efficiently compensated and local facial structures can be appropriately enhanced. Our method obtains good performance according to the experiments.

  • Image Retargeting with Protection of Object Arrangement

    Kazu MISHIBA  Takeshi YOSHITOME  

     
    PAPER-Image Processing and Video Processing

      Vol:
    E97-D No:6
      Page(s):
    1583-1589

    The relative arrangement, such as relative positions and orientations among objects, can play an important role in expressing the situation such as sports games and race scenes. In this paper, we propose a retargeting method that allows maintaining the relative arrangement. Our proposed retargeting method is based on a warping method which finds an optimal transformation by solving an energy minimization problem. To achieve protection of object arrangement, we introduce an energy that enforces all the objects and the relative positions among these objects to be transformed by the same transformation in the retargeting process. In addition, our method imposes the following three types of conditions in order to obtain more satisfactory results: protection of important regions, avoiding extreme deformation, and cropping with preservation of the balance of visual importance. Experimental results demonstrate that our proposed method maintains the relative arrangement while protecting important regions.

  • Longest Fault-Free Cycles in Folded Hypercubes with Conditional Faulty Elements

    Wen-Yin HUANG  Jia-Jie LIU  Jou-Ming CHANG  Ro-Yu WU  

     
    PAPER

      Vol:
    E97-A No:6
      Page(s):
    1187-1191

    An n-dimensional folded hypercube, denoted by FQn, is an enhanced n-dimensional hypercube with one extra link between nodes that have the furthest Hamming distance. Let FFv (respectively, FFe) denote the set of faulty nodes (respectively, faulty links) in FQn. Under the assumption that every fault-free node in FQn is incident to at least two fault-free links, Hsieh et al. (Inform. Process. Lett. 110 (2009) pp.41-53) showed that if |FFv|+|FFe| ≤ 2n-4 for n ≥ 3, then FQn-FFv-FFe contains a fault-free cycle of length at least 2n-2|FFv|. In this paper, we show that, under the same conditional fault model, FQn with n ≥ 5 can tolerate more faulty elements and provides the same lower bound of the length of a longest fault-free cycle, i.e., FQn-FFv-FFe contains a fault-free cycle of length at least 2n-2|FFv| if |FFv|+|FFe| ≤ 2n-3 for n ≥ 5.

  • Queue Layouts of Toroidal Grids

    Kung-Jui PAI  Jou-Ming CHANG  Yue-Li WANG  Ro-Yu WU  

     
    PAPER

      Vol:
    E97-A No:6
      Page(s):
    1180-1186

    A queue layout of a graph G consists of a linear order of its vertices, and a partition of its edges into queues, such that no two edges in the same queue are nested. The queuenumber qn(G) is the minimum number of queues required in a queue layout of G. The Cartesian product of two graphs G1 = (V1,E1) and G2 = (V2,E2), denoted by G1 × G2, is the graph with {:v1 ∈ V1 and v2 ∈ V2} as its vertex set and an edge (,) belongs to G1×G2 if and only if either (u1,v1) ∈ E1 and u2 = v2 or (u2,v2) ∈ E2 and u1 = v1. Let Tk1,k2,...,kn denote the n-dimensional toroidal grid defined by the Cartesian product of n cycles with varied lengths, i.e., Tk1,k2,...,kn = Ck1 × Ck2 × … × Ckn, where Cki is a cycle of length ki ≥ 3. If k1 = k2 = … = kn = k, the graph is also called the k-ary n-cube and is denoted by Qnk. In this paper, we deal with queue layouts of toroidal grids and show the following bound: qn(Tk1,k2,...,kn) ≤ 2n-2 if n ≥ 2 and ki ≥ 3 for all i = 1,2,...,n. In particular, for n = 2 and k1,k2 ≥ 3, we acquire qn(Tk1,k2) = 2. Recently, Pai et al. (Inform. Process. Lett. 110 (2009) pp.50-56) showed that qn(Qnk) ≤ 2n-1 if n ≥1 and k ≥9. Thus, our result improves the bound of qn(Qnk) when n ≥2 and k ≥9.

  • Feature Fusion for Blurring Detection in Image Forensics

    BenJuan YANG  BenYong LIU  

     
    LETTER-Image Processing and Video Processing

      Vol:
    E97-D No:6
      Page(s):
    1690-1693

    Artificial blurring is a typical operation in image forging. Most existing image forgery detection methods consider only one single feature of artificial blurring operation. In this manuscript, we propose to adopt feature fusion, with multifeatures for artificial blurring operation in image tampering, to improve the accuracy of forgery detection. First, three feature vectors that address the singular values of the gray image matrix, correlation coefficients for double blurring operation, and image quality metrics (IQM) are extracted and fused using principal component analysis (PCA), and then a support vector machine (SVM) classifier is trained using the fused feature extracted from training images or image patches containing artificial blurring operations. Finally, the same procedures of feature extraction and feature fusion are carried out on the suspected image or suspected image patch which is then classified, using the trained SVM, into forged or non-forged classes. Experimental results show the feasibility of the proposed method for image tampering feature fusion and forgery detection.

  • Fingerprint Verification and Identification Based on Local Geometric Invariants Constructed from Minutiae Points and Augmented with Global Directional Filterbank Features

    Chuchart PINTAVIROOJ  Fernand S. COHEN  Woranut IAMPA  

     
    PAPER-Image Recognition, Computer Vision

      Vol:
    E97-D No:6
      Page(s):
    1599-1613

    This paper addresses the problems of fingerprint identification and verification when a query fingerprint is taken under conditions that differ from those under which the fingerprint of the same person stored in a database was constructed. This occurs when using a different fingerprint scanner with a different pressure, resulting in a fingerprint impression that is smeared and distorted in accordance with a geometric transformation (e.g., affine or even non-linear). Minutiae points on a query fingerprint are matched and aligned to those on one of the fingerprints in the database, using a set of absolute invariants constructed from the shape and/or size of minutiae triangles depending on the assumed map. Once the best candidate match is declared and the corresponding minutiae points are flagged, the query fingerprint image is warped against the candidate fingerprint image in accordance with the estimated warping map. An identification/verification cost function using a combination of distance map and global directional filterbank (DFB) features is then utilized to verify and identify a query fingerprint against candidate fingerprint(s). Performance of the algorithm yields an area of 0.99967 (perfect classification is a value of 1) under the receiver operating characteristic (ROC) curve based on a database consisting of a total of 1680 fingerprint images captured from 240 fingers. The average probability of error was found to be 0.713%. Our algorithm also yields the smallest false non-match rate (FNMR) for a comparable false match rate (FMR) when compared to the well-known technique of DFB features and triangulation-based matching integrated with modeling non-linear deformation. This work represents an advance in resolving the fingerprint identification problem beyond the state-of-the-art approaches in both performance and robustness.

  • Semi-Supervised Learning via Geodesic Weighted Sparse Representation

    Jianqiao WANG  Yuehua LI  Jianfei CHEN  Yuanjiang LI  

     
    LETTER-Pattern Recognition

      Vol:
    E97-D No:6
      Page(s):
    1673-1676

    The label estimation technique provides a new way to design semi-supervised learning algorithms. If the labels of the unlabeled data can be estimated correctly, the semi-supervised methods can be replaced by the corresponding supervised versions. In this paper, we propose a novel semi-supervised learning algorithm, called Geodesic Weighted Sparse Representation (GWSR), to estimate the labels of the unlabeled data. First, the geodesic distance and geodesic weight are calculated. The geodesic weight is utilized to reconstruct the labeled samples. The Euclidean distance between the reconstructed labeled sample and the unlabeled sample equals the geodesic distance between the original labeled sample and the unlabeled sample. Then, the unlabeled samples are sparsely reconstructed and the sparse reconstruction weight is obtained by minimizing the L1-norm. Finally, the sparse reconstruction weight is utilized to estimate the labels of the unlabeled samples. Experiments on synthetic data and USPS hand-written digit database demonstrate the effectiveness of our method.

  • Motion Pattern Study and Analysis from Video Monitoring Trajectory

    Kai KANG  Weibin LIU  Weiwei XING  

     
    PAPER-Pattern Recognition

      Vol:
    E97-D No:6
      Page(s):
    1574-1582

    This paper introduces an unsupervised method for motion pattern learning and abnormality detection from video surveillance. In the preprocessing steps, trajectories are segmented based on their locations, and the sub-trajectories are represented as codebooks. Under our framework, Hidden Markov Models (HMMs) are used to characterize the motion pattern feature of the trajectory groups. The state of trajectory is represented by a HMM and has a probability distribution over the possible output sub-trajectories. Bayesian Information Criterion (BIC) is introduced to measure the similarity between groups. Based on the pairwise similarity scores, an affinity matrix is constructed which indicates the distance between different trajectory groups. An Adaptable Dynamic Hierarchical Clustering (ADHC) tree is proposed to gradually merge the most similar groups and form the trajectory motion patterns, which implements a simpler and more tractable dynamical clustering procedure in updating the clustering results with lower time complexity and avoids the traditional overfitting problem. By using the HMM models generated for the obtained trajectory motion patterns, we may recognize motion patterns and detect anomalies by computing the likelihood of the given trajectory, where a maximum likelihood for HMM indicates a pattern, and a small one below a threshold suggests an anomaly. Experiments are performed on EIFPD trajectory datasets from a structureless scene, where pedestrians choose their walking paths randomly. The experimental results show that our method can accurately learn motion patterns and detect anomalies with better performance.

5301-5320hit(21534hit)