The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] PAR(2741hit)

401-420hit(2741hit)

  • Blur Map Generation Based on Local Natural Image Statistics for Partial Blur Segmentation

    Natsuki TAKAYAMA  Hiroki TAKAHASHI  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2017/09/05
      Vol:
    E100-D No:12
      Page(s):
    2984-2992

    Partial blur segmentation is one of the most interesting topics in computer vision, and it has practical value. The generation of blur maps is a crucial part of partial blur segmentation because partial blur segmentation involves producing a blur map and applying a segmentation algorithm to the blur map. In this study, we address two important issues in order to improve the discrimination of blur maps: (1) estimating a robust local blur feature to consider variations in the intensity amplitude and (2) a scheme for generating blur maps. We propose the ANGHS (Amplitude-Normalized Gradient Histogram Span) as a local blur feature. ANGHS represents the heavy-tailedness of a gradient distribution, where it is calculated from an image gradient normalized using the intensity amplitude. ANGHS is robust to variations in the intensity amplitude, and it can handle local regions in a more appropriate manner than previously proposed local blur features. Blur maps are affected by local blur features but also by the contents and sizes of local regions, and the assignment of blur feature values to pixels. Thus, multiple-sized grids and the EAI (Edge-Aware Interpolation) are employed in each task to improve the discrimination of blur maps. The discrimination of the generated blur maps is evaluated visually and statistically using numerous partial blur images. Comparisons with the results obtained by state-of-the-art methods demonstrate the high discrimination of the blur maps generated using the proposed method.

  • Distributed Pareto Local Search for Multi-Objective DCOPs

    Maxime CLEMENT  Tenda OKIMOTO  Katsumi INOUE  

     
    PAPER-Information Network

      Pubricized:
    2017/09/15
      Vol:
    E100-D No:12
      Page(s):
    2897-2905

    Many real world optimization problems involving sets of agents can be modeled as Distributed Constraint Optimization Problems (DCOPs). A DCOP is defined as a set of variables taking values from finite domains, and a set of constraints that yield costs based on the variables' values. Agents are in charge of the variables and must communicate to find a solution minimizing the sum of costs over all constraints. Many applications of DCOPs include multiple criteria. For example, mobile sensor networks must optimize the quality of the measurements and the quality of communication between the agents. This introduces trade-offs between solutions that are compared using the concept of Pareto dominance. Multi-Objective Distributed Constraint Optimization Problems (MO-DCOPs) are used to model such problems where the goal is to find the set of Pareto optimal solutions. This set being exponential in the number of variables, it is important to consider fast approximation algorithms for MO-DCOPs. The bounded multi-objective max-sum (B-MOMS) algorithm is the first and only existing approximation algorithm for MO-DCOPs and is suited for solving a less-constrained problem. In this paper, we propose a novel approximation MO-DCOP algorithm called Distributed Pareto Local Search (DPLS) that uses a local search approach to find an approximation of the set of Pareto optimal solutions. DPLS provides a distributed version of an existing centralized algorithm by complying with the communication limitations and the privacy concerns of multi-agent systems. Experiments on a multi-objective extension of the graph-coloring problem show that DPLS finds significantly better solutions than B-MOMS for problems with medium to high constraint density while requiring a similar runtime.

  • Gauss-Seidel HALS Algorithm for Nonnegative Matrix Factorization with Sparseness and Smoothness Constraints

    Takumi KIMURA  Norikazu TAKAHASHI  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:12
      Page(s):
    2925-2935

    Nonnegative Matrix Factorization (NMF) with sparseness and smoothness constraints has attracted increasing attention. When these properties are considered, NMF is usually formulated as an optimization problem in which a linear combination of an approximation error term and some regularization terms must be minimized under the constraint that the factor matrices are nonnegative. In this paper, we focus our attention on the error measure based on the Euclidean distance and propose a new iterative method for solving those optimization problems. The proposed method is based on the Hierarchical Alternating Least Squares (HALS) algorithm developed by Cichocki et al. We first present an example to show that the original HALS algorithm can increase the objective value. We then propose a new algorithm called the Gauss-Seidel HALS algorithm that decreases the objective value monotonically. We also prove that it has the global convergence property in the sense of Zangwill. We finally verify the effectiveness of the proposed algorithm through numerical experiments using synthetic and real data.

  • An Efficient GPU Implementation of CKY Parsing Using the Bitwise Parallel Bulk Computation Technique

    Toru FUJITA  Koji NAKANO  Yasuaki ITO  Daisuke TAKAFUJI  

     
    PAPER-GPU computing

      Pubricized:
    2017/08/04
      Vol:
    E100-D No:12
      Page(s):
    2857-2865

    The main contribution of this paper is to present an efficient GPU implementation of bulk computation of the CKY parsing for a context-free grammar, which determines if a context-free grammar derives each of a lot of input strings. The bulk computation is to execute the same algorithm for a lot of inputs in turn or at the same time. The CKY parsing is to determine if a context-free grammar derives a given string. We show that the bulk computation of the CKY parsing can be implemented in the GPU efficiently using Bitwise Parallel Bulk Computation (BPBC) technique. We also show the rule minimization technique and the dynamic scheduling method for further acceleration of the CKY parsing on the GPU. The experimental results using NVIDIA TITAN X GPU show that our implementation of the bitwise-parallel CKY parsing for strings of length 32 takes 395µs per string with 131072 production rules for 512 non-terminal symbols.

  • A Region-Based Through-Silicon via Repair Method for Clustered Faults

    Tianming NI  Huaguo LIANG  Mu NIE  Xiumin XU  Aibin YAN  Zhengfeng HUANG  

     
    PAPER-Integrated Electronics

      Vol:
    E100-C No:12
      Page(s):
    1108-1117

    Three-dimensional integrated circuits (3D ICs) that employ through-silicon vias (TSVs) integrating multiple dies vertically have opened up the potential of highly improved circuit designs. However, various types of TSV defects may occur during the assembly process, especially the clustered TSV faults because of the winding level of thinned wafer, the surface roughness and cleanness of silicon dies,inducing TSV yield reduction greatly. To tackle this fault clustering problem, router-based and ring-based TSV redundancy architectures were previously proposed. However, these schemes either require too much area overhead or have limited reparability to tolerant clustered TSV faults. Furthermore, the repairing lengths of these schemes are too long to be ignored, leading to additional delay overhead, which may cause timing violation. In this paper, we propose a region-based TSV redundancy design to achieve relatively high reparability as well as low additional delay overhead. Simulation results show that for a given number of TSVs (8*8) and TSV failure rate (1%), our design achieves 11.27% and 20.79% reduction of delay overhead as compared with router-based design and ring-based scheme, respectively. In addition, the reparability of our proposed scheme is much better than ring-based design by 30.84%, while it is close to that of the router-based scheme. More importantly, the overall TSV yield of our design achieves 99.88%, which is slightly higher than that of both router-based method (99.53%) and ring-based design (99.00%).

  • An Efficient Weighted Bit-Flipping Algorithm for Decoding LDPC Codes Based on Log-Likelihood Ratio of Bit Error Probability

    Tso-Cho CHEN  Erl-Huei LU  Chia-Jung LI  Kuo-Tsang HUANG  

     
    PAPER-Fundamental Theories for Communications

      Pubricized:
    2017/05/29
      Vol:
    E100-B No:12
      Page(s):
    2095-2103

    In this paper, a weighted multiple bit flipping (WMBF) algorithman for decoding low-density parity-check (LDPC) codes is proposed first. Then the improved WMBF algorithm which we call the efficient weighted bit-flipping (EWBF) algorithm is developed. The EWBF algorithm can dynamically choose either multiple bit-flipping or single bit-flipping in each iteration according to the log-likelihood ratio of the error probability of the received bits. Thus, it can efficiently increase the convergence speed of decoding and prevent the decoding process from falling into loop traps. Compared with the parallel weighted bit-flipping (PWBF) algorithm, the EWBF algorithm can achieve significantly lower computational complexity without performance degradation when the Euclidean geometry (EG)-LDPC codes are decoded. Furthermore, the flipping criterion does not require any parameter adjustment.

  • Query Rewriting or Ontology Modification? Toward a Faster Approximate Reasoning on LOD Endpoints

    Naoki YAMADA  Yuji YAMAGATA  Naoki FUKUTA  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2017/09/15
      Vol:
    E100-D No:12
      Page(s):
    2923-2930

    On an inference-enabled Linked Open Data (LOD) endpoint, usually a query execution takes longer than on an LOD endpoint without inference engine due to its processing of reasoning. Although there are two separate kind of approaches, query modification approaches, and ontology modifications have been investigated on the different contexts, there have been discussions about how they can be chosen or combined for various settings. In this paper, for reducing query execution time on an inference-enabled LOD endpoint, we compare these two promising methods: query rewriting and ontology modification, as well as trying to combine them into a cluster of such systems. We employ an evolutionary approach to make such rewriting and modification of queries and ontologies based on the past-processed queries and their results. We show how those two approaches work well on implementing an inference-enabled LOD endpoint by a cluster of SPARQL endpoints.

  • A Segmentation Method of Single- and Multiple-Touching Characters in Offline Handwritten Japanese Text Recognition

    Kha Cong NGUYEN  Cuong Tuan NGUYEN  Masaki NAKAGAWA  

     
    PAPER-Pattern Recognition

      Pubricized:
    2017/08/23
      Vol:
    E100-D No:12
      Page(s):
    2962-2972

    This paper presents a method to segment single- and multiple-touching characters in offline handwritten Japanese text recognition with practical speed. Distortions due to handwriting and a mix of complex Chinese characters with simple phonetic and alphanumeric characters leave optical handwritten text recognition (OHTR) for Japanese still far from perfection. Segmentation of characters, which touch neighbors on multiple points, is a serious unsolved problem. Therefore, we propose a method to segment them which is made in two steps: coarse segmentation and fine segmentation. The coarse segmentation employs vertical projection, stroke-width estimation while the fine segmentation takes a graph-based approach for thinned text images, which employs a new bridge finding process and Voronoi diagrams with two improvements. Unlike previous methods, it locates character centers and seeks segmentation candidates between them. It draws vertical lines explicitly at estimated character centers in order to prevent vertically unconnected components from being left behind in the bridge finding. Multiple candidates of separation are produced by removing touching points combinatorially. SVM is applied to discard improbable segmentation boundaries. Then, ambiguities are finally solved by the text recognition employing linguistic context and geometric context to recognize segmented characters. The results of our experiments show that the proposed method can segment not only single-touching characters but also multiple-touching characters, and each component in our proposed method contributes to the improvement of segmentation and recognition rates.

  • Deformable Part Model Based Arrhythmia Detection Using Time Domain Features

    Yuuka HIRAO  Yoshinori TAKEUCHI  Masaharu IMAI  Jaehoon YU  

     
    PAPER-Digital Signal Processing

      Vol:
    E100-A No:11
      Page(s):
    2221-2229

    Heart disease is one of the major causes of death in many advanced countries. For prevention or treatment of heart disease, getting an early diagnosis from a long time period of electrocardiogram (ECG) examination is necessary. However, it could be a large burden on medical experts to analyze this large amount of data. To reduce the burden and support the analysis, this paper proposes an arrhythmia detection method based on a deformable part model, which absorbs individual variation of ECG waveform and enables the detection of various arrhythmias. Moreover, to detect the arrhythmia in low processing delay, the proposed method only utilizes time domain features. In an experimental result, the proposed method achieved 0.91 F-measure for arrhythmia detection.

  • Colored Magnetic Janus Particles Open Access

    Hiroshi YABU  

     
    INVITED PAPER

      Vol:
    E100-C No:11
      Page(s):
    955-957

    The aim of this research is realizing a high resolution and a fast color switching of electronic papers. In this report, we realized basis of electric papers comprised on magnetic Janus particles was established. Colored and magnetic Janus particles were successfully prepared, and magnetic Janus particles were introduced into honeycomb matrices. Introduced magnetic Janus particles quickly respond to an external magnetic field.

  • Formation of Polymer Walls by Monomer Aggregation Control Utilizing Substrate-Surface Wettability for Flexible LCDs Open Access

    Seiya KAWAMORITA  Yosei SHIBATA  Takahiro ISHINABE  Hideo FUJIKAKE  

     
    INVITED PAPER

      Vol:
    E100-C No:11
      Page(s):
    1005-1011

    We examined the novel aggregation control of the LC and monomer during formation of the polymer walls from a LC/monomer mixture in order to suppress the presence of the residual monomers and polymer networks in the pixel areas. The method is utilization of the differing wettabilities among LC and monomer molecules on a substrate surface. We patterned a substrate surface with a fluororesin and a polyimide film, and promoted phase separation of the LC and monomer by cooling process. This resulted in the LC and monomer aggregates primarily existing in the pixel areas and non-pixel areas, respectively. Moreover, the polymer-walls structure which was formed in this method partitioned into individual pixels in a lattice region and prevented the LC from flowing. This polymer-walls formation technique will be useful for developing high-quality flexible LCDs.

  • AIGIF: Adaptively Integrated Gradient and Intensity Feature for Robust and Low-Dimensional Description of Local Keypoint

    Songlin DU  Takeshi IKENAGA  

     
    PAPER-Vision

      Vol:
    E100-A No:11
      Page(s):
    2275-2284

    Establishing local visual correspondences between images taken under different conditions is an important and challenging task in computer vision. A common solution for this task is detecting keypoints in images and then matching the keypoints with a feature descriptor. This paper proposes a robust and low-dimensional local feature descriptor named Adaptively Integrated Gradient and Intensity Feature (AIGIF). The proposed AIGIF descriptor partitions the support region surrounding each keypoint into sub-regions, and classifies the sub-regions into two categories: edge-dominated ones and smoothness-dominated ones. For edge-dominated sub-regions, gradient magnitude and orientation features are extracted; for smoothness-dominated sub-regions, intensity feature is extracted. The gradient and intensity features are integrated to generate the descriptor. Experiments on image matching were conducted to evaluate performances of the proposed AIGIF. Compared with SIFT, the proposed AIGIF achieves 75% reduction of feature dimension (from 128 bytes to 32 bytes); compared with SURF, the proposed AIGIF achieves 87.5% reduction of feature dimension (from 256 bytes to 32 bytes); compared with the state-of-the-art ORB descriptor which has the same feature dimension with AIGIF, AIGIF achieves higher accuracy and robustness. In summary, the AIGIF combines the advantages of gradient feature and intensity feature, and achieves relatively high accuracy and robustness with low feature dimension.

  • Ball State Based Parallel Ball Tracking and Event Detection for Volleyball Game Analysis

    Xina CHENG  Norikazu IKOMA  Masaaki HONDA  Takeshi IKENAGA  

     
    PAPER-Vision

      Vol:
    E100-A No:11
      Page(s):
    2285-2294

    The ball state tracking and detection technology plays a significant role in volleyball game analysis, whose performance is limited due to the challenges include: 1) the inaccurate ball trajectory; 2) multiple numbers of the ball event category; 3) the large intra-class difference of one event. With the goal of broadcasting supporting for volleyball games which requires a real time system, this paper proposes a ball state based parallel ball tracking and event detection method based on a sequential estimation method such as particle filter. This method employs a parallel process of the 3D ball tracking and the event detection so that it is friendly for real time system implementation. The 3D ball tracking process uses the same models with the past work [8]. For event detection process, a ball event change estimation based multiple system model, a past trajectory referred hit point likelihood and a court-line distance feature based event type detection are proposed. First, the multiple system model transits the ball event state, which consists the event starting time and the event type, through three models dealing with different ball motion situations in the volleyball game, such as the motion keeping and changing. The mixture of these models is decided by estimation of the ball event change estimation. Secondly, the past trajectory referred hit point likelihood avoids the processing time delay between the ball tracking and the event detection process by evaluating the probability of the ball being hit at certain time without using future ball trajectories. Third, the feature of the distance between the ball and the specific court line are extracted to detect the ball event type. Experimental results based on multi-view HDTV video sequences (2014 Inter High School Men's Volleyball Games, Japan), which contains 606 events in total, show that the detection rate reaches 88.61% while the success rate of 3D ball tracking keeps more than 99%.

  • Identification of Time-Varying Parameters of Hybrid Dynamical System Models and Its Application to Driving Behavior

    Thomas WILHELEM  Hiroyuki OKUDA  Tatsuya SUZUKI  

     
    PAPER-Systems and Control

      Vol:
    E100-A No:10
      Page(s):
    2095-2105

    This paper presents a novel identification method for hybrid dynamical system models, where parameters have stochastic and time-varying characteristics. The proposed parameter identification scheme is based on a modified implementation of particle filtering, together with a time-smoothing technique. Parameters of the identified model are considered as time-varying random variables. Parameters are identified independently at each time step, using the Bayesian inference implemented as an iterative particle filtering method. Parameters time dynamics are smoothed using a distribution based moving average technique. Modes of the hybrid system model are handled independently, allowing any type of nonlinear piecewise model to be identified. The proposed identification scheme has low computation burden, and it can be implemented for online use. Effectiveness of the scheme is verified by numerical experiments, and an application of the method is proposed: analysis of driving behavior through identified time-varying parameters.

  • Fraud Detection in Comparison-Shopping Services: Patterns and Anomalies in User Click Behaviors

    Sang-Chul LEE  Christos FALOUTSOS  Dong-Kyu CHAE  Sang-Wook KIM  

     
    LETTER-Artificial Intelligence, Data Mining

      Pubricized:
    2017/07/10
      Vol:
    E100-D No:10
      Page(s):
    2659-2663

    This paper deals with a novel, interesting problem of detecting frauds in comparison-shopping services (CSS). In CSS, there exist frauds who perform excessive clicks on a target item. They aim at making the item look very popular and subsequently ranked high in the search and recommendation results. As a result, frauds may distort the quality of recommendations and searches. We propose an approach of detecting such frauds by analyzing click behaviors of users in CSS. We evaluate the effectiveness of the proposed approach on a real-world clickstream dataset.

  • Ground Plane Detection with a New Local Disparity Texture Descriptor

    Kangru WANG  Lei QU  Lili CHEN  Jiamao LI  Yuzhang GU  Dongchen ZHU  Xiaolin ZHANG  

     
    LETTER-Pattern Recognition

      Pubricized:
    2017/06/27
      Vol:
    E100-D No:10
      Page(s):
    2664-2668

    In this paper, a novel approach is proposed for stereo vision-based ground plane detection at superpixel-level, which is implemented by employing a Disparity Texture Map in a convolution neural network architecture. In particular, the Disparity Texture Map is calculated with a new Local Disparity Texture Descriptor (LDTD). The experimental results demonstrate our superior performance in KITTI dataset.

  • Next-Activity Set Prediction Based on Sequence Partitioning to Reduce Activity Pattern Complexity in the Multi-User Smart Space

    Younggi KIM  Younghee LEE  

     
    PAPER-Pattern Recognition

      Pubricized:
    2017/07/18
      Vol:
    E100-D No:10
      Page(s):
    2587-2596

    Human activity prediction has become a prerequisite for service recommendation and anomaly detection systems in a smart space including ambient assisted living (AAL) and activities of daily living (ADL). In this paper, we present a novel approach to predict the next-activity set in a multi-user smart space. Differing from the majority of the previous studies considering single-user activity patterns, our study considers multi-user activities that occur with a large variety of patterns. Its complexity increases exponentially according to the number of users. In the multi-user smart space, there can be inevitably multiple next-activity candidates after multi-user activities occur. To solve the next-activity problem in a multi-user situation, we propose activity set prediction rather than one activity prediction. We also propose activity sequence partitioning to reduce the complexity of the multi-user activity pattern. This divides an activity sequence into start, ongoing, and finish zones based on the features in the tendency of activity occurrences. The majority of the activities in a multi-user environment occur at the beginning or end, rather than the middle, of an activity sequence. Furthermore, the types of activities typically occurring in each zone can be sufficiently distinguishable. Exploiting these characteristics, we suggest a two-step procedure to predict the next-activity set utilizing a long short-term memory (LSTM) model. The first step identifies the zones to which current activities belong. In the next step, we construct three different LSTM models to predict the next-activity set in each zone. To evaluate the proposed approach, we experimented using a real dataset generated from our campus testbed. Our experiments confirmed the complexity reduction and high accuracy in the next-activity set prediction. Thus, it can be effectively utilized for various applications with context-awareness in a multi-user smart space.

  • Random-Valued Impulse Noise Removal Using Non-Local Search for Similar Structures and Sparse Representation

    Kengo TSUDA  Takanori FUJISAWA  Masaaki IKEHARA  

     
    PAPER-Image

      Vol:
    E100-A No:10
      Page(s):
    2146-2153

    In this paper, we introduce a new method to remove random-valued impulse noise in an image. Random-valued impulse noise replaces the pixel value at a random position by a random value. Due to the randomness of the noisy pixel values, it is difficult to detect them by comparison with neighboring pixels, which is used in many conventional methods. Then we improve the recent noise detector which uses a non-local search of similar structure. Next we propose a new noise removal algorithm by sparse representation using DCT basis. Furthermore, the sparse representation can remove impulse noise by using the neighboring similar image patch. This method has much more superior noise removal performance than conventional methods at images. We confirm the effectiveness of the proposed method quantitatively and qualitatively.

  • Fast Parameter Estimation for Polyphase P Codes Modulated Radar Signals

    Qi ZHANG  Pei WANG  Jun ZHU  Bin TANG  

     
    LETTER-Digital Signal Processing

      Vol:
    E100-A No:10
      Page(s):
    2162-2166

    A fast parameter estimation method with a coarse estimation and a fine estimation for polyphase P coded signals is proposed. For a received signal with N sampling points, the proposed method has an improved performance when the signal-to-noise ratio (SNR) is larger than 2dB and a lower computational complexity O(N logs N) compared with the latest time-frequency rate estimation method whose computational complexity is O(N2).

  • Generalized Framework to Attack RSA with Special Exposed Bits of the Private Key

    Shixiong WANG  Longjiang QU  Chao LI  Shaojing FU  

     
    PAPER-Cryptography and Information Security

      Vol:
    E100-A No:10
      Page(s):
    2113-2122

    In this paper, we study partial key exposure attacks on RSA where the number of unexposed blocks of the private key is greater than or equal to one. This situation, called generalized framework of partial key exposure attack, was first shown by Sarkar [22] in 2011. Under a certain condition for the values of exposed bits, we present a new attack which needs fewer exposed bits and thus improves the result in [22]. Our work is a generalization of [28], and the approach is based on Coppersmith's method and the technique of unravelled linearization.

401-420hit(2741hit)