The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] OMP(3945hit)

3161-3180hit(3945hit)

  • Simulation Algorithms among Enhanced Mesh Models

    Susumu MATSUMAE  Nobuki TOKURA  

     
    PAPER-Algorithm and Computational Complexity

      Vol:
    E82-D No:10
      Page(s):
    1324-1337

    In this paper, we present simulation algorithms among enhanced mesh models. The enhanced mesh models here include reconfigurable mesh and mesh with multiple broadcasting. A reconfigurable mesh (RM) is a processor array that consists of processors arranged to a 2-dimensional grid with a reconfigurable bus system. The bus system can be used to dynamically obtain various interconnection patterns among the processors during the execution of programs. A horizontal-vertical RM (HV-RM) is obtained from the general RM model, by restricting the network topology it can take to the ones in which each bus segment must be along row or column. A mesh with multiple broadcasting (MWMB) is an enhanced mesh, which has additional broadcasting buses endowed to every row and column. We present two algorithms:1) an algorithm that simulates a HV-RM of size nn time-optimally in θ(n) time on a MWMB of size nn, and 2) an algorithm that simulates a RM of size nn in θ(log2 n) time on a HV-RM of size nn. Both algorithms use a constant number of storage in each processor. Furthermore, we show that a RM of size nn can be simulated in θ((n/m)2 log n log m) time on a HV-RM of size mm, in θ ((n/m)2 m log n log m) time on a MWMB of size mm (m < n). These simulations use θ((n/m)2) storage in each processor, which is optimal.

  • Universal Variable-to-Fixed Length Codes Achieving Optimum Large Deviations Performance for Empirical Compression Ratio

    Tomohiko UYEMATSU  

     
    PAPER-Information Theory and Coding Theory

      Vol:
    E82-A No:10
      Page(s):
    2246-2250

    This paper clarifies two variable-to-fixed length codes which achieve optimum large deviations performance of empirical compression ratio. One is Lempel-Ziv code with fixed number of phrases, and the other is an arithmetic code with fixed codeword length. It is shown that Lempel-Ziv code is asymptotically optimum in the above sense, for the class of finite-alphabet and finite-state sources, and that the arithmetic code is asymptotically optimum for the class of finite-alphabet unifilar sources.

  • Pattern Formation in Reaction-Diffusion Enzyme Transistor Circuits

    Masahiko HIRATSUKA  Takafumi AOKI  Tatsuo HIGUCHI  

     
    PAPER

      Vol:
    E82-A No:9
      Page(s):
    1809-1817

    This paper explores a possibility of constructing massively parallel molecular computing systems using molecular electronic devices called enzyme transistors. The enzyme transistor is, in a sense, an artificial catalyst which selects a specific substrate molecule and transforms it into a specific product. Using this primitive function, various active continuous media for signal transfer/processing can be realized. Prominent examples discussed in this paper are: (i) Turing pattern formation and (ii) excitable wave propagation in a two-dimensional enzyme transistor array. This paper demonstrates the potential of enzyme transistors for creating reaction-diffusion dynamics that performs useful computations in a massively parallel fashion.

  • A Low-Power Half-Swing Clocking Scheme for Flip-Flop with Complementary Gate and Source Drive

    Jin-Cheon KIM  Sang-Hoon LEE  Hong-June PARK  

     
    LETTER-Integrated Electronics

      Vol:
    E82-C No:9
      Page(s):
    1777-1779

    A half-swing clocking scheme with a complementary gate and source drive is proposed for a CMOS flip-flop to reduce the power consumption of the clock system by 43%, while keeping the flip-flop delay time the same as that of the conventional full-swing clocking scheme. The delay time of the preceding half stage of a flip-flop using this scheme is less than half of that using the previous half-swing clocking scheme.

  • Computational Complexity of Finding Meaningful Association Rules

    Yeon-Dae KWON  Ryuichi NAKANISHI  Minoru ITO  Michio NAKANISHI  

     
    PAPER-Algorithms and Data Structures

      Vol:
    E82-A No:9
      Page(s):
    1945-1952

    Recent developments in computer technology allow us to analyze all the data in a huge database. Data mining is to analyze all the data in such a database and to obtain useful information for database users. One of the well-studied problems in data mining is the search for meaningful association rules in a market basket database which contains massive amounts of transactions. One way to find meaningful association rules is to find all the large itemsets first, and then to find meaningful association rules from the large itemsets. Although a number of algorithms for computing all the large itemsets have been proposed, the computational complexity of them is scarcely disscussed. In this paper, we show that it is NP-complete to decide whether there exists a large itemset that has a given cardinality. Also, we propose subclasses of databases in which all the meaningful association rules can be computed in time polynomial of the size of a database.

  • Competitive Learning Methods with Refractory and Creative Approaches

    Michiharu MAEDA  Hiromi MIYAJIMA  

     
    PAPER

      Vol:
    E82-A No:9
      Page(s):
    1825-1833

    This paper presents two competitive learning methods with the objective of avoiding the initial dependency of weight (reference) vectors. The first is termed the refractory and competitive learning algorithm. The algorithm has a refractory period: Once the cell has fired, a winner unit corresponding to the cell is not selected until a certain amount of time has passed. Thus, a specific unit does not become a winner in the early stage of processing. The second is termed the creative and competitive learning algorithm. The algorithm is presented as follows: First, only one output unit is prepared at the initial stage, and a weight vector according to the unit is updated under the competitive learning. Next, output units are created sequentially to a prespecified number based on the criterion of the partition error, and competitive learning is carried out until the ternimation condition is satisfied. Finally, we discuss algorithms which have little dependence on the initial values and compare them with the proposed algorithms. Experimental results are presented in order to show that the proposed methods are effective in the case of average distortion.

  • CooPs: A Cooperative Process Planning System to Negotiate Process Change Requests

    Kagetomo GENJI  Katsuro INOUE  

     
    PAPER-Sofware System

      Vol:
    E82-D No:9
      Page(s):
    1261-1277

    In order to lead an ongoing software project to success, it is important to flexibly control its dynamically-changing software process. However, it is generally impossible not only to exactly pre-define the production process but also to prescribe the process change process (meta-process). To solve the problem, we have focused on communication between the project staff through which process change requests presented by individuals can be immediately shared, designed, verified, validated and implemented. This paper proposes a communication model which can represent a wide variety of communication states between the project manager and developers discussing how to implement process change requests. The communication model has been derived by investigating the sort of process change requests and, based on the model, we have implemented a cooperative process planning system (called CooPs). CooPs is a communication environment designed for software projects and supports information sharing for discussing the process change requests. By using CooPs, the software project can flexibly deal with not only expected change requests but also unexpected ones. To evaluate the applicability of the communication model and the capabilities of CooPs, we have conducted an experiment which is an application of CooPs to the ISPW6 example problem. This paper describes the concepts of CooPs, the system implementation, and the experiment.

  • Fully-Parallel VLSI Implementation of Vector Quantization Processor Using Neuron-MOS Technology

    Akira NAKADA  Masahiro KONDA  Tatsuo MORIMOTO  Takemi YONEZAWA  Tadashi SHIBATA  Tadahiro OHMI  

     
    PAPER-Processors

      Vol:
    E82-C No:9
      Page(s):
    1730-1738

    An analog vector quantization processor has been designed based on the neuron-MOS (νMOS) technology. In order to achieve a high integrating density, template information is merged into the matching cell (the absolute value circuitry) using the νMOS ROM technology. A new-architecture νMOS winner-take-all (WTA) circuit is employed for fully-parallel search for the minimum-distance vector. The WTA performs multi-resolution winner search with an automatic feedback gain control. A test chip having 256 16-element fixed template vectors has been built in a 1.5-µm double-polysilicon CMOS technology with the chip size of 7.2 mm 7.2 mm, and the basic operation of the circuits has been demonstrated.

  • A Code-Division Multiplexing Technique for Efficient Data Transmission in VLSI Systems

    Yasushi YUMINAKA  Kazuhiko ITOH  Yoshisato SASAKI  Takafumi AOKI  Tatsuo HIGUCHI  

     
    PAPER-Non-Binary Architectures

      Vol:
    E82-C No:9
      Page(s):
    1669-1677

    This paper proposes applications of a code-division multiplexing technique to VLSI systems free from interconnection problems. We employ a pseudo-random orthogonal m-sequence carrier as a multiplexable information carrier to achieve efficient data transmission. Using orthogonal property of m-sequences, we can multiplex several computational activities into a single circuit, and execute in parallel using multiplexed data transmission with reduced interconnection. Also, randomness of m-sequences offers the high tolerance to interference (jamming), and suppression of dynamic range of signals while maintaining a sufficient signal-to-noise ratio (SNR). We demonstrate application examples of multiplex computing circuits, neural networks, and spread-spectrum image processing to show the advantages.

  • Analog Computation Using Coupled-Quantum-Dot Spin Glass

    Nan-Jian WU  Hassu LEE  Yoshihito AMEMIYA  Hitoshi YASUNAGA  

     
    PAPER-Quantum Devices and Circuits

      Vol:
    E82-C No:9
      Page(s):
    1623-1629

    A novel analog-computation system using quantum-dot spin glass is proposed. Analog computation is a processing method that solves a mathematical problem by applying an analogy of a physical system to the problem. A 2D array of quantum dots is constructed by mixing two-dot (antiferromagnetic interaction) and three-dot (ferromagnetic interaction) systems. The simulation results show that the array shows spin-glass-like behavior. We then mapped two combinatorial optimization problems onto the quantum-dot spin glasses, and found their optimal solutions. The results demonstrate that quantum-dot spin glass can perform analog computation and solve a complex mathematical problem.

  • Signed-Weight Arithmetic and Its Application to a Field-Programmable Digital Filter Architecture

    Takafumi AOKI  Yoshiki SAWADA  Tatsuo HIGUCHI  

     
    PAPER-Configurable Computing and Fault Tolerance

      Vol:
    E82-C No:9
      Page(s):
    1687-1698

    This paper presents a new number representation called the Signed-Weight (SW) number system, which is useful for designing configurable counter-tree architectures for digital signal processing applications. The SW number system allows the unified manipulation of positive and negative numbers in arithmetic circuits by adjusting the signs assigned to individual digit positions. This makes possible the construction of highly regular arithmetic circuits without introducing irregular arithmetic operations, such as negation and sign extension in the two's complement representation. This paper also presents the design of a Field-Programmable Digital Filter (FPDF) architecture--a special-purpose FPGA architecture for high-speed FIR filtering--using the proposed SW arithmetic system.

  • Skew-Compensation Technique for Parallel Optical Interconnections

    Takeshi SAKAMOTO  Nobuyuki TANAKA  Yasuhiro ANDO  

     
    PAPER-Optical Systems and Technologies

      Vol:
    E82-B No:8
      Page(s):
    1162-1168

    We have developed a low-latency, error-correcting-code-(ECC-)adaptable skew-compensation technique, which is needed for high-speed and long-distance parallel optical interconnections. A new frame-coding technique called shuffled mB1C encoding, which requires no clock-rate conversion circuit and no data buffering, and a new skew-measurement method which is suitable for ECC adaptation have been developed for the compensation. Full-digital skew-compensation circuits using these new techniques were able to compensate for a two-clock-cycle skew, even when one transmission channel was removed. The maximum latency for skew compensation was only five clock cycles.

  • Skew-Compensation Technique for Parallel Optical Interconnections

    Takeshi SAKAMOTO  Nobuyuki TANAKA  Yasuhiro ANDO  

     
    PAPER-Optical Systems and Technologies

      Vol:
    E82-C No:8
      Page(s):
    1428-1434

    We have developed a low-latency, error-correcting-code-(ECC-)adaptable skew-compensation technique, which is needed for high-speed and long-distance parallel optical interconnections. A new frame-coding technique called shuffled mB1C encoding, which requires no clock-rate conversion circuit and no data buffering, and a new skew-measurement method which is suitable for ECC adaptation have been developed for the compensation. Full-digital skew-compensation circuits using these new techniques were able to compensate for a two-clock-cycle skew, even when one transmission channel was removed. The maximum latency for skew compensation was only five clock cycles.

  • Comparative Study of Discrete Orthogonal Transforms in Adaptive Signal Processing

    Susanto RAHARDJA  Bogdan J. FALKOWSKI  

     
    PAPER

      Vol:
    E82-A No:8
      Page(s):
    1386-1390

    In this paper, comparison of various orthogonal transforms in Wiener filtering is discussed. The study involves the family of discrete orthogonal transforms called Complex Hadamard Transform, which has been recently introduced by the same authors. Basic definitions, properties and transformation kernel of Complex Hadamard Transform are also shown.

  • Parameter Estimation of Inhomogeneous AR Model Expanded with Unknown Basis

    Yukiko YOKOYAMA  Mineo KUMAZAWA  Naoki MIKAMI  

     
    LETTER

      Vol:
    E82-A No:8
      Page(s):
    1582-1587

    We proposed a new model for non-stationary time series analysis based on the IAR (inhomogeneous autoregressive) model, and a method for model parameter estimation when the set of basis is given. In this paper, we further propose a method for parameter estimation including that of basis set: we set a new condition that power of the input sequence is concentrated in low-frequency domain, and developed an iterative estimation method. We firstly select an initial set of basis, from which new sets are created in order to minimize the difference between the model and data. Among new sets of basis, we select a good one that gives minimum standard deviation of estimated frequencies.

  • A Two-Stage Discrete Optimization Method for Largest Common Subgraph Problems

    Nobuo FUNABIKI  Junji KITAMICHI  

     
    PAPER-Algorithm and Computational Complexity

      Vol:
    E82-D No:8
      Page(s):
    1145-1153

    A novel combinatorial optimization algorithm called 2-stage discrete optimization method (2DOM) is proposed for the largest common subgraph problem (LCSP) in this paper. Given two graphs G=(V1, E1) and H=(V2, E2), the goal of LCSP is to find a subgraph G'=(V1', E1') of G and a subgraph H'=(V2', E2') of H such that G' and H' are not only isomorphic to each other but also their number of edges is maximized. The two graphs G' and H' are isomorphic when |V1'|=|V2'| and |E1'|=|E2'|, and there exists one-to-one vertex correspondence f: V1' V2' such that {u, v} E1' if and only if{f(u), f(v)} E2'. LCSP is known to be NP-complete in general. The 2DOM consists of a construction stage and a refinement stage to achieve the high solution quality and the short computation time for large size difficult combinatorial optimization problems. The construction stage creates a feasible initial solution with considerable quality, based on a greedy heuristic method. The refinement stage improves it keeping the feasibility, based on a random discrete descent method. The performance is evaluated by solving two types of randomly generated 1200 LCSP instances with a maximum of 500 vertices for G and 1000 vertices for H. The simulation result shows the superiority of 2DOM to the simulated annealing in terms of the solution quality and the computation time.

  • A Memory Reduction Approach for MPEG Decoding System

    Hideo OHIRA  Fumitoshi KARUBE  

     
    LETTER

      Vol:
    E82-A No:8
      Page(s):
    1588-1591

    An approach to an MPEG decoding system with reduced memory capacity will be presented. This method relies on the simple technique of one-dimensional DPCM to recompress reconstructed Macro Block (MB) prior to being stored on frame memory. Simulation results suggest that image quality is subjectively acceptable when using approximately one-half of the memory size required by that of conventional decoder. The degradation in the signal-to-noise ratio introduced by this compression method ranged from 0.1 dB to 0.7 dB for MPEG MP@ML standard test sequences at 4 Mbps. This technique can be implemented to achieve a cost effective MPEG decoder.

  • A Hybrid Speech Coder Based on CELP and Sinusoidal Coding

    Mohammad NAKHAI  Farokh MARVASTI  

     
    PAPER-Speech Processing and Acoustics

      Vol:
    E82-D No:8
      Page(s):
    1190-1199

    In this paper, we study a new hybrid speech coder which employs a modified version of the harmonic sinusoidal analysis to encode the periodic contents of speech waveform and to split the speech spectrum into two frequency regions of harmonic and random components. A reliable fundamental frequency is estimated for the harmonic region using both speech and its linear predictive (LP) residual spectrum. The peak envelope of speech spectrum is encoded in terms of the coefficients of an all-pole spectrum. A harmonic tracking algorithm appropriately interpolates the sinusoidal parameters to achieve a smooth transition between the parameter update points and to reconstruct an essential level of periodicity in the synthetic voiced speech. The random part of spectrum and unvoiced speech are coded using the conventional CELP algorithm. The individual components are then combined at the decoder to obtain the synthetic speech. The proposed hybrid coder which combines the powerful features of the sinusoidal and CELP coding algorithms yeilds a high quality synthetic speech at 4.05 kbps.

  • Comparison of Performance between AND and Majority Logic Type Nonlinear Feedforward Logic Pseudonoise Sequence Generators

    Kari H. A. KARKKAINEN  

     
    PAPER-Spread Spectrum Technologies and Applications

      Vol:
    E82-A No:8
      Page(s):
    1641-1647

    Two classes of nonlinear feedforward logic (NLFFL) pseudonoise (PN) code generators based on the use of AND and majority logic (ML) gates are compared. Cross-correlation and code-division multiple-access (CDMA) properties of properly designed NLFFL sequences are found to be comparable with the properties of well-known linear PN codes. It is determined that code design employing ML gates with an odd number of inputs is easier compared with designing with AND gates. This is especially true when the degree of nonlinearity is large, since the nonbalance problem, e. g. , at the output of an AND gate, can be avoided. ML type sequences are less vulnerable to correlation attack and jamming by the m-sequence of an NLFFL generator

  • Speech Enhancement Using Nonlinear Microphone Array Based on Complementary Beamforming

    Hiroshi SARUWATARI  Shoji KAJITA  Kazuya TAKEDA  Fumitada ITAKURA  

     
    PAPER

      Vol:
    E82-A No:8
      Page(s):
    1501-1510

    This paper describes a spatial spectral subtraction method by using the complementary beamforming microphone array to enhance noisy speech signals for speech recognition. The complementary beamforming is based on two types of beamformers designed to obtain complementary directivity patterns with respect to each other. In this paper, it is shown that the nonlinear subtraction processing with complementary beamforming can result in a kind of the spectral subtraction without the need for speech pause detection. In addition, the optimization algorithm for the directivity pattern is also described. To evaluate the effectiveness, speech enhancement experiments and speech recognition experiments are performed based on computer simulations under both stationary and nonstationary noise conditions. In comparison with the optimized conventional delay-and-sum (DS) array, it is shown that: (1) the proposed array improves the signal-to-noise ratio (SNR) of degraded speech by about 2 dB and performs more than 20% better in word recognition rates under the conditions that the white Gaussian noise with the input SNR of -5 or -10 dB is used, (2) the proposed array performs more than 5% better in word recognition rates under the nonstationary noise conditions. Also, it is shown that these improvements of the proposed array are same as or superior to those of the conventional spectral subtraction method cascaded with the DS array.

3161-3180hit(3945hit)