Hiroshi KONDO Suharno AGUS Mariko MOROKUMA
One approach for image edge sharpening which includes a phase-only synthesis (POS) is presented. The technique presented here is the generalized version of the traditional Laplacian image enhancement. The utilizing of an internally dividing point between the POS and the Laplacian enhancement image makes this approach more flexible, more effective.
Youichi SATO Naoaki YAMANAKA Ken-ichi SATO
The benefits of ATM techniques have been widely recognized and many organizations envisage the introduction of ATM techniques into their telecommunication networks. The ATM benefits can, however, be fully exploited only after effective network resource management techniques have been developed. This paper focuses on CBR-VP management techniques. The ATM transport network architecture and VP roles are summarized. Next, the issues of VP accommodation design are described. The point is how to create a design that accommodates cell loss and cell delay jitter, both of which depend on various network parameters and conditions. For this purpose, analytical procedures based on an M/D/1 queueing model are adopted. The approximation method is shown to be very effective in practical use through computer analysis. The method guarantees conservative QOSs. Finally, the proposed method is applied to several design examples to illustrate VP management issues. The proposed method will enable ATM techniques to be introduced to our telecommunication networks by the mid-1990's.
Hector PEREZ-MEANA Mariko NAKANO-MIYATAKE Laura ORTIZ-BALBUENA Alejandro MARTINEZ-GONZALEZ Juan Carlos SANCHEZ-GARCIA
This letter propose a fast frequency domain adaptive filter algorithm (FADF) for applications in which large order adaptive filters are required. Proposed FADF algorithm reduces the block delay of conventional FADF algorithms allowing a more efficient selection of the fast Fourier Transform (FFT) size. Proposed FADF algorithm also provides faster convergence rates than conventional FBAF algorithms by using a near-optimum convergence factor derived by using the FFT. Computer simulations using white and colored signals are given to show the desirable features of proposed scheme.
Yasuhiro SUGIMOTO Takaaki TSUJI
This paper examines the feasibility of a high frequency (moro than 1 GHz) ring-oscillator-type CMOS VCO, able to maintain a good linearity between the oscillator output frequency and control voltage, while preserving low voltage and low power operation capabilities. A CMOS VCO circuit, with a newly developed corrent-controlled delay cell and an architecture combining the transitions of each delay cell output, with high-frequency operation, was designed and simulated using the CMOS 0.6 µm device paramenters. We analyzed the generation of unnecessary harmonics and sub-harmonics when a delay cell's propagation delay time varied. The simulation indicated that a CMOS VCO with a frequency range of 200 MHz to 1.4 GHz, a power dissipation of 8.5 mW at 900 MHz from a 3 V power supply, and an operation voltage of 1 V to 3 V can be implemented on a chip.
This paper is described on the realization of simulated inductance cercuit with parallel negative conductance and its application for an oscillator. The design's condition for realizing the circuit needs stability, narrow expance of elements, larger dynamic-range and lower sensitivity. A new floating simulated inductance circuit with parallel nagative conductance with two operational amplifiers, four resistors, and four capacitors is created by using the design's algorithm. And the elements sensitivity of the simulated circuit is superior to that of the conventional circuits. By experimenting with a resonance circuit, the author tested the sinusoidal oscillator's circuit of a parallel -GLC as an application in order to confirm the operation of the simulated inductance circuit with parallel negative conductance.
Kazuhiro NAKAI Gaishi YAMAMOTO Toshio NAKAMURA
A filter configuration that allows configuration of any transfer function used the state variable is discribed as an application of the second generation current conveyors (CCIIs) to RC networks. The filter types discussed are low-pass filter (LPF), high-pass filter (HPF), band-pass filter (BPF), all-pass filter (APF), and band-elimination filter (BEF). The filter circuit consists of four CCIIs and allows tandem connections. The device sensitivity and CCII's sensitivity to transfer coefficient are relatively low. The filter circuit that allow simultaneous configuration wewe fabricated. An experimental result at around 10kHz was obtained for the filters. In the case, the LPF, HPF, BPF, APF, and BEF characteristics are obtained at Q value of 5.0.
Fernando Gil V. RESENDE Jr. Keiichi TOKUDA Mineo KANEKO
A new adaptive AR spectral estimation method is proposed. While conventional least-squares methods use a single windowing function to analyze the linear prediction error, the proposed method uses a different window for each frequency band of the linear prediction error to define a cost function to be meinemized. With this approach, since time and frequency resolutions can be traded off throughout the frequency spectrum, an improvement on the precision of the estimates is achieved. In this paper, a wavelet-like time-frequency resolution grid is used so that low-frequency components of the linear prediction error are analyzed through long windows and high-frequency components are analyzed through short ones. To solve the optimization problem for the new cost function, special properties of the correlation matrix are used to derive an RLS algorithm on the order of M2, where M is the number of parameters of the AR model. Computer simulations comparing the performance of conventional RLS and the proposed methods are shown. In particular, it can be observed that the wavelet-based spectral estimation method gives fine frequency resolution at low frequencies and sharp time resolution at high frequencies, while with conventional methods it is possible to obtain only one of these characteristics.
Hiroshi TSURUMI Tadahiko MAEDA Hiroshi TANIMOTO Yasuo SUZUKI Masayuki SAITO Kunio YOSHIHARA Kenji ISHIDA Naotaka UCHITOMI
A miniature transceiver, including highly integrated MMIC front-end, for 1.9 GHz band personal handy phone system(PHS) has been developed. The terminal, adopting direct conversion transmitter and receiver technology, consists of four high-density RF circuit modules and a digital signal processing LSI with 2.7 V power supply. The four functional modules are a power amplifier, a transmitter,a receiver, and a frequency synthesizer. Each functional module includes one IC chip and passive LCR components connected with solder bumps on module substrate. The experimental miniature PHS handset has been fabricated to verify the design concepts of the miniature transceiver. The total volume of the developed PHS terminal is 60cc, including the 12cc front-end which comprises the four RF functional circuit modules. The air interface connection with the PHS base station simulator has been confirmed.
Toshio TSUTSUMIDA Toshihiro MATSUI Tadashi NOUMI Toru WAKAHARA
Through comparing the results of two successive IPTP Character Recognition Competitions which focused on 3-digit handprinted postal codes, we herein analyze the methodologies of the submitted algorithms along with the substituted or rejected patterns of these algorithms. Regarding their methodologies, lesser diversity was apparent specifically concerning the contour-chain code based on local stroke directions and statistical discriminant functions for feature extraction and discrimination. Analysis of the patterns demonstrated that the misrecognized patterns being most often improved were categorized as a decrease in peculiarly shaped handwritten characters or heavy-handed and disconnected strokes. However, most of the remaining misrecognitions were still classed as peculiarly shaped handwriting as commonly shared between the best three algorithms. From these analyses, we could delineate a direction to be taken for developing more effective methodologies and clarify the remaining problems to be overcome by the subsequent intensive research. Furthermore, we evaluate in this article our multi-expert recognition system for achieving higher recognition performances by means of combining complementary recognition algorithms. We performed a subsequent investigation of the Candidate Appearance Likelihood Method using novel experimental conditions and a new examination of the application of the neural network as the combining method for accumulating the broader candidate appearances. The results obtained confirm that combining through the neural network constitutes one of the most effective ways of making the multi-expert recognition system a reality.
Keiji GYOHTEN Tomoko SUMIYA Noboru BABAGUCHI Koh KAKUSHO Tadahiro KITAHASHI
This paper describes COCE (COordinative Character Extractor), a method for extracting printed Japanese characters and their character strings from all sorts of document images. COCE is based on a multi-agent system where each agent tries to find a character string and extracts the characters in it. For the adaptability, the agents are allowed to look after arbitrary parts of documents and extract the characters using only the knowledge independent of the layouts. Moreover, the agents check and correct their results sometimes with the help of the other agents. From experimental results, we have verified the effectiveness of our approach.
Naoaki YAMANAKA Francis PITCHO Hiroaki SATO
This letter studies the Peak Cell Rate (PCR) policing of ATM connections that consist of multiple cell flow components. It is shown that the conventional methods proposed for policing the aggregate flow do not use the network's resources efficiently. This letter proposes a simple and efficient UPC (Usage Parameter Control) mechanism based on a tandem leaky bucket for multi-component ATM connections. The results show that network resource requirements can be minimized, with reasonable hardware complexity.
Most conventional methods used in character recognition extract geometrical features, such as stroke direction and connectivity, and compare them with reference patterns in a stored dictionary. Unfortunately, geometrical features are easily degraded by blurs and stains, and by the graphical designs such as used in Japanese newspaper headlines. This noise must be removed before recognition commences, but no preprocessing method is perfectly accurate. This paper proposes a method for recognizing degraded characters as well as characters printed on graphical designs. This method extracts features from binary images, and a new similarity measure, the complementary similarity measure, is used as a discriminant function; it compares the similarity and dissimilarity of binary patterns with reference dictionary patterns. Experiments are conducted using the standard character database ETL-2, which consists of machine-printed Kanji, Hiragana, Katakana, alphanumeric, and special characters. The results show that our method is much more robust against noise than the conventional geometrical-feature method. It also achieves high recognition rates of over 97% for characters with textured foregrounds, over 99% for characters with textured backgrounds, over 98% for outline fonts and over 99% for reverse contrast characters. The experiments for recognizing both the fontstyles and character category show that it also achieves high recognition rates against noise.
Ju YE Masahiro TANAKA Tetsuzo TANINO
The problem of genetic algorithm's efficiency has been attracting the attention of genetic algorithm community. Over the last decade, considerable researches have focused on improving genetic algorithm's performance. However, they are generally under the framework of natural evolutionary mechanism and the major genetic operators, crossover and mutation, are activated by the prior probabilities. An operator based on a prior probability possesses randomness, that is, the unexpected individuals are frequently operated, but the expected individuals are sometimes not operated. Moreover, as the evaluation function is the link between the genetic algorithm and the problem to be solved, the evaluation function provides the heuristic information for evolutionary search. Therefore, how to use this kind of heuristic information (present and past) is influential in the efficiency of evolutionary search. This paper, as an attempt, presents a eugenics-based genetic algorithm (EGA) -- a genetic algorithm that reflects the human's decision will (eugenics), and fully utilizes the heuristic information provided by the evaluation function for the decisions. In other words, EGA = evolutionary mechanisms + human's decision will + heuristic information. In EGA, the ideas of the positive eugenics and the negative eugenics are applied as the principle of selections and the selections are not activated by the prior probabilities but by the evaluation values of individuals. A method of genealogical chain-based selection for mutation is proposed, which avoids the blindness of stochastic mutation and the disruptive problem of mutation. A control strategy of reasonable competitions is proposed, which brings the effects of crossover and mutation into full play. Three examples, the minimum problem of a standard optimizing function--De Jong's test function F2, a typical combinatorial optimization problem--the traveling salesman problem, and a problem of identifying nonlinear system, are given to show the good performance of EGA.
Among various characters invented by the human being ,Kanji (Chinese character) is outstanding in its diversity and complexity, in contrast to Roman alphanumerics. So machine recognition of handwritten characters requires particular technics. This paper concerns commercially available optical character readers (OCRs) and the recognition techniques using therein. Methods of character recognition is classified the pattern types and the method of extracting features, to discuss the present states of the character recognition. In regard to the structural analysis, the proposed techniques are classified and discussed in accordance with the extraction method of line segment and axial correlation. Finally, various techniques are compared with one another by using a common database, so as to understand the present states of character recognition and to discuss the technical trends.
Masato SUZUKI Nei KATO Hirotomo ASO Yoshiaki NEMOTO
In recognition of handprinted characters, it is important to dissolve distortions of character caused by writer's habits. In order to dissolve distortions and to obtain better features, many image conversion methods have been proposed. But there are distortions that cannot be dissolved by these methods. One example is the case of parallel strokes which are spread out in fan shape. In this paper, in order to dissolve distortions, we propose a new image conversion method, Transformation based on Partial Inclination Detection (TPID)", which is employed just before normalization, and is intended to dissolve several kinds of distortions in images of each character. TPID constructs transformation functions from inclination angles which are detected in some subspaces of the character's image, and converts images using the transformation functions. TPID is especially suitable for correcting the inclinations of horizontal and vertical strokes of a character. This has a powerful impact on the quality of the characteristic features. In recognition experiments using ETL9B, the largest database of handprinted characters in Japan, we have obtained a recognition rate of 99.08%, which is the best to our knowledge.
Kanad KEENI Hiroshi SHIMODAIRA Tetsuro NISHINO Yasuo TAN
Devanagari is the most widely used script in India. Here, a method is introduced for recognizing Devanagari characters using Neural network. The proposed method reduces the number of output unit necessary for a conventional neural network where the classification is based on a winner take all basis. An automatic coding procedure for representing the output layer of the network and a different method for the final classification is also proposed. Along with the automatic coding procedure, a heuristic method for representing the output units by exploiting the structural information of Devanagari character is also demonstrated. Besides, it has been shown by random representation of the output layer that the representation effects the generalization/performance of the network. The proposed automatic representation gave the recognition rate of 98.09% for 44 categories.
Table-form document structure analysis is an important problem in the document processing domain. This paper presents a new method called Box-Driven Reasoning (BDR) to robustly analyze the structure of table-form documents that include touching characters and broken lines. Real documents are copied repeatedly and overlaid with printed data, resulting in characters that touch cells and lines that are broken. Most previous methods employ a line-oriented approach, but touching characters and broken lines make the procedure fail at an early stage. BDR deals with regions directly in contrast with other previous methods and a reduced resolution image is introduced to supplement information deteriorated by noise. Experimental tests show that BDR reliably recognizes cells and strings in document images with touching characters and broken lines.
Tsutomu MIYASATO Haruo NOMA Fumio KISHINO
This paper describes the results of tests that measured the allowable delay between images and tactile information via a force feedback device. In order to investigate the allowable delay, two experiments were performed: 1) subjective evaluation in real space and 2) subjective evaluation in virtual space using a force feedback device.
For similarity methods to work well, the image must be blurred before being input. However, the relationship between the blurring operation and the similarity is not fully understood. To solve the problem of this relationship, in this paper, the effect of blurring is investigated by expressing figure f(x) in the form of the sum of higher derivatives of f (x,σ), and then a simple similarity between figures was mathematically formulated in terms of the relation between visual patterns. By modifying this formulation, we propose pluralized simple similarity to increase the allowance in different view of multiple similarity. The similarity maintains higher allowance without any discernible loss of distinguishing power. We verify the effectiveness of the pluralized simple similarity throughout some experiments.
Continuous nonlinearity" is stressed as a fundamental principle in pattern recognition including handprinted Kanji character recognition. Continuity" in template matching and spatial nonlinearity" in structural analysis should be unified toward deriving a higher level of recognition algorithm. At the same time, continuous nonlinearity in the temporal axis is important, as is the case of simultaneous processing of segmentation and recognition for touching characters. The above viewpoint is discussed in the following examples: nonlinear normalization, directional pattern matching, locally maximized similarity, relaxation matching, dynamic programming matching, segmentation of character string using dynamic programming, and exhaustive matching for character extraction on complex background.