1-5hit |
Mauricio KUGLER Teemu TOSSAVAINEN Miku NAKATSU Susumu KUROYANAGI Akira IWATA
The development of assistive devices for automated sound recognition is an important field of research and has been receiving increased attention. However, there are still very few methods specifically developed for identifying environmental sounds. The majority of the existing approaches try to adapt speech recognition techniques for the task, usually incurring high computational complexity. This paper proposes a sound recognition method dedicated to environmental sounds, designed with its main focus on embedded applications. The pre-processing stage is loosely based on the human hearing system, while a robust set of binary features permits a simple k-NN classifier to be used. This gives the system the capability of in-field learning, by which new sounds can be simply added to the reference set in real-time, greatly improving its usability. The system was implemented in an FPGA based platform, developed in-house specifically for this application. The design of the proposed method took into consideration several restrictions imposed by the hardware, such as limited computing power and memory, and supports up to 12 reference sounds of around 5.3 s each. Experimental results were performed in a database of 29 sounds. Sensitivity and specificity were evaluated over several random subsets of these signals. The obtained values for sensitivity and specificity, without additional noise, were, respectively, 0.957 and 0.918. With the addition of +6 dB of pink noise, sensitivity and specificity were 0.822 and 0.942, respectively. The in-field learning strategy presented no significant change in sensitivity and a total decrease of 5.4% in specificity when progressively increasing the number of reference sounds from 1 to 9 under noisy conditions. The minimal signal-to-noise ration required by the prototype to correctly recognize sounds was between -8 dB and 3 dB. These results show that the proposed method and implementation have great potential for several real life applications.
Takeshi SAITOH Ryosuke KONISHI
This paper describes a recognition method of Japanese single sounds for application to lip reading. Related researches investigated only five or ten sounds. In this paper, experiments were conducted for 45 Japanese single sounds by classifying them into five vowels category, ten consonants category, and 45 sounds category. We obtained recognition rates of 94.7, 30.9 and 30.0% with trajectory feature.
Jun-Pyo HONG Jung-Jun LEE Sang-Bong JUNG Seung-Hong HONG
Heart sound is an acoustic wave generated by the mechanical movement of the heart and blood flow, and is a complicated, non-stationary signal composed of many signal sources. It can be divided into normal heart sounds and heart murmurs. Murmurs are abnormal signals that appear over wider ranges of frequency than normal heart sounds. They are generated at random spots in the whole period of heart sounds. The recognition of heart sounds is to differentiate heart murmurs through patterns that appear according to the generation time of murmurs. In this paper, a group of heart sounds was classified into normal heart sounds, pre-systolic murmurs, early systolic murmurs, late systolic murmurs, early diastolic murmurs, and continuous murmurs. The suggested algorithm was standardized by re-sampling and then added as an input to the neural network through wavelet transform. The neural network used Error Back - Propagation algorithm, which is a representative learning method, and controlled the number of hidden layers and the learning rate for optimal construction of networks. As a result of recognition, the suggested algorithm obtained a higher recognition rate than that of existing research methods. The best result was obtained with the average of 88% of the recognition rate when it consisted of 15 hidden layers. The suggested algorithm was considered effective for the recognition of automatic diagnosis of heart sound recognition programs.
Sang Min LEE In Young KIM Seung Hong HONG
Recently many researches concerning heart sound analysis are being processed with development of digital signal processing and electronic components. But there are few researches about recognition of heart sound, especially full cardiac cycled heart sound. In this paper, three new recognition methods about full cardiac cycled heart sound were proposed. The first method recognizes the characteristics of heart sound by integrating important peaks and analyzing statistical variables in time domain. The second method builds a database by principal components analysis on training heart sound set in time domain. This database is used to recognize new input of heart sound. The third method builds the same sort of the database not in time domain but in time-frequency domain. We classify the heart sounds into seven classes such as normal (NO) class, pre-systolic murmur (PS) class, early systolic murmur (ES) class, late systolic murmur (LS) class, early diastolic murmur (ED) class, late diastolic murmur (LD) class and continuous murmur (CM) class. As a result, we could verify that the third method is better efficient to recognize the characteristics of heart sound than the others and also than any precedent research. The recognition rates of the third method are 100% for NO, 80% for PS and ES, 67% for LS, 93 for ED, 80% for LD and 30% for CM.
Takehiko ASHIYA Masao NAKAGAWA
In the future, it will be necessary that robot technology or environmental technology has an auditory function of recognizing sound expect for speech. In this letter, we propose a recognition system for the species of birds receiving birdcalls, based on network technology. We show the first step of a recognition system for the species of birds, as an application of a recognition system for environmental sound.