Automatic disorder recognition in speech can be very helpful for the therapist while monitoring therapy progress of the patients with disordered speech. In this article we focus on prolongations. We analyze the signal using Continuous Wavelet Transform with 18 bark scales, we divide the result into vectors (using windowing) and then we pass such vectors into Kohonen network. Quite large search analysis was performed (5 variables were checked) during which, recognition above 90% was achieved. All the analysis was performed and the results were obtained using the authors' program - "WaveBlaster". It is very important that the recognition ratio above 90% was obtained by a fully automatic algorithm (without a teacher) from the continuous speech. The presented problem is part of our research aimed at creating an automatic prolongation recognition system.
A novel method for designing near-perfect reconstruction oversampled nonuniform cosine-modulated filter banks is proposed, which combines frequency warping and subband merging, and thus offers more flexibility than known techniques. On the one hand, desirable frequency partitionings can be better approximated. On the other hand, at the price of only a small loss in partitioning accuracy, both warping strength and number of channels before merging can be adjusted so as to minimize the computational complexity of a system. In particular, the coefficient of the function behind warping can be constrained to be a negative integer power of two, so that multiplications related to allpass filtering can be replaced with more efficient binary shifts. The main idea is accompanied by some contributions to the theory of warped filter banks. Namely, group delay equalization is thoroughly investigated, and it is shown how to avoid significant aliasing by channel oversampling. Our research revolves around filter banks for perceptual processing of sound, which are required to approximate the psychoacoustic scales well and need not guarantee perfect reconstruction.
Automatic disorders recognition in speech can be very helpful for therapist while monitoring therapy progress of patients with disordered speech. This article is focused on sound repetitions. The signal is analyzed using Continuous Wavelet Transform with 16 bark scales, the result is divided into vectors and passed into Kohonen network. Finally, the Kohonen winning neuron result is put on the 3-layer perceptron. The recognition ratio was increased by about 20% by adding a modification into the Kohonen network training process as well as into CWT computation algorithm. All the analysis was performed and the results were obtained using the authors' program ”WaveBlaster“, The problem presented in this article is a part of our research work aimed at creating an automatic disordered speech recognition system.
JavaScript jest wyłączony w Twojej przeglądarce internetowej. Włącz go, a następnie odśwież stronę, aby móc w pełni z niej korzystać.