Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 4 de 4
Filter
Add more filters










Database
Language
Publication year range
1.
Adv Exp Med Biol ; 787: 81-7; discussion 87-8, 2013.
Article in English | MEDLINE | ID: mdl-23716212

ABSTRACT

Lyon (J Acoust Soc Am 130:3893-3904, 2011) has described how a cascade of simple asymmetric resonators (CAR) can be used to simulate the filtering of the basilar membrane and how the gain of the resonators can be manipulated by a feedback network to simulate the fast-acting compression (FAC) characteristic of cochlear processing. When the compression is applied to complex tones, each pair of primary components produces both quadratic and cubic distortion tones (DTs), and the cascade architecture of the CAR-FAC system propagates them down to their appropriate place along the basilar membrane, where they combine additively with each other and any primary components at that frequency. This suggests that CAR-FAC systems might be used to study the role of compressive distortion in the perception of complex sounds and that behavioural measurements of cochlear distortion data might be useful when tuning the parameters of CAR-FAC systems.


Subject(s)
Basilar Membrane/physiology , Cochlea/physiopathology , Models, Biological , Perceptual Distortion/physiology , Acoustic Stimulation/methods , Humans , Nonlinear Dynamics , Pitch Perception/physiology
2.
Neural Comput ; 22(9): 2390-416, 2010 Sep 01.
Article in English | MEDLINE | ID: mdl-20569181

ABSTRACT

To create systems that understand the sounds that humans are exposed to in everyday life, we need to represent sounds with features that can discriminate among many different sound classes. Here, we use a sound-ranking framework to quantitatively evaluate such representations in a large-scale task. We have adapted a machine-vision method, the passive-aggressive model for image retrieval (PAMIR), which efficiently learns a linear mapping from a very large sparse feature space to a large query-term space. Using this approach, we compare different auditory front ends and different ways of extracting sparse features from high-dimensional auditory images. We tested auditory models that use an adaptive pole-zero filter cascade (PZFC) auditory filter bank and sparse-code feature extraction from stabilized auditory images with multiple vector quantizers. In addition to auditory image models, we compare a family of more conventional mel-frequency cepstral coefficient (MFCC) front ends. The experimental results show a significant advantage for the auditory models over vector-quantized MFCCs. When thousands of sound files with a query vocabulary of thousands of words were ranked, the best precision at top-1 was 73% and the average precision was 35%, reflecting a 18% improvement over the best competing MFCC front end.


Subject(s)
Auditory Perception/physiology , Models, Neurological , Humans , Sound
3.
J Acoust Soc Am ; 125(4): 2374-86, 2009 Apr.
Article in English | MEDLINE | ID: mdl-19354411

ABSTRACT

This paper investigates the theoretical basis for estimating vocal-tract length (VTL) from the formant frequencies of vowel sounds. A statistical inference model was developed to characterize the relationship between vowel type and VTL, on the one hand, and formant frequency and vocal cavity size, on the other. The model was applied to two well known developmental studies of formant frequency. The results show that VTL is the major source of variability after vowel type and that the contribution due to other factors like developmental changes in oral-pharyngeal ratio is small relative to the residual measurement noise. The results suggest that speakers adjust the shape of the vocal tract as they grow to maintain a specific pattern of formant frequencies for individual vowels. This formant-pattern hypothesis motivates development of a statistical-inference model for estimating VTL from formant-frequency data. The technique is illustrated using a third developmental study of formant frequencies. The VTLs of the speakers are estimated and used to provide a more accurate description of the complicated relationship between VTL and glottal pulse rate as children mature into adults.


Subject(s)
Larynx/anatomy & histology , Larynx/growth & development , Models, Biological , Mouth/anatomy & histology , Mouth/growth & development , Phonetics , Adolescent , Adult , Algorithms , Child , Child Development , Child Language , Child, Preschool , Female , Humans , Male , Speech , Young Adult
4.
J Acoust Soc Am ; 122(6): 3628-39, 2007 Dec.
Article in English | MEDLINE | ID: mdl-18247770

ABSTRACT

A recent study [Smith and Patterson, J. Acoust. Soc. Am. 118, 3177-3186 (2005)] demonstrated that both the glottal-pulse rate (GPR) and the vocal-tract length (VTL) of vowel sounds have a large effect on the perceived sex and age (or size) of a speaker. The vowels for all of the "different" speakers in that study were synthesized from recordings of the sustained vowels of one, adult male speaker. This paper presents a follow-up study in which a range of vowels were synthesized from recordings of four different speakers--an adult man, an adult woman, a young boy, and a young girl--to determine whether the sex and age of the original speaker would have an effect upon listeners' judgments of whether a vowel was spoken by a man, woman, boy, or girl, after they were equated for GPR and VTL. The sustained vowels of the four speakers were scaled to produce the same combinations of GPR and VTL, which covered the entire range normally encountered in every day life. The results show that listeners readily distinguish children from adults based on their sustained vowels but that they struggle to distinguish the sex of the speaker.


Subject(s)
Aging/physiology , Glottis/physiology , Phonetics , Speech Acoustics , Speech Perception , Vocal Cords/physiology , Adult , Age Factors , Body Height , Child , Female , Glottis/anatomy & histology , Humans , Male , Middle Aged , Organ Size , Probability , Sex Factors , Vocal Cords/anatomy & histology
SELECTION OF CITATIONS
SEARCH DETAIL
...