Loading AI tools
From Wikipedia, the free encyclopedia
Auditory phonetics is the branch of phonetics concerned with the hearing of speech sounds and with speech perception. It thus entails the study of the relationships between speech stimuli and a listener's responses to such stimuli as mediated by mechanisms of the peripheral and central auditory systems, including certain areas of the brain. It is said to compose one of the three main branches of phonetics along with acoustic and articulatory phonetics,[1][2] though with overlapping methods and questions.[3]
There is no direct connection between auditory sensations and the physical properties of sound that give rise to them. While the physical (acoustic) properties are objectively measurable, auditory sensations are subjective and can only be studied by asking listeners to report on their perceptions.[4] The table below shows some correspondences between physical properties and auditory sensations.
Physical property | Auditory perception |
---|---|
amplitude or intensity | loudness |
fundamental frequency | pitch |
spectral structure | sound quality |
duration | length |
Auditory phonetics is concerned with both segmental (chiefly vowels and consonants) and prosodic (such as stress, tone, rhythm and intonation) aspects of speech. While it is possible to study the auditory perception of these phenomena without context, in continuous speech all these variables are processed in parallel with significant variability and complex interactions between them.[5][6] For example, it has been observed that vowels, which are usually described as different from each other in the frequencies of their formants, also have intrinsic values of fundamental frequency (and presumably therefore of pitch) that are different according to the height of the vowel. Thus open vowels typically have lower fundamental frequency than close vowels in a given context,[7] and vowel recognition is likely to interact with the perception of prosody.
If there is a distinction to be made between auditory phonetics and speech perception, it is that the former is more closely associated with traditional non-instrumental approaches to phonology and other aspects of linguistics, while the latter is closer to experimental, laboratory-based study. Consequently, the term auditory phonetics is often used to refer to the study of speech without the use of instrumental analysis: the researcher may make use of technology such as recording equipment, or even a simple pen and paper (as used by William Labov in his study of the pronunciation of English in New York department stores),[8] but will not use laboratory techniques such as spectrography or speech synthesis, or methods such as EEG and fMRI that allow phoneticians to directly study the brain's response to sound. Most research in sociolinguistics and dialectology has been based on auditory analysis of data and almost all pronunciation dictionaries are based on impressionistic, auditory analysis of how words are pronounced. It is possible to claim an advantage for auditory analysis over instrumental: Kenneth L. Pike stated "Auditory analysis is essential to phonetic study since the ear can register all those features of sound waves, and only those features, which are above the threshold of audibility ... whereas analysis by instruments must always be checked against auditory reaction".[9] Herbert Pilch attempted to define auditory phonetics in such a way as to avoid any reference to acoustic parameters.[10] In the auditory analysis of phonetic data such as recordings of speech, it is clearly an advantage to have been trained in analytical listening. Practical phonetic training has since the 19th century been seen an essential foundation for phonetic analysis and for the teaching of pronunciation; it is still a significant part of modern phonetics. The best-known type of auditory training has been in the system of cardinal vowels; there is disagreement about the relative importance of auditory and articulatory factors underlying the system, but the importance of auditory training for those who are to use it is indisputable.[11]
Training in the auditory analysis of prosodic factors such as pitch and rhythm is also important. Not all research on prosody has been based on auditory techniques: some pioneering work on prosodic features using laboratory instruments was carried out in the 20th century (e.g. Elizabeth Uldall's work using synthesized intonation contours,[12] Dennis Fry's work on stress perception[13] or Daniel Jones's early work on analyzing pitch contours by means of manually operating the pickup arm of a gramophone to listen repeatedly to individual syllables, checking where necessary against a tuning fork),.[14] However, the great majority of work on prosody has been based on auditory analysis until the recent arrival of approaches explicitly based on computer analysis of the acoustic signal, such as ToBI, INTSINT or the IPO system.[15]
Seamless Wikipedia browsing. On steroids.
Every time you click a link to Wikipedia, Wiktionary or Wikiquote in your browser's search results, it will show the modern Wikiwand interface.
Wikiwand extension is a five stars, simple, with minimum permission required to keep your browsing private, safe and transparent.