Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 122
Filter
1.
Front Neurosci ; 18: 1373232, 2024.
Article in English | MEDLINE | ID: mdl-38952924

ABSTRACT

Timbre is a central aspect of music that allows listeners to identify musical sounds and conveys musical emotion, but also allows for the recognition of actions and is an important structuring property of music. The former functions are known to be implemented in a ventral auditory stream in processing musical timbre. While the latter functions are commonly attributed to areas in a dorsal auditory processing stream in other musical domains, its involvement in musical timbre processing is so far unknown. To investigate if musical timbre processing involves both dorsal and ventral auditory pathways, we carried out an activation likelihood estimation (ALE) meta-analysis of 18 experiments from 17 published neuroimaging studies on musical timbre perception. We identified consistent activations in Brodmann areas (BA) 41, 42, and 22 in the bilateral transverse temporal gyri, the posterior superior temporal gyri and planum temporale, in BA 40 of the bilateral inferior parietal lobe, in BA 13 in the bilateral posterior Insula, and in BA 13 and 22 in the right anterior insula and superior temporal gyrus. The vast majority of the identified regions are associated with the dorsal and ventral auditory processing streams. We therefore propose to frame the processing of musical timbre in a dual-stream model. Moreover, the regions activated in processing timbre show similarities to the brain regions involved in processing several other fundamental aspects of music, indicating possible shared neural bases of musical timbre and other musical domains.

2.
Front Psychol ; 15: 1363329, 2024.
Article in English | MEDLINE | ID: mdl-38933586

ABSTRACT

Aim: Timbre in piano performance plays a critical role in enhancing musical expression. However, timbre control in current piano performance education relies mostly on descriptive characterization, which involves large variations of interpretation. The current study aimed to mitigate the limitations by identifying quantitative indices with adequate precision to characterize piano timbre. Methods: A total of 24 sounds of G6 were recorded from 3 grand pianos, by 2 performers, and with 4 repetitions. The sounds were processed and analyzed with audio software for the frequencies and volumes of harmonic series in the spectrum curves. Ten quantitative timbre indices were calculated. Precision validation with statistical gage R&R analysis was conducted to gage the repeatability (between repetitions) and reproducibility (between performers) of the indices. The resultant percentage study variation (%SV) of an index must be ≤10% to be considered acceptable for characterizing piano timbre with enough precision. Results: Out of the 10 indices, 4 indices had acceptable precision in characterizing piano timbre with %SV ≤10%, including the square sum of relative volume (4.40%), the frequency-weighted arithmetic mean of relative volume (4.29%), the sum of relative volume (3.11%), and the frequency-weighted sum of relative volume (2.09%). The novel indices identified in the current research will provide valuable tools to advance the measurement and communication of timbre and advance music performance education.

3.
J Voice ; 2024 May 15.
Article in English | MEDLINE | ID: mdl-38755075

ABSTRACT

Timbre is a central quality of singing, yet remains a complex notion poorly understood in psychoacoustic studies. Previous studies note how no single acoustic variable or combinations of variables consistently predict timbre dimensions. Timbre varies on a continuum from darkest to lightest. These extremes are associated with laryngeal and vocal tract adjustments related to smaller and larger vocal tract area and variations in vocal fold vibratory characteristics. Perceptually, timbre assessment is influenced by spectral characteristics and formant frequency adjustments, though these dimensions are not independently perceived. Perceptual studies repeatedly demonstrate difficulties in correlating variations in timbre stimuli to specific measures. A recent study demonstrated how acoustic predictive salience of voice category and voice weight across pitches contribute to timbre assessments and concludes that timbre may be related to as-of-yet unknown factor(s). The purpose of this study was to test four different models for assessing timbre; one model focused on specific anatomy, one on listener intuition, one utilizing auditory anchors, and one using expert raters in a deconstructed timbre model with five specific dimensions. METHODS: Four independent panels were conducted with separate cohorts of professional singing teachers. Forty-one assessors took part in the anatomically focused panel, 54 in the intuition-based panel, 30 in the anchored panel, and 12 in the expert listener panel. Stimuli taken from live performances of well-known singers were used for all panels, representing all genders, genres, and styles across a large pitch range. All stimuli are available as Supplementary Materials. Fleiss' kappa values, descriptive statistics, and significance tests are reported for all panel assessments. RESULTS: Panels 1 through 4 varied in overall accuracy and agreement. The intuition-based model showed overall 45% average accuracy (SD ± 4%), k = 0.289 (<0.001) compared to overall 71% average accuracy (SD ± 3%), k = 0.368 (<0.001) of the anatomical focused panel. The auditory-anchored model showed overall 75% average accuracy (SD ± 8%), k = 0.54 (<0.001) compared with overall 83% average accuracy and agreement of k = 0.63 (<0.001) for panel 4. Results revealed that the highest accuracy and reliability were achieved in a deconstructed timbre model and that providing anchoring improved reliability but with no further increase in accuracy. CONCLUSION: Deconstructing timbre into specific parameters improved auditory perceptual accuracy and overall agreement. Assessing timbre along with other perceptual dimensions improves accuracy and reliability. Panel assessors' expert level of listening skills remain an important factor in obtaining reliable and accurate assessments of auditory stimuli for timbre dimensions. Anchoring improved reliability but with no further increase in accuracy. The study suggests that timbre assessment can be improved by approaching the percept through a prism of five specific dimensions each related to specific physiology and auditory-perceptual subcategories. Further tests are needed with framework-naïve listeners, nonmusically educated listeners, artificial intelligence comparisons, and synthetic stimuli to further test the reliability.

4.
Psychol Music ; 52(3): 305-321, 2024 May.
Article in English | MEDLINE | ID: mdl-38708378

ABSTRACT

Music that evokes strong emotional responses is often experienced as autobiographically salient. Through emotional experience, the musical features of songs could also contribute to their subjective autobiographical saliency. Songs which have been popular during adolescence or young adulthood (ages 10-30) are more likely to evoke stronger memories, a phenomenon known as a reminiscence bump. In the present study, we sought to determine how song-specific age, emotional responsiveness to music, musical features, and subjective memory functioning contribute to the subjective autobiographical saliency of music in older adults. In a music listening study, 112 participants rated excerpts of popular songs from the 1950s to the 1980s for autobiographical saliency. Additionally, they filled out questionnaires about emotional responsiveness to music and subjective memory functioning. The song excerpts' musical features were extracted computationally using MIRtoolbox. Results showed that autobiographical saliency was best predicted by song-specific age and emotional responsiveness to music and musical features. Newer songs that were more similar in rhythm to older songs were also rated higher in autobiographical saliency. Overall, this study contributes to autobiographical memory research by uncovering a set of factors affecting the subjective autobiographical saliency of music.

5.
Cognition ; 243: 105672, 2024 Feb.
Article in English | MEDLINE | ID: mdl-38086279

ABSTRACT

Pleasure in music has been linked to predictive coding of melodic and rhythmic patterns, subserved by connectivity between regions in the brain's auditory and reward networks. Specific musical anhedonics derive little pleasure from music and have altered auditory-reward connectivity, but no difficulties with music perception abilities and no generalized physical anhedonia. Recent research suggests that specific musical anhedonics experience pleasure in nonmusical sounds, suggesting that the implicated brain pathways may be specific to music reward. However, this work used sounds with clear real-world sources (e.g., babies laughing, crowds cheering), so positive hedonic responses could be based on the referents of these sounds rather than the sounds themselves. We presented specific musical anhedonics and matched controls with isolated short pleasing and displeasing synthesized sounds of varying timbres with no clear real-world referents. While the two groups found displeasing sounds equally displeasing, the musical anhedonics gave substantially lower pleasure ratings to the pleasing sounds, indicating that their sonic anhedonia is not limited to musical rhythms and melodies. Furthermore, across a large sample of participants, mean pleasure ratings for pleasing synthesized sounds predicted significant and similar variance in six dimensions of musical reward considered to be relatively independent, suggesting that pleasure in sonic timbres play a role in eliciting reward-related responses to music. We replicate the earlier findings of preserved pleasure ratings for semantically referential sounds in musical anhedonics and find that pleasure ratings of semantic referents, when presented without sounds, correlated with ratings for the sounds themselves. This association was stronger in musical anhedonics than in controls, suggesting the use of semantic knowledge as a compensatory mechanism for affective sound processing. Our results indicate that specific musical anhedonia is not entirely specific to melodic and rhythmic processing, and suggest that timbre merits further research as a source of pleasure in music.


Subject(s)
Music , Humans , Music/psychology , Anhedonia/physiology , Pleasure/physiology , Brain/physiology , Reward , Auditory Perception/physiology
6.
Br J Psychol ; 115(2): 206-225, 2024 May.
Article in English | MEDLINE | ID: mdl-37851369

ABSTRACT

Musicians outperform non-musicians in vocal emotion perception, likely because of increased sensitivity to acoustic cues, such as fundamental frequency (F0) and timbre. Yet, how musicians make use of these acoustic cues to perceive emotions, and how they might differ from non-musicians, is unclear. To address these points, we created vocal stimuli that conveyed happiness, fear, pleasure or sadness, either in all acoustic cues, or selectively in either F0 or timbre only. We then compared vocal emotion perception performance between professional/semi-professional musicians (N = 39) and non-musicians (N = 38), all socialized in Western music culture. Compared to non-musicians, musicians classified vocal emotions more accurately. This advantage was seen in the full and F0-modulated conditions, but was absent in the timbre-modulated condition indicating that musicians excel at perceiving the melody (F0), but not the timbre of vocal emotions. Further, F0 seemed more important than timbre for the recognition of all emotional categories. Additional exploratory analyses revealed a link between time-varying F0 perception in music and voices that was independent of musical training. Together, these findings suggest that musicians are particularly tuned to the melody of vocal emotions, presumably due to a natural predisposition to exploit melodic patterns.


Subject(s)
Music , Voice , Humans , Acoustic Stimulation , Emotions , Fear , Recognition, Psychology , Music/psychology , Auditory Perception
7.
Laryngoscope ; 134(3): 1381-1387, 2024 Mar.
Article in English | MEDLINE | ID: mdl-37665102

ABSTRACT

OBJECTIVE: Music is a highly complex acoustic stimulus in both spectral and temporal contents. Accurate representation and delivery of high-fidelity information are essential for music perception. However, it is unclear how well bone-anchored hearing implants (BAHIs) transmit music. The study objective is to establish music perception performance baselines for BAHI users and normal hearing (NH) listeners and compare outcomes between the cohorts. METHODS: A case-controlled, cross-sectional study was conducted among 18 BAHI users and 11 NH controls. Music perception was assessed via performance on seven major musical element tasks: pitch discrimination, melodic contour identification, rhythmic clocking, basic tempo discrimination, timbre identification, polyphonic pitch detection, and harmonic chord discrimination. RESULTS: BAHI users performed comparably well on all music perception tasks with their device compared with the unilateral condition with their better-hearing ear. BAHI performance was not statistically significantly different from NH listeners' performance. BAHI users performed just as well, if not better than NH listeners when using their control contralateral ear; there was no significant difference between the two groups except for the rhythmic timing (BAHI non-implanted ear 69% [95% CI: 62%-75%], NH 56% [95% CI: 49%-63%], p = 0.02), and basic tempo tasks (BAHI non-implanted ear 80% [95% CI: 65%-95%]; NH 75% [95% CI: 68%-82%, p = 0.03]). CONCLUSIONS: This study represents the first comprehensive study of basic music perception performance in BAHI users. Our results demonstrate that BAHI users perform as well with their implanted ear as with their contralateral better-hearing ear and NH controls in the major elements of music perception. LEVEL OF EVIDENCE: 3 Laryngoscope, 134:1381-1387, 2024.


Subject(s)
Cochlear Implantation , Cochlear Implants , Music , Humans , Auditory Perception , Cross-Sectional Studies , Hearing , Pitch Perception
8.
Brain Sci ; 13(11)2023 Nov 07.
Article in English | MEDLINE | ID: mdl-38002523

ABSTRACT

Musicians outperform non-musicians in vocal emotion recognition, but the underlying mechanisms are still debated. Behavioral measures highlight the importance of auditory sensitivity towards emotional voice cues. However, it remains unclear whether and how this group difference is reflected at the brain level. Here, we compared event-related potentials (ERPs) to acoustically manipulated voices between musicians (n = 39) and non-musicians (n = 39). We used parameter-specific voice morphing to create and present vocal stimuli that conveyed happiness, fear, pleasure, or sadness, either in all acoustic cues or selectively in either pitch contour (F0) or timbre. Although the fronto-central P200 (150-250 ms) and N400 (300-500 ms) components were modulated by pitch and timbre, differences between musicians and non-musicians appeared only for a centro-parietal late positive potential (500-1000 ms). Thus, this study does not support an early auditory specialization in musicians but suggests instead that musicality affects the manner in which listeners use acoustic voice cues during later, controlled aspects of emotion evaluation.

9.
Diagnostics (Basel) ; 13(18)2023 Sep 06.
Article in English | MEDLINE | ID: mdl-37761237

ABSTRACT

Vocal arts medicine provides care and prevention strategies for professional voice disorders in performing artists. The issue of correct "Fach" determination depending on the presence of a lyric or dramatic voice structure is of crucial importance for opera singers, as chronic overuse often leads to vocal fold damage. To avoid phonomicrosurgery or prevent a premature career end, our aim is to offer singers an improved, objective fach counseling using digital sound analyses and machine learning procedures. For this purpose, a large database of 2004 sound samples from professional opera singers was compiled. Building on this dataset, we employed a classic ensemble learning method, namely the Random Forest algorithm, to construct an efficient fach classifier. This model was trained to learn from features embedded within the sound samples, subsequently enabling voice classification as either lyric or dramatic. As a result, the developed system can decide with an accuracy of about 80% in most examined voice types whether a sound sample has a lyric or dramatic character. To advance diagnostic tools and health in vocal arts medicine and singing voice pedagogy, further machine learning methods will be applied to find the best and most efficient classification method based on artificial intelligence approaches.

10.
J Audiol Otol ; 27(4): 193-204, 2023 Oct.
Article in English | MEDLINE | ID: mdl-37533349

ABSTRACT

BACKGROUND AND OBJECTIVES: The present study aimed to explore meaningful change in the musical performance of cochlear implant (CI) users according to age after receiving music training compared to the other CI users who did not receive any music training and normal hearing (NH) listeners. Materials and. METHODS: Article search was conducted from five journal databases. In accordance with PICOS criteria, 29 articles and 15 articles were included for the CI and NH groups, respectively. A total of 44 articles consisting of 407 CI users and 486 NH participants were then analyzed. Using the data gathered from pre- and post-training, meta-analysis was applied by the random-effects model, and then piecewise meta-regression was employed. RESULTS: CI users with music training showed better performance than those without the training across all ages. The curve change in musical performance when training showed that the CI users was similar to the general pattern of NH performance, indicating that the CI users gained benefits from music training and developed perceptual ability, even though it was not as high as the level of NH counterparts. CONCLUSIONS: This finding supports the benefits of active engagement with music for both CI users and NH listeners, while noting the important role of music as one of the primary aural rehabilitation approaches.

11.
J Neurophysiol ; 130(2): 291-302, 2023 08 01.
Article in English | MEDLINE | ID: mdl-37377190

ABSTRACT

Traditionally, pitch variation in a sound stream has been integral to music identity. We attempt to expand music's definition, by demonstrating that the neural code for musicality is independent of pitch encoding. That is, pitchless sound streams can still induce music-like perception and a neurophysiological hierarchy similar to pitched melodies. Previous work reported that neural processing of sounds with no-pitch, fixed-pitch, and irregular-pitch (melodic) patterns, exhibits a right-lateralized hierarchical shift, with pitchless sounds favorably processed in Heschl's gyrus (HG), ascending laterally to nonprimary auditory areas for fixed-pitch and even more laterally for melodic patterns. The objective of this EEG study was to assess whether sound encoding maintains a similar hierarchical profile when musical perception is driven by timbre irregularities in the absence of pitch changes. Individuals listened to repetitions of three musical and three nonmusical sound-streams. The nonmusical streams were comprised of seven 200-ms segments of white, pink, or brown noise, separated by silent gaps. Musical streams were created similarly, but with all three noise types combined in a unique order within each stream to induce timbre variations and music-like perception. Subjects classified the sound streams as musical or nonmusical. Musical processing exhibited right dominant α power enhancement, followed by a lateralized increase in θ phase-locking and spectral power. The θ phase-locking was stronger in musicians than in nonmusicians. The lateralization of activity suggests higher-level auditory processing. Our findings validate the existence of a hierarchical shift, traditionally observed with pitched-melodic perception, underscoring that musicality can be achieved with timbre irregularities alone.NEW & NOTEWORTHY EEG induced by streams of pitchless noise segments varying in timbre were classified as music-like and exhibited a right-lateralized hierarchy in processing similar to pitched melodic processing. This study provides evidence that the neural-code of musicality is independent of pitch encoding. The results have implications for understanding music processing in individuals with degraded pitch perception, such as in cochlear-implant listeners, as well as the role of nonpitched sounds in the induction of music-like perceptual states.


Subject(s)
Cochlear Implants , Music , Humans , Pitch Perception/physiology , Auditory Perception/physiology , Sound , Acoustic Stimulation
12.
J Voice ; 2023 Jun 09.
Article in English | MEDLINE | ID: mdl-37302910

ABSTRACT

OBJECTIVE/HYPOTHESES: The terms "soprano" and "mezzo-soprano" are frequently used by vocal pedagogues to describe a main category of singing timbre categorization, while the terms "lyric" and "dramatic" are often used to describe sub-categories of "soprano" and "mezzo-soprano". A handful of studies have reported on the perceptual dissimilarity of main voice categories, but few, if any, have focused on within voice category perceptual distinctions such as dramatic and lyric vocal timbre. Using stimuli collected from cisgender female singers of varying voice categories and voice weights across the pitches C4, G4, and F5, this study sought (1) to visualize an experienced listener's perception of vocal timbre dissimilarity within and between voice categories using the statistical technique of multidimensional scaling (MDS), (2) to identify salient acoustic predictors of voice category and voice weight, and (3) to determine any dependencies on pitch for the perception of vocal timbre. METHOD/DESIGN: For the pitches C4, G4, and F5, experienced listeners (N=18) rated the dissimilarity of pairs of sung vowels produced by classically trained singers classified as follows: six mezzo-sopranos (three lighter and three heavier) and six sopranos (three lighter and three heavier). The resulting dissimilarity data were analyzed using MDS. Backward linear regression was used to see if one or more of the following variables predicted MDS dimensions: spectral centroid from 0 to 5 kHz, spectral centroid from 0 to 2 kHz, spectral centroid from 2 to 5 kHz, frequency vibrato rate, and frequency vibrato extent. Listeners also completed a categorization task where they rated each individual stimulus on two dimensions: voice category and voice weight. RESULTS: Visual analysis of the MDS solutions appears to show that both voice category and voice weight emerged as dimensions at pitches C4 and G4. Discriminant analysis, on the other hand, statistically confirmed both these dimensions at G4, but only voice weight at C4. At pitch F5, only voice weight emerged as a dimension, both visually and statistically. Acoustic predictors of MDS dimensions were highly variable across pitches. At the pitch C4, no MDS dimension was predicted by the acoustic variables. At pitch G4, the dimension associated with voice weight was predicted by spectral centroid from 0 to 2 kHz. At pitch F5, the dimension associated with voice weight was predicted by spectral centroid from 2 to 5 kHz and frequency vibrato rate. In the categorization task, voice category and voice weight were highly correlated at the pitches C4, G4, and when all pitches were presented together, but weakly correlated at the pitch F5. CONCLUSION: While voice category and sub-category distinctions are commonly used by singing voice professionals to describe the overall timbre of voices, these distinctions may not be able to consistently predict the perceptual difference between any given pair of vocal stimuli, particularly across pitch. Nonetheless, these dimensions do emerge in some fashion when listeners are presented with paired vocal stimuli. On the other hand, when asked to rate stimuli according to the specific labels of mezzo-soprano/soprano and dramatic/lyric, experienced listeners have a very difficult time disentangling voice category from voice weight when presented with a single-note stimulus or even a 3-note stimulus consisting of the pitches C3, G4, and F5.

13.
Front Psychol ; 14: 1046672, 2023.
Article in English | MEDLINE | ID: mdl-37205083

ABSTRACT

Introduction: A singer's or speaker's Fach (voice type) should be appraised based on acoustic cues characterizing their voice. Instead, in practice, it is often influenced by the individual's physical appearance. This is especially distressful for transgender people who may be excluded from formal singing because of perceived mismatch between their voice and appearance. To eventually break down these visual biases, we need a better understanding of the conditions under which they occur. Specifically, we hypothesized that trans listeners (not actors) would be better able to resist such biases, relative to cis listeners, precisely because they would be more aware of appearance-voice dissociations. Methods: In an online study, 85 cisgender and 81 transgender participants were presented with 18 different actors singing or speaking short sentences. These actors covered six voice categories from high/bright (traditionally feminine) to low/dark (traditionally masculine) voices: namely soprano, mezzo-soprano (referred to henceforth as mezzo), contralto (referred to henceforth as alto), tenor, baritone, and bass. Every participant provided voice type ratings for (1) Audio-only (A) stimuli to get an unbiased estimate of a given actor's voice type, (2) Video-only (V) stimuli to get an estimate of the strength of the bias itself, and (3) combined Audio-Visual (AV) stimuli to see how much visual cues would affect the evaluation of the audio. Results: Results demonstrated that visual biases are not subtle and hold across the entire scale, shifting voice appraisal by about a third of the distance between adjacent voice types (for example, a third of the bass-to-baritone distance). This shift was 30% smaller for trans than for cis listeners, confirming our main hypothesis. This pattern was largely similar whether actors sung or spoke, though singing overall led to more feminine/high/bright ratings. Conclusion: This study is one of the first demonstrations that transgender listeners are in fact better judges of a singer's or speaker's voice type because they are better able to separate the actors' voice from their appearance, a finding that opens exciting avenues to fight more generally against implicit (or sometimes explicit) biases in voice appraisal.

14.
Cogn Emot ; 37(4): 731-747, 2023.
Article in English | MEDLINE | ID: mdl-37104118

ABSTRACT

Research into voice perception benefits from manipulation software to gain experimental control over acoustic expression of social signals such as vocal emotions. Today, parameter-specific voice morphing allows a precise control of the emotional quality expressed by single vocal parameters, such as fundamental frequency (F0) and timbre. However, potential side effects, in particular reduced naturalness, could limit ecological validity of speech stimuli. To address this for the domain of emotion perception, we collected ratings of perceived naturalness and emotionality on voice morphs expressing different emotions either through F0 or Timbre only. In two experiments, we compared two different morphing approaches, using either neutral voices or emotional averages as emotionally non-informative reference stimuli. As expected, parameter-specific voice morphing reduced perceived naturalness. However, perceived naturalness of F0 and Timbre morphs were comparable with averaged emotions as reference, potentially making this approach more suitable for future research. Crucially, there was no relationship between ratings of emotionality and naturalness, suggesting that the perception of emotion was not substantially affected by a reduction of voice naturalness. We hold that while these findings advocate parameter-specific voice morphing as a suitable tool for research on vocal emotion perception, great care should be taken in producing ecologically valid stimuli.


Subject(s)
Speech Perception , Voice , Humans , Emotions
15.
Trends Hear ; 27: 23312165221138390, 2023.
Article in English | MEDLINE | ID: mdl-36789758

ABSTRACT

The study tests the hypothesis that vibrotactile stimulation can affect timbre perception. A multidimensional scaling experiment was conducted. Twenty listeners with normal hearing and nine cochlear implant users were asked to judge the dissimilarity of a set of synthetic sounds that varied in attack time and amplitude modulation depth. The listeners were simultaneously presented with vibrotactile stimuli, which varied also in attack time and amplitude modulation depth. The results showed that alterations to the temporal waveform of the tactile stimuli affected the listeners' dissimilarity judgments of the audio. A three-dimensional analysis revealed evidence of crossmodal processing where the audio and tactile equivalents combined accounted for their dissimilarity judgments. For the normal-hearing listeners, 86% of the first dimension was explained by audio impulsiveness and 14% by tactile impulsiveness; 75% of the second dimension was explained by the audio roughness or fast amplitude modulation, while its tactile counterpart explained 25%. Interestingly, the third dimension revealed a combination of 43% of audio impulsiveness and 57% of tactile amplitude modulation. For the CI listeners, the first dimension was mostly accounted for by the tactile roughness and the second by the audio impulsiveness. This experiment shows that the perception of timbre can be affected by tactile input and could lead to the developing of new audio-tactile devices for people with hearing impairment.


Subject(s)
Cochlear Implants , Music , Humans , Acoustic Stimulation/methods , Auditory Perception/physiology , Timbre Perception , Hearing
16.
Atten Percept Psychophys ; 85(2): 525-542, 2023 Feb.
Article in English | MEDLINE | ID: mdl-36690914

ABSTRACT

Absolute pitch (AP) is the rare ability to name any musical note without the use of a reference note. Given that genuine AP representations are based on the identification of isolated notes by their tone chroma, they are considered to be invariant to (1) surrounding tonal context, (2) changes in instrumental timbre, and (3) changes in octave register. However, there is considerable variability in the literature in terms of how AP is trained and tested along these dimensions, making recent claims about AP learning difficult to assess. Here, we examined the effect of tonal context on participant success with a single-note identification training paradigm, including how learning generalized to an untested instrument and octave. We found that participants were able to rapidly learn to distinguish C from other notes, with and without feedback and regardless of the tonal context in which C was presented. Participants were also able to partly generalize this skill to an untrained instrument. However, participants displayed the weakest generalization in recognizing C in a higher octave. The results indicate that participants were likely attending to pitch height in addition to pitch chroma - a conjecture that was supported by analyzing the pattern of response errors. These findings highlight the complex nature of note representation in AP, which requires note identification across contexts, going beyond the simple storage of a note fundamental. The importance of standardizing testing that spans both timbre and octave in assessing AP and further implications on past literature and future work are discussed.


Subject(s)
Music , Humans , Acoustic Stimulation/methods , Learning/physiology , Generalization, Psychological , Pitch Perception/physiology , Pitch Discrimination/physiology
17.
Cognition ; 232: 105327, 2023 03.
Article in English | MEDLINE | ID: mdl-36495710

ABSTRACT

Information in speech and music is often conveyed through changes in fundamental frequency (f0), perceived by humans as "relative pitch". Relative pitch judgments are complicated by two facts. First, sounds can simultaneously vary in timbre due to filtering imposed by a vocal tract or instrument body. Second, relative pitch can be extracted in two ways: by measuring changes in constituent frequency components from one sound to another, or by estimating the f0 of each sound and comparing the estimates. We examined the effects of timbral differences on relative pitch judgments, and whether any invariance to timbre depends on whether judgments are based on constituent frequencies or their f0. Listeners performed up/down and interval discrimination tasks with pairs of spoken vowels, instrument notes, or synthetic tones, synthesized to be either harmonic or inharmonic. Inharmonic sounds lack a well-defined f0, such that relative pitch must be extracted from changes in individual frequencies. Pitch judgments were less accurate when vowels/instruments were different compared to when they were the same, and were biased by the associated timbre differences. However, this bias was similar for harmonic and inharmonic sounds, and was observed even in conditions where judgments of harmonic sounds were based on f0 representations. Relative pitch judgments are thus not invariant to timbre, even when timbral variation is naturalistic, and when such judgments are based on representations of f0.


Subject(s)
Music , Pitch Perception , Humans , Pitch Discrimination , Acoustic Stimulation
18.
Front Psychol ; 13: 908543, 2022.
Article in English | MEDLINE | ID: mdl-36248517

ABSTRACT

This study exploratively conducted two investigations of timbre influenced by harmonic loudness. Investigation 1 examined piano timbre preference between non-processed melodies and two kinds of adjustments created via an equalizer-a basic tool used for sound engineering and for producing audio materials. Using the paired comparison method, 98 respondents were surveyed. The results show that in the bass range of the piano sound, the preferred audio material was that produced via an equalizer by reducing the loudness of the 7th and 9th harmonics by 15 dB, while enhancing the loudness of the 8th harmonic by 15 dB. Investigation 2 examined three degrees of the processed melodies-adjusting the harmonics by 15, 10, and 5 dB. The results show that the 15 dB change was the best, while the 10 dB change was the worst. These results provide a new approach to improving the sound timbre for sound engineering and artificial intelligence music production.

19.
Neonatology ; 119(6): 760-768, 2022.
Article in English | MEDLINE | ID: mdl-36116434

ABSTRACT

INTRODUCTION: Crying newborns signal a need or discomfort as part of the innate communication system. Exposure to pain is related to infants' unfavorable neurodevelopmental outcomes. There is a tremendous need for more objective methods to assess neonatal pain. An audio analysis of acoustic utterances could provide specific information on the patient's pain level. METHODS: We analyzed 67 videos of 33 term-born newborns recorded during a planned capillary blood sample, including the stimuli, non-noxious thermal stimulus, short noxious stimulus, and prolonged unpleasant stimulus, between December 2020 and March 2021. Two expert raters evaluated the infants' pain responses using the Neonatal Facial Coding System (NFCS). The mean values of 123 timbre features of the recorded audio data were analyzed by using specific toolboxes and libraries from the following programming environments: MIRtoolbox (MATLAB), MiningSuite (MATLAB), Essentia (Python), AudioCommons timbral models (Python), and Librosa (Python). RESULTS: The NFCS values were significantly higher during the short noxious stimulus (p < 0.001) and prolonged unpleasant stimulus (p < 0.001) than during the non-noxious thermal stimulus, whereas NFCS values during the short noxious stimulus and prolonged unpleasant stimulus were similar (p = 0.79). Brightness, roughness, percussive energy, and attack times were identified as the features having the highest impact on the NFCS. CONCLUSION: This hypothesis-generating study identified several salient acoustic features highly associated with pain responses in term newborns. Our analysis is an encouraging starting point for the targeted analysis of pain-specific acoustic features of neonatal cries and vocalizations from the perspective of real-time acoustic processing.


Subject(s)
Acoustics , Pain , Infant, Newborn , Humans , Pain/diagnosis
20.
Polymers (Basel) ; 14(14)2022 Jul 10.
Article in English | MEDLINE | ID: mdl-35890589

ABSTRACT

This article presents a proposal of thermal modification of Norway spruce and sycamore maple for special wood products, mainly for musical instruments. Selected physical and acoustical characteristics (PACHs), including the density (ρ), dynamic modulus of elasticity along the wood grain (EL), specific modulus (Esp), speed of sound along the wood grain (cL), resonant frequency (fr) and acoustic constant (A), logarithmic decrement (ϑ), loss coefficient (η), acoustic conversion efficiency (ACE), sound quality factor (Q), and the timbre of sound, were evaluated. These two wood species were chosen regarding their use in the production or repair of musical instruments. For the thermal modification, a similar process to the ThermoWood process was chosen. Thermal modification was performed at the temperatures 135 °C, 160 °C and 185 °C. The resonant dynamic method was used to obtain the PACHs. Fast Fourier transform (FFT) was used to analyze the sound produced. The changes in the observed wood properties depended on the treatment temperature. Based on our results of all properties, the different temperature modified wood could find uses in the making of musical instruments or where the specific values of these wood characteristics are required. The mild thermal modification resulted in a decrease in mass, density, and increased speed of sound and dynamic modulus of elasticity at all temperatures of modification. The thermally modified wood showed higher sound radiation and lower loss coefficients than unmodified wood. The modification also influenced the timbre of sound of both wood species.

SELECTION OF CITATIONS
SEARCH DETAIL
...