Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

DATENSATZ AKTIONENEXPORT
  Decoding time for the identification of musical key

Farbood, M. M., Rowland, J., Marcus, G., Ghitza, O., & Poeppel, D. (2015). Decoding time for the identification of musical key. Attention, Perception & Psychophysics, 77(1), 28-35. doi:10.3758/s13414-014-0806-0.

Item is

Basisdaten

einblenden: ausblenden:
Genre: Zeitschriftenartikel

Externe Referenzen

einblenden:

Urheber

einblenden:
ausblenden:
 Urheber:
Farbood, M. M.1, Autor
Rowland, J.1, Autor
Marcus, Gary1, Autor
Ghitza, O., Autor
Poeppel, David1, 2, Autor           
Affiliations:
1New York University, New York, NY, USA , ou_persistent22              
2Department of Neuroscience, Max Planck Institute for Empirical Aesthetics, Max Planck Society, ou_2421697              

Inhalt

einblenden:
ausblenden:
Schlagwörter: Key finding Tonal induction Neuronal oscillations Music structure Brain rhythms Speech rate COMPRESSED SPEECH LANGUAGE INTELLIGIBILITY BRAIN INTEGRATION PERCEPTION DURATION SYNTAX Psychology Psychology, Experimental
 Zusammenfassung: This study examines the decoding times at which the brain processes structural information in music and compares them to timescales implicated in recent work on speech. Combining an experimental paradigm based on Ghitza and Greenberg (Phonetica, 66(1-2), 113-126, 2009) for speech with the approach of Farbood et al. (Journal of Experimental Psychology: Human Perception and Performance, 39(4), 911-918, 2013) for musical key-finding, listeners were asked to judge the key of short melodic sequences that were presented at a highly a compressed rate with varying durations of silence inserted in a periodic manner in the audio signal. The distorted audio signals comprised signal-silence alternations showing error rate curves that identify peak performance centered around an event rate of 5-7 Hz (143-200 ms interonset interval; 300-420 beats/min), where event rate is defined as the average rate of pitch change. The data support the hypothesis that the perceptual analysis of music entails the processes of parsing the signal into chunks of the appropriate temporal granularity and decoding the signal for recognition. The music-speech comparison points to similarities in how auditory processing builds on the specific temporal structure of the input, and how that structure interacts with the internal temporal dynamics of the neural mechanisms underpinning perception.

Details

einblenden:
ausblenden:
Sprache(n): eng - English
 Datum: 2014-12-102015-01
 Publikationsstatus: Erschienen
 Seiten: -
 Ort, Verlag, Ausgabe: -
 Inhaltsverzeichnis: -
 Art der Begutachtung: -
 Identifikatoren: Anderer: WOS:000347454200002
DOI: 10.3758/s13414-014-0806-0
 Art des Abschluß: -

Veranstaltung

einblenden:

Entscheidung

einblenden:

Projektinformation

einblenden:

Quelle 1

einblenden:
ausblenden:
Titel: Attention, Perception & Psychophysics
  Kurztitel : Atten Percept Psychophys
Genre der Quelle: Zeitschrift
 Urheber:
Affiliations:
Ort, Verlag, Ausgabe: Psychonomic Society
Seiten: - Band / Heft: 77 (1) Artikelnummer: - Start- / Endseite: 28 - 35 Identifikator: ISSN: 1943-3921
CoNE: https://pure.mpg.de/cone/journals/resource/1943-3921