Stellen Sie sich vor, Sie befinden sich im dichten Straßenverkehr, inmitten von Fußgängern, Radfahrern und Autos, die sich alle in unterschiedliche Richtungen bewegen. In dieser und vielen anderen Situationen ist es überlebenswichtig, genau zu wissen wo und wann Ereignisse in unserer Umgebung stattfinden. Um möglichst schnell und korrekt auf externe Reize zu reagieren, erzeugt unser Gehirn dabei ständig Vorhersagen über zukünftige Ereignisse. Zum Beispiel, wo ein heranfahrendes Auto sich zu dem Zeitpunkt befinden wird, wenn wir die Straße überqueren wollen. Nicht nur für uns Menschen sind diese Vorhersagen zentral. Auch andere Primaten könnten ähnliche Mechanismen verwenden, etwa wenn sie sich durch dichtes Dschungelgebiet bewegen. Inwiefern die Evolution diese Mechanismen bei Menschen im Vergleich zu anderen Spezies geformt hat, ist bis heute unklar.
Des Weiteren sind unsere Sinnesinformationen oft mehrdeutig, sodass unser Gehirn mehrere parallele Interpretationen und Vorhersagen erzeugt und sich letztlich auf eine festlegen muss. Gegenwärtig stammt der Großteil unseres Wissens über diese Wahrnehmungsprozesse aus Studien zum Sehsinn. Vergleichsweise wenig ist darüber für unseren Hörsinn bekannt, welcher aber gleichermaßen zentral für unser Überleben und Sozialverhalten ist.
Das Zukunftskolleg Dynamates möchte diese zentralen Wissenslücken zur Hörwahrnehmung schließen indem es die Vorhersagemechanismen nahe verwandter Spezies in realistischen aber hochkontrollierbaren virtuellen akustischen Umgebungen testen und mit Computermodellen abbilden wird. Zusätzlich wird Dynamates mittels hochauflösender Elektroenzephalographie (EEG) bei Menschen die neuronalen Grundlagen der zugrunde liegenden Prozesse untersuchen. Das Projekt basiert damit auf einer interdisziplinären Zusammenarbeit zwischen Expertinnen und Experten aus dem Bereich der Computermodellierung (Robert Baumgartner), der Neurowissenschaft (Ulrich Pomper), und der Kognitionsbiologie (Michelle Spierings).
Dynamates wird somit den ersten systematischen Vergleich von dynamischen Vorhersage- und Entscheidungsprozessen des Hörsinns zwischen Menschen und nicht-menschlichen Primaten durchführen. Ein besseres Verständnis der neuronalen Grundlagen dieser Prozesse kann Anwendung bei der Behandlung von Personen mit Störungen im Wahrnehmungs- und Entscheidungsverhalten (z.B. bei Autismus oder Schizophrenie) finden. Die erstellten mathematischen Modelle lassen sich in Zukunft auch in anderen Spezies oder bei komplexeren Entscheidungsprozessen (z.B. in sozialen Interaktionen) testen, und können direkte Anwendung in der Entwicklung von künstlicher Intelligenz und virtuellen Realitäten finden.
In folgender Online-Lecture erklärt Robert Baumgartner weitere Hintergründe zu dieser Forschung: ÖAW Science Bites: Gefahr - wie wir sie hören.
Unsere neuen Teammitglieder starten in Kürze: Roberto Barumerli, Sophie Hanke, and David Meijer.
Selective hearing refers to the ability of the human auditory system to selectively attend to a desired speaker while ignoring undesired, concurrent speakers. This is often referred to as the cocktail-party problem. In normal hearing, selective hearing is remarkably powerful. However, in so-called electric hearing, i.e., hearing with cochlear implants (CIs), selective hearing is severely degraded, close to not present at all. CIs are commonly used for treatment of severe-to-profound hearing loss or deafness because they provide good speech understanding in quiet. The reasons for the deficits in selective hearing are mainly twofold. First, they arise from structural limitations of current CI electrode designs which severely limit the spectral resolution. Second, they arise from a lack of salient timing cues, most importantly interaural time difference (ITD) and temporal pitch. The second limitation is assumed to be partly “software”-sided and conquerable with perception-driven signal processing. Yet, success achieved so far is at best moderate.
A recently proposed approach to provide precise ITD and temporal-pitch cues in addition to speech understanding is to insert extra pulses with short inter-pulse intervals (so-called SIPI pulses) into periodic high-rate pulse trains. Results gathered so far in our previous project ITD PsyPhy in single-electrode configurations are encouraging in that both ITD and temporal-pitch sensitivity improved when SIPI pulses were inserted at the signals’ temporal-envelope peaks. Building on those results, this project aims to answer the most urgent research questions towards determining whether the SIPI approach improves selective hearing in CI listeners: Does the SIPI benefit translate into multi-electrode configurations? Does the multi-electrode SIPI approach harm speech understanding? Does the multi-electrode SIPI approach improve speech-in-speech understanding?
Psychophysical experiments with CI listeners are planned to examine the research questions. To ensure high temporal precision and stimulus control, clinical CI signal processors will be bypassed by using a laboratory stimulation system directly connecting the CIs with a laboratory computer. The results are expected to shed light on parts of both electric and acoustic hearing that are still not fully understood to date, such as the role and the potential of temporal cues in selective hearing.
References from our Lab:
Duration: May 2020 - April 2022
Funding: DOC Fellowship Program of the Austrian Academy of Sciences (A-25606)
Supervisors: Bernhard Laback and Ulrich Ansorge (University of Vienna)
See also:
Normal-hearing (NH) listeners use two binaural cues, the interaural time difference (ITD) and the interaural level difference (ILD), for sound localization in the horizontal plane. They apply frequency-dependent weights when combining them to determine the perceived azimuth of a sound source. Cochlear implant (CI) listeners, however, rely almost entirely on ILDs. This is partly due to the properties of current envelope-based CI-systems, which do not explicitly encode carrier ITDs. However, even if they are artificially conveyed via a research system, CI listeners perform worse on average than NH listeners. Since current CI-systems do not reliably convey ITD information, CI listeners might learn to ignore ITDs and focus on ILDs instead. A recent study in our lab provided first evidence that such reweighting of binaural cues is possible in NH listeners.
This project aims to further investigate the phenomenon: First, we will test whether a changed ITD/ILD weighting will generalize to different frequency regions. Second, the effect of ITD/ILD reweighting on spatial release from speech-on-speech masking will be investigated, as listeners benefit particularly from ITDs in such tasks. And third, we will test, whether CI listeners can also be trained to weight ITDs more strongly and whether that translates to an increase in ITD sensitivity. Additionally, we will explore and evaluate different training methods to induce ITD/ILD reweighting.
The results are expected to shed further light on the plasticity of the binaural auditory system in acoustic and electric hearing.
Start: October 2018
Duration: 3 years
Funding: uni:docs fellowship program for doctoral candidates of the University of Vienna
The auditory system constantly monitors the environment to protect us from harmful events such as collisions with approaching objects. Auditory looming bias is an astoundingly fast perceptual bias favoring approaching compared to receding auditory motion and was demonstrated behaviorally even in infants of four months in age. The role of learning in developing this perceptual bias and its underlying mechanisms are yet to be investigated. Supervised learning and statistical learning are the two distinct mechanisms enabling neural plasticity. In the auditory system, statistical learning refers to the implicit ability to extract and represent regularities, such as frequently occurring sound patterns or frequent acoustic transitions, with or without attention while supervised learning refers to the ability to attentively encode auditory events based on explicit feedback. It is currently unclear how these two mechanisms are involved in learning auditory spatial cues at different stages of life. While newborns already possess basic skills of spatial hearing, adults are still able to adapt to changing circumstances such as modifications of spectral-shape cues. Spectral-shape cues are naturally induced when the complex geometry especially of the human pinna shapes the spectrum of an incoming sound depending on its source location. Auditory stimuli lacking familiarized spectral-shape cues are often perceived to originate from inside the head instead of perceiving them as naturally external sound sources. Changes in the salience or familiarity of spectral-shape cues can thus be used to elicit auditory looming bias. The importance of spectral-shape cues for both auditory looming bias and auditory plasticity makes it ideal for studying them together.
Born2Hear will combine auditory psychophysics and neurophysiological measures in order to 1) identify auditory cognitive subsystems underlying auditory looming bias, 2) investigate principle cortical mechanisms for statistical and supervised learning of auditory spatial cues, and 3) reveal cognitive and neural mechanisms of auditory plasticity across the human lifespan. These general research questions will be addressed within three studies. Study 1 will investigate the differences in the bottom-up processing of different spatial cues and the top-down attention effects on auditory looming bias by analyzing functional interactions between brain regions in young adults and then test in newborns whether these functional interactions are innate. Study 2 will investigate the cognitive and neural mechanisms of supervised learning of spectral-shape cues in young and older adults based on an individualized perceptual training on sound source localization. Study 3 will focus on the cognitive and neural mechanisms of statistical learning of spectral-shape cues in infants as well as young and older adults.
Key publication: 2017): Asymmetries in behavioral and neural responses to spectral cues demonstrate the generality of auditory looming bias, in: Proceedings of the National Academy of Sciences of the USA 114, 9743-9748.
(Project investigator (PI): Robert Baumgartner
Project partner / Co-PI: Brigitta Tóth, Institute of Cognitive Neuroscience and Psychology, Research Centre for Natural Sciences, Hungarian Academy of Sciences, Budapest, Hungary
Collaboration partners:
Duration: April 2020 - March 2024
Supported by the Austrian Science Fund (FWF, I 4294-B) and NKFIH.
Normal 0 21 false false false DE-AT X-NONE X-NONE
/* Style Definitions */ table.MsoNormalTable {mso-style-name:"Normale Tabelle"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:12.0pt; font-family:"Liberation Serif","serif"; mso-bidi-font-family:Mangal; mso-fareast-language:ZH-CN; mso-bidi-language:HI;}
Current cochlear implants (CIs) are very successful in restoring speech understanding in individuals with profound or complete hearing loss by electrically stimulating the auditory nerve. However, the ability of CI users to localize sound sources and to understand speech in complex listening situations, e.g. with interfering speakers, is dramatically reduced as compared to normal (acoustically) hearing listeners. From acoustic hearing studies it is known that interaural time difference (ITD) cues are essential for sound localization and speech understanding in noise. Users of current bilateral CI systems are, however, rather limited in their ability to perceive salient ITDs cues. One particular problem is that their ITD sensitivity is especially low when stimulating at relatively high pulses rates which are required for proper encoding of speech signals.
In this project we combine psychophysical studies in human bilaterally implanted listeners and physiological studies in bilaterally implanted animals to find ways in order to improve ITD sensitivity in electric hearing. We build on the previous finding that ITD sensitivity can be enhanced by introducing temporal jitter (Laback and Majdak, 2008) or short inter-pulse intervals (Hancock et al., 2012) in high-rate pulse sequences. Physiological experiments, performed at the Eaton-Peabody Laboratories Neural Coding Group (Massachusetts Eye and Ear Infirmary, Harvard Medical School, PI: Bertrand Delgutte), are combined with matched psychoacoustic experiments, performed at the EAP group of ARI (PI: Bernhard Laback). The main project milestones are the following:
· Aim 1: Effects of auditory deprivation and electric stimulation through CI on neural ITD sensitivity. In physiological experiments it is studied if chronic CI stimulation can reverse the effect of neonatal deafness on neural ITD sensitivity.
· Aim 2: Improving the delivery of ITD information with high-rate strategies for CI processors.
◦ A. Improving ITD sensitivity at high pulse rates by introducing short inter-pulse intervals
◦ B. Using short inter-pulse intervals to enhance ITD sensitivity with “pseudo-syllable” stimuli.
· External: Eaton-Peabody Laboratories Neural Coding Group des Massachusetts Eye and Ear Infirmary an der Harvard Medical School (PI: Bertrand Delgutte)
· Internal: Mathematik und Signalverarbeitung in der Akustik
· This project is funded by the National Institute of Health (NIH). http://grantome.com/grant/NIH/R01-DC005775-11A1
· It is planned to run from 2014 to 2019.
· Article in DER STANDARD: http://derstandard.at/2000006635467/OeAW-und-Harvard-Medical-School-forschenCochleaimplantaten
· Article in DIE PRESSE: http://diepresse.com/home/science/3893396/Eine-Prothese-die-in-der-Horschnecke-sitzt
· OEAW website: http://www.oeaw.ac.at/oesterreichische-akademie-der-wissenschaften/news/article/transatlantische-hoerhilfe/
Binaural hearing is extremely important in everyday life, most notably for sound localization and for understanding speech embedded in competing sound sources (e.g., other speech sources). While bilateral implantation has been shown to provide cochlear implant (CIs) listeners with some basic left/right localization ability, the performance with current CI systems is clearly reduced compared to normal hearing. Moreover, the binaural advantage in speech understanding in noise has been shown to be mediated mainly by the better-ear effect, while there is only very little binaural unmasking.
There exists now a body of literature on binaural sensitivity of CI listeners stimulated at a single interaural electrode pair. However, the CI listener’s sensitivity to binaural cues under more realistic conditions, i.e., with stimulation at multiple electrodes, has not been systematically addressed in depth so far.
This project attempts to fill this gap. In particular, given the high perceptual importance of ITDs, this project focuses on the systematic investigation of the sensitivity to ITD under various conditions of multi-electrode stimulation, including interference from neighboring channels, integration of ITD information across channels, and the perceptually tolerable room for degradations of binaural timing information.
Involved people:
Start: January 2013
Duration: 3 years
Funding: MED-EL
Localization of sound sources is an important task of the human auditory system and much research effort has been put into the development of audio devices for virtual acoustics, i.e. the reproduction of spatial sounds via headphones. Even though the process of sound localization is not completely understood yet, it is possible to simulate spatial sounds via headphones by using head-related transfer functions (HRTFs). HRTFs describe the filtering of the incoming sound due to head, torso and particularly the pinna and thus they strongly depend on the particular details in the listener's geometry. In general, for realistic spatial-sound reproduction via headphones, the individual HRTFs must be measured. As of 2012, the available HRTF acquisition methods were acoustic measurements: a technically-complex process, involving placing microphones into the listener's ears, and lasting for tens of minutes.
In LocaPhoto, we were working on an easily accessible method to acquire and evaluate listener-specific HRTFs. The idea was to numerically calculate HRTFs based on a geometrical representation of the listener (3-D mesh) obtained from 2-D photos by means of photogrammetric reconstruction.
As a result, we have developed a software package for numerical HRTF calculations, a method for geometry acquisition, and models able to evaluate HRTFs in terms of broadband ITDs and sagittal-plane sound localization performance.
http://www.kfs.oeaw.ac.at/LocaPhoto
The spatially oriented format for acoustics (SOFA) is dedicated to store all kinds of acoustic informations related to a specified geometrical setup. The main task is to describe simple HRTF measurements, but SOFA also aims to provide the functionality to store measurements of something fancy like BRIRs with a 64-channel mic-array in a multi-source excitation situation or directivity measurement of a loudspeaker. The format is intended to be easily extendable, highly portable, and actually the greatest common divider of all publicly available HRTF databases at the moment of writing.
SOFA defines the structure of data and meta data and stores them in a numerical container. The data description will be a hierarchical description when coming from free-field HRTFs (simple setup) and going to more complex setups like mic-array measurements in reverberant spaces, excited by a loudspeaker array (complex setup). We will use global geometry description (related to the room), and local geometry description (related to the listener/source) without limiting the number of acoustic transmitters and receivers. Room descriptions will be available by linking a CAD file within SOFA. Networking support will be provided as well allowing to remotely access HRTFs and BRIRs from client computers.
SOFA is being developed by many contributors worldwide. The development is coordinated at ARI by Piotr Majdak.
While it is often assumed that our auditory system is phase-deaf, there is a body of literature showing that listeners are very sensitive to phase differences between spectral components of a sound. Particularly, for spectral components falling into the same perceptual filter, the so-called auditory filter, a change in relative phase across components causes a change in the temporal pattern at the output of the filter. The phase response of the auditory filter is thus important for any auditory tasks that rely on within-channel temporal envelope information, most notably temporal pitch or interaural time differences.
Within-channel phase sensitivity has been used to derive a psychophysical measure of the phase response of auditory filters (Kohlrausch and Sanders, 1995). The basic idea of the widely used masking paradigm is that a harmonic complex whose phase curvature roughly mirrors the phase response of the auditory filter spectrally centered on the complex causes a maximally modulated (peaked) internal representation and, thus, elicits minimal masking of a pure tone target at the same center frequency. Therefore, systematic variation of the phase curvature of the harmonic complex (the masker) allows to estimate the auditory filter’s phase response: the masker phase curvature causing minimal masking reflects the mirrored phase response of the auditory filter.
Besides the obvious importance of detecting the target in the temporal dips of the masker, particularly of the target is short compared to the modulation period of the masker (Kohlrausch and Sanders, 1995), there are several indications that fast compression in the cochlea is important to obtain the masker-phase effect (e.g., Carlyon and Datta, 1997; Oxenham and Dau, 2004). One indication is that listeners with sensorineural hearing impairment (HI), characterized by reduced or absent cochlear compression due to loss of outer hair cells, show only a very weak masker-phase effect, making it difficult to estimate the cochlear phase response.
In the BiPhase project we propose a new paradigm for measuring the cochlear phase response that does not rely on cochlear compression and thus should be applicable in HI listeners. It relies on the idea that the amount of modulation (peakedness) in the internal representation of a harmonic complex, as given by its phase curvature, determines the listener’s sensitivity to envelope interaural time difference (ITD) imposed on the stimulus. Assuming that listener’s sensitivity to envelope ITD does not rely on compression, systematic variation of the stimulus phase curvature should allow to estimate the cochlear phase response both in normal-hearing (NH) and HI listeners. The main goals of BiPhase are the following:
This project is funded by the Austrian Science Fund (FWF, Project # P24183-N24, awarded to Bernhard Laback). It run from 2013 to 2017
The dependency of perceived loudness from electrical current in Cochlear Implant (CI) stimulation has been investigated in several existing studies. This investigation has two main goals:
Loudness functions are measured at single electrodes (or interaural electrode pairs) using the method of categorical loudness scaling. The efficiency of this method for hearing impaired listeners has been demonstrated in previous studies (Brand and Hohmann, JASA 112, p.1597-1604). Both an adaptive method and the method of constant stimuli are used. Binaural functions are measured subsequently to monaural function, including monaural measurements as control conditions.
The results indicate the suitability and efficiency of the adaptive categorical loudness scaling method as a tool for the fast determination of the loudness function. This can be applied to the clinical fitting of implant processors as well as for pre-measurements in psychoaoustic CI studies. The measurement results also provide new insights into monaural and binaural loudness perception of CI listeners.
internal