from #Audiology via ola Kala on Inoreader http://ift.tt/2dDBLDf
via IFTTT
OtoRhinoLaryngology by Sfakianakis G.Alexandros Sfakianakis G.Alexandros,Anapafseos 5 Agios Nikolaos 72100 Crete Greece,tel : 00302841026182,00306932607174
The purpose of this study was to determine the effects of age on the spectro-temporal integration of speech. The hypothesis was that the integration of speech fragments distributed over frequency, time, and ear of presentation is reduced in older listeners—even for those with good audiometric hearing. Younger, middle-aged, and older listeners (10 per group) with good audiometric hearing participated. They were each tested under seven conditions that encompassed combinations of spectral, temporal, and binaural integration. Sentences were filtered into two bands centered at 500 Hz and 2500 Hz, with criterion bandwidth tailored for each participant. In some conditions, the speech bands were individually square wave interrupted at a rate of 10 Hz. Configurations of uninterrupted, synchronously interrupted, and asynchronously interrupted frequency bands were constructed that constituted speech fragments distributed across frequency, time, and ear of presentation. The over-arching finding was that, for most configurations, performance was not differentially affected by listener age. Although speech intelligibility varied across condition, there was no evidence of performance deficits in older listeners in any condition. This study indicates that age, per se, does not necessarily undermine the ability to integrate fragments of speech dispersed across frequency and time.
The purpose of this study was to determine the effects of age on the spectro-temporal integration of speech. The hypothesis was that the integration of speech fragments distributed over frequency, time, and ear of presentation is reduced in older listeners—even for those with good audiometric hearing. Younger, middle-aged, and older listeners (10 per group) with good audiometric hearing participated. They were each tested under seven conditions that encompassed combinations of spectral, temporal, and binaural integration. Sentences were filtered into two bands centered at 500 Hz and 2500 Hz, with criterion bandwidth tailored for each participant. In some conditions, the speech bands were individually square wave interrupted at a rate of 10 Hz. Configurations of uninterrupted, synchronously interrupted, and asynchronously interrupted frequency bands were constructed that constituted speech fragments distributed across frequency, time, and ear of presentation. The over-arching finding was that, for most configurations, performance was not differentially affected by listener age. Although speech intelligibility varied across condition, there was no evidence of performance deficits in older listeners in any condition. This study indicates that age, per se, does not necessarily undermine the ability to integrate fragments of speech dispersed across frequency and time.
The purpose of this study was to determine the effects of age on the spectro-temporal integration of speech. The hypothesis was that the integration of speech fragments distributed over frequency, time, and ear of presentation is reduced in older listeners—even for those with good audiometric hearing. Younger, middle-aged, and older listeners (10 per group) with good audiometric hearing participated. They were each tested under seven conditions that encompassed combinations of spectral, temporal, and binaural integration. Sentences were filtered into two bands centered at 500 Hz and 2500 Hz, with criterion bandwidth tailored for each participant. In some conditions, the speech bands were individually square wave interrupted at a rate of 10 Hz. Configurations of uninterrupted, synchronously interrupted, and asynchronously interrupted frequency bands were constructed that constituted speech fragments distributed across frequency, time, and ear of presentation. The over-arching finding was that, for most configurations, performance was not differentially affected by listener age. Although speech intelligibility varied across condition, there was no evidence of performance deficits in older listeners in any condition. This study indicates that age, per se, does not necessarily undermine the ability to integrate fragments of speech dispersed across frequency and time.
by Islay Mactaggart, Hannah Kuper, G. V. S. Murthy, Joseph Oye, Sarah Polack
PurposeTo investigate the relationship between two distinct measures of disability: self-reported functional limitations and objectively-screened clinical impairments.
MethodsWe undertook an all age population-based survey of disability in two areas: North-West Cameroon (August/October 2013) and Telangana State, India (Feb/April 2014). Participants were selected for inclusion via two-stage cluster randomised sampling (probability proportionate to size cluster selection and compact segment sampling within clusters). Disability was defined as the presence of self-reported functional limitations across eight domains, or presence of moderate or greater clinical impairments. Clinical impairment screening comprised of visual acuity testing for vision impairment, pure tone audiometry for hearing impairment, musculoskeletal functioning assessment for musculoskeletal impairment, reported seizure history for epilepsy and reported symptoms of clinical depression (depression adults only). Information was collected using structured questionnaires, observations and examinations.
ResultsSelf-reported disability prevalence was 5.9% (95% CI 4.7–7.4) and 7.5% (5.9–9.4) in Cameroon and India respectively. The prevalence of moderate or greater clinical impairments in the same populations were 8.4% (7.5–9.4) in Cameroon and 10.5% (9.4–11.7) in India. Overall disability prevalence (self-report and/or screened positive to a moderate or greater clinical impairment) was 10.5% in Cameroon and 12.2% in India, with limited overlap between the sub-populations identified using the two types of tools. 33% of participants in Cameroon identified to have a disability, and 45% in India, both reported functional limitations and screened positive to objectively-screened impairments, whilst the remainder were identified via one or other tool only. A large proportion of people with moderate or severe clinical impairments did not self-report functional difficulties despite reporting participation restrictions.
ConclusionTools to assess reported functional limitation alone are insufficient to identify all persons with participation restrictions and moderate or severe clinical impairments. A self-reported functional limitation tool followed by clinical screening of all those who report any level of difficulty would identify 94% of people with disabilities in Cameroon and 95% in India, meeting the study criteria.
Athletes from the United States won 117 medals. Those from Ukraine, Great Britain, and China won even more. We are talking about medals at the 2016 Paralympics that followed the summer Olympics in Rio da Janeiro. While the summer Olympics get all their deserved attention, the Paralympics that follow hold a place of particular importance in the Olympic movement. This year’s Paralympics in Rio had a unique twist for us sound geeks.
Available traffic noise prediction models are usually based on regression analysis of experimental data, and this paper presents the application of soft computing techniques in traffic noise prediction. Two mathematical models are proposed and their predictions are compared to data collected by traffic noise monitoring in urban areas, as well as to predictions of commonly used traffic noise models. The results show that application of evolutionary algorithms and neural networks may improve process of development, as well as accuracy of traffic noise prediction.
This study investigates the effects of the dialect of the speaker on the spectral properties of stop bursts. Forty-five female speakers—20 Standard Modern Greek and 25 Cypriot Greek speakers—participated in this study. The spectral properties of stop bursts were calculated from the burst spectra and analyzed using spectral moments. The findings show that besides linguistic information, i.e., the place of articulation and the stress, the speech signals of bursts can encode social information, i.e., the dialects. A classification model using decision trees showed that skewness and standard deviation have a major contribution for the classification of bursts across dialects.
Related Articles |
Maturation of Spontaneous Firing Properties after Hearing Onset in Rat Auditory Nerve Fibers: Spontaneous Rates, Refractoriness, and Interfiber Correlations.
J Neurosci. 2016 Oct 12;36(41):10584-10597
Authors: Wu JS, Young ED, Glowatzki E
Abstract
Auditory nerve fibers (ANFs) exhibit a range of spontaneous firing rates (SRs) that are inversely correlated with threshold for sounds. To probe the underlying mechanisms and time course of SR differentiation during cochlear maturation, loose-patch extracellular recordings were made from ANF dendrites using acutely excised rat cochlear preparations of different ages after hearing onset. Diversification of SRs occurred mostly between the second and the third postnatal week. Statistical properties of ANF spike trains showed developmental changes that approach adult-like features in older preparations. Comparison with intracellularly recorded EPSCs revealed that most properties of ANF spike trains derive from the characteristics of presynaptic transmitter release. Pharmacological tests and waveform analysis showed that endogenous firing produces some fraction of ANF spikes, accounting for their unusual properties; the endogenous firing diminishes gradually during maturation. Paired recordings showed that ANFs contacting the same inner hair cell could have different SRs, with no correlation in their spike timing.
SIGNIFICANCE STATEMENT: The inner hair cell (IHC)/auditory nerve fiber (ANF) synapse is the first synapse of the auditory pathway. Remarkably, each IHC is the sole partner of 10-30 ANFs with a range of spontaneous firing rates (SRs). Low and high SR ANFs respond to sound differently, and both are important for encoding sound information across varying acoustical environments. Here we demonstrate SR diversification after hearing onset by afferent recordings in acutely excised rat cochlear preparations. We describe developmental changes in spike train statistics and endogenous firing in immature ANFs. Dual afferent recordings provide the first direct evidence that fibers with different SRs contact the same IHCs and do not show correlated spike timing at rest. These results lay the groundwork for understanding the differential sensitivity of ANFs to acoustic trauma.
PMID: 27733610 [PubMed - in process]
Related Articles |
Maturation of Spontaneous Firing Properties after Hearing Onset in Rat Auditory Nerve Fibers: Spontaneous Rates, Refractoriness, and Interfiber Correlations.
J Neurosci. 2016 Oct 12;36(41):10584-10597
Authors: Wu JS, Young ED, Glowatzki E
Abstract
Auditory nerve fibers (ANFs) exhibit a range of spontaneous firing rates (SRs) that are inversely correlated with threshold for sounds. To probe the underlying mechanisms and time course of SR differentiation during cochlear maturation, loose-patch extracellular recordings were made from ANF dendrites using acutely excised rat cochlear preparations of different ages after hearing onset. Diversification of SRs occurred mostly between the second and the third postnatal week. Statistical properties of ANF spike trains showed developmental changes that approach adult-like features in older preparations. Comparison with intracellularly recorded EPSCs revealed that most properties of ANF spike trains derive from the characteristics of presynaptic transmitter release. Pharmacological tests and waveform analysis showed that endogenous firing produces some fraction of ANF spikes, accounting for their unusual properties; the endogenous firing diminishes gradually during maturation. Paired recordings showed that ANFs contacting the same inner hair cell could have different SRs, with no correlation in their spike timing.
SIGNIFICANCE STATEMENT: The inner hair cell (IHC)/auditory nerve fiber (ANF) synapse is the first synapse of the auditory pathway. Remarkably, each IHC is the sole partner of 10-30 ANFs with a range of spontaneous firing rates (SRs). Low and high SR ANFs respond to sound differently, and both are important for encoding sound information across varying acoustical environments. Here we demonstrate SR diversification after hearing onset by afferent recordings in acutely excised rat cochlear preparations. We describe developmental changes in spike train statistics and endogenous firing in immature ANFs. Dual afferent recordings provide the first direct evidence that fibers with different SRs contact the same IHCs and do not show correlated spike timing at rest. These results lay the groundwork for understanding the differential sensitivity of ANFs to acoustic trauma.
PMID: 27733610 [PubMed - in process]
A novel founder MYO15A frameshift duplication is the major cause of genetic hearing loss in Oman.
J Hum Genet. 2016 Oct 13;:
Authors: Palombo F, Al-Wardy N, Ruscone GA, Oppo M, Kindi MN, Angius A, Al Lamki K, Girotto G, Giangregorio T, Benelli M, Magi A, Seri M, Gasparini P, Cucca F, Sazzini M, Al Khabori M, Pippucci T, Romeo G
Abstract
The increased risk for autosomal recessive disorders is one of the most well-known medical implications of consanguinity. In the Sultanate of Oman, a country characterized by one of the highest rates of consanguineous marriages worldwide, prevalence of genetic hearing loss (GHL) is estimated to be 6/10 000. Families of GHL patients have higher consanguinity rates than the general Omani population, indicating a major role for recessive forms. Mutations in GJB2, the most commonly mutated GHL gene, have been sporadically described. We collected 97 DNA samples of GHL probands, affected/unaffected siblings and parents from 26 Omani consanguineous families. Analyzing a first family by whole-exome sequencing, we identified a novel homozygous frameshift duplication (c.1171_1177dupGCCATCT) in MYO15A, the gene linked to the deafness locus DFNB3. This duplication was then found in a total of 8/26 (28%) families, within a 849 kb founder haplotype. Reconstruction of haplotype structure at MYO15A surrounding genomic regions indicated that the founder haplotype branched out in the past two to three centuries from a haplotype present worldwide. The MYO15A duplication emerges as the major cause of GHL in Oman. These findings have major implications for the design of GHL diagnosis and prevention policies in Oman.Journal of Human Genetics advance online publication, 13 October 2016; doi:10.1038/jhg.2016.120.
PMID: 27734841 [PubMed - as supplied by publisher]
![]() |
Related Articles |
Speech Recognition and Parent Ratings From Auditory Development Questionnaires in Children Who Are Hard of Hearing.
Ear Hear. 2015 Nov-Dec;36 Suppl 1:60S-75S
Authors: McCreery RW, Walker EA, Spratford M, Oleson J, Bentler R, Holte L, Roush P
Abstract
OBJECTIVES: Progress has been made in recent years in the provision of amplification and early intervention for children who are hard of hearing. However, children who use hearing aids (HAs) may have inconsistent access to their auditory environment due to limitations in speech audibility through their HAs or limited HA use. The effects of variability in children's auditory experience on parent-reported auditory skills questionnaires and on speech recognition in quiet and in noise were examined for a large group of children who were followed as part of the Outcomes of Children with Hearing Loss study.
DESIGN: Parent ratings on auditory development questionnaires and children's speech recognition were assessed for 306 children who are hard of hearing. Children ranged in age from 12 months to 9 years. Three questionnaires involving parent ratings of auditory skill development and behavior were used, including the LittlEARS Auditory Questionnaire, Parents Evaluation of Oral/Aural Performance in Children rating scale, and an adaptation of the Speech, Spatial, and Qualities of Hearing scale. Speech recognition in quiet was assessed using the Open- and Closed-Set Test, Early Speech Perception test, Lexical Neighborhood Test, and Phonetically Balanced Kindergarten word lists. Speech recognition in noise was assessed using the Computer-Assisted Speech Perception Assessment. Children who are hard of hearing were compared with peers with normal hearing matched for age, maternal educational level, and nonverbal intelligence. The effects of aided audibility, HA use, and language ability on parent responses to auditory development questionnaires and on children's speech recognition were also examined.
RESULTS: Children who are hard of hearing had poorer performance than peers with normal hearing on parent ratings of auditory skills and had poorer speech recognition. Significant individual variability among children who are hard of hearing was observed. Children with greater aided audibility through their HAs, more hours of HA use, and better language abilities generally had higher parent ratings of auditory skills and better speech-recognition abilities in quiet and in noise than peers with less audibility, more limited HA use, or poorer language abilities. In addition to the auditory and language factors that were predictive for speech recognition in quiet, phonological working memory was also a positive predictor for word recognition abilities in noise.
CONCLUSIONS: Children who are hard of hearing continue to experience delays in auditory skill development and speech-recognition abilities compared with peers with normal hearing. However, significant improvements in these domains have occurred in comparison to similar data reported before the adoption of universal newborn hearing screening and early intervention programs for children who are hard of hearing. Increasing the audibility of speech has a direct positive effect on auditory skill development and speech-recognition abilities and also may enhance these skills by improving language abilities in children who are hard of hearing. Greater number of hours of HA use also had a significant positive impact on parent ratings of auditory skills and children's speech recognition.
PMID: 26731160 [PubMed - indexed for MEDLINE]