The current dataset provides all the stimuli (folder ../01-Stimuli/), raw data (folder ../02-Raw-data/) and post-processed data (../03-Post-proc-data/) used in a prosody reverse correlation study with the title "prosodic cues to word boundaries in a segmentation task using reverse correlation" by the same authors. The listening experiment was implemented using one-interval trials with target words of the structure l'aX (option 1) and la'X (option 2). The experiment was designed and implemented using the fastACI toolbox under the name 'segmentation'. A between-subject design was used with a total of 47 participants, who evaluated one of five conditions, LAMI (N=16), LAPEL (N=18), LACROCH (N=5), LALARM (N=5), and LAMI_SHIFTED (N=3). More details are given in the related publication (to be submitted to JASA-EL in May 2023).
Natural speech is highly complex and variable. Particularly, spoken language, in contrast to written language, has no clear word boundaries. Adult listeners can exploit different types of information to segment the continuous stream such as acoustic and semantic information. However, the weight of these cues, when co-occurring, remains to be determined. Behavioural tasks are not conclusive on this point as they focus participants’ attention on certain sources of information, thus biasing the results. Here, we looked at the processing of homophonic utterances such as l’amie vs la mie (both /lami/) which include fine acoustic differences and for which the meaning changes depending on segmentation. To examine the perceptual resolution of such ambiguities when semantic information is available, we measured the online processing of sentences containing such sequences in an ERP experiment involving no active task. In a congruent context, semantic information matched the acoustic signal of the word amie, while, in the incongruent condition, the semantic information carried by the sentence and the acoustic signal were leading to different lexical candidates. No clear neural markers for the use of acoustic cues were found. Our results suggest a preponderant weight of semantic information over acoustic information during natural spoken sentence processing.
Speech perception involves segmenting a continuous stream of speech into its word components. This can be challenging in the case of homophonous utterances only differing in non-contrastive subphonemic features. Yet, the speech perception system seems able to discriminate subphonemic deviation in homophonous utterances, since it has been shown to elicit a mismatch response (MMN). Here, we focused on the oscillatory correlates, namely phase resetting and power, of non-contrastive subphonemic deviation processing in language. An oddball task that considered natural intraspeaker variability was used. Subphonemic deviance elicited intertrial phase coherence (ITC) differences in the theta band at Fz during the time window of the MMN. No differences in power were found. This suggests that the processing of subphonemic deviation in speech signals, reflected by the MMN, might rely on mechanisms of phase resetting. ITC might facilitate the synchronous firing of functional networks involved in the processing of subphonemic deviance
We designed two experiments that tested the listeners' perceptual capacities during online segmentation of homophonic word boundaries while processing sentential information. In French, listeners often use variations in fine acoustic indices to detect word beginnings. We measured event-related potentials (ERPs) evoked by phonemically identical sequences, such as l'affiche ("the poster") and la fiche ("the sheet"), both [lafiʃ], which were contained in either congruent or incongruent sentences. Results showed that although listeners can detect acoustic variations in homophonic sequences, these may not be salient enough when contextual information is also present. Shifting attention from sentence meaning (Task 1) to lexical information (Task 2), enhanced the listeners' perception of fine-grained acoustic details. Thus, topdown processes are likely to modulate speech perception and segmentation.
In the present preregistered study, we evaluated the possibility of a shared cognitive mechanism during verbal and non-verbal tasks and therefore the implication of domain-general cognitive control during language comprehension. We hypothesized that a behavioral cost will be observed during a dual-task including both verbal and non-verbal difficult processing. Specifically, to test this claim, we designed a dual-task paradigm involving: an auditory language comprehension task (sentence comprehension) and a non-verbal Flanker task (including congruent and incongruent trials). We manipulated sentence ambiguity and evaluated if the ambiguity effect modified behavioral performances in the non-verbal Flanker task. Under the assumption that ambiguous sentences induce a more difficult process than unambiguous sentences, we expected non-verbal flanker task performances to be impaired only when a simultaneous difficult language processing is performed. This would be specifically reflected by a performance cost during incongruent Flanker items only during ambiguous sentence presentation. Conversely, we observed a facilitatory effect for the incongruent Flanker items during ambiguous sentence suggesting better non-verbal inhibitory performances when an ambiguous sentence was simultaneously processed. Exploratory data analysis suggests that this effect is not only related to a more difficult language processing but also to the previous ( n-1 ) Flanker item. Indeed, results showed that incongruent n-1 Flanker items led to a facilitation of the incongruent synchronized Flanker items only when ambiguous sentences were conjointly presented. This result, even if it needs to be corroborated in future studies, suggests that the recruitment of executive control mechanisms facilitates subsequent executive control implication during difficult language processing. The present study suggests a common executive control mechanism during difficult verbal and non-verbal tasks.
Interspeech 2020 - 21st Annual Conference of the International Speech Communication Association, Oct 2020, Shanghai (Virtual Conf), China. pp.1600-1604, ⟨10.21437/Interspeech.2020-2683⟩
annee_publi
2020
resume
Whistled speech is a form of modified speech where some frequencies of vowels and consonants are augmented and transposed to whistling, modifying the timbre and the construction of each phoneme. These transformations cause only some elements of the signal to be intelligible for naive listeners, which, according to previous studies, includes vowel recognition. Here, we analyze naive listeners’ capacities for whistled consonant categorization for four consonants: /p/, /k/, /t/ and /s/ by presenting the findings of two behavioral experiments. Though both experiments measure whistled consonant categorization, we used modified frequencies — lowered with a phase vocoder — of the whistled stimuli in the second experiment to better identify the relative nature of pitch cues employed in this process. Results show that participants obtained approximately 50% of correct responses (when chance is at 25%). These findings show specific consonant preferences for “s” and “t” over “k” and “p”, specifically when stimuli is unmodified. Previous research on whistled consonants systems has often opposed “s” and “t” to “k” and “p”, due to their strong pitch modulations. The preference for these two consonants underlines the importance of these cues in phoneme processing.
Interspeech 2020 - 21st Annual Conference of the International Speech Communication Association, Oct 2020, Shanghai (Virtual Conf), China. pp.1605-1609, ⟨10.21437/Interspeech.2020-2697⟩
annee_publi
2020
resume
In this paper, we analyzed whistled vowel categorization by native French listeners. Whistled speech, a natural, yet modified register of speech, is used here as a tool to investigate perceptual processes in languages. We focused on four whistled vowels: /i, e, a, o/. After a detailed description of the vowels, we built and ran a behavioral experiment in which we asked native French speakers to categorize whistled vowel stimuli in which we introduced intra- and inter- production variations. In addition, half of the participants performed the experiment in person (at the laboratory) while the other half participated online, allowing us to evaluate the impact of the testing set up. Our results confirm that the categorization rate of whistled vowels is above chance. They reveal significant differences in performance for different vowels and suggest an influence of certain acoustic parameters from the whistlers’ vowel range on categorization. Moreover, no effect or interaction was found for testing location and circumstances in our data set. This study confirms that whistled stimuli are a useful tool for studying how listeners process modified speech and which parameters impact sound categorization.