Refine
Year of publication
Document Type
- Article (17)
- Doctoral Thesis (8)
- Postprint (5)
- Monograph/Edited Volume (2)
- Part of a Book (1)
- Conference Proceeding (1)
Keywords
- prosody (34) (remove)
Institute
- Department Linguistik (22)
- Humanwissenschaftliche Fakultät (3)
- Institut für Anglistik und Amerikanistik (2)
- Department Psychologie (1)
- Extern (1)
- Institut für Germanistik (1)
- Institut für Philosophie (1)
- Language Acquisition (1)
- Philosophische Fakultät (1)
- Sonderforschungsbereich 632 - Informationsstruktur (1)
An exploration of rhythmic grouping of speech sequences by french- and german-learning infants
(2016)
Rhythm in music and speech can be characterized by a constellation of several acoustic cues. Individually, these cues have different effects on rhythmic perception: sequences of sounds alternating in duration are perceived as short-long pairs (weak-strong/iambicpattern), whereas sequences of sounds alternating in intensity or pitch are perceived as loud-soft, or high-low pairs (strong-weak/trochaic pattern). This perceptual bias-called the lambic-Trochaic Law (ITL) has been claimed to be an universal property of the auditory system applying in both the music and the language domains. Recent studies have shown that language experience can modulate the effects of the ITL on rhythmic perception of both speech and non-speech sequences in adults, and of non-speech sequences in 7.5-month-old infants. The goal of the present study was to explore whether language experience also modulates infants' grouping of speech. To do so, we presented sequences of syllables to monolingual French- and German-learning 7.5-month-olds. Using the Headturn Preference Procedure (HPP), we examined whether they were able to perceive a rhythmic structure in sequences of syllables that alternated in duration, pitch, or intensity. Our findings show that both French- and German-learning infants perceived a rhythmic structure when it was cued by duration or pitch but not intensity. Our findings also show differences in how these infants use duration and pitch cues to group syllable sequences, suggesting that pitch cues were the easier ones to use. Moreover, performance did not differ across languages, failing to reveal early language effects on rhythmic perception. These results contribute to our understanding of the origin of rhythmic perception and perceptual mechanisms shared across music and speech, which may bootstrap language acquisition.
An Exploration of Rhythmic Grouping of Speech Sequences by French- and German-Learning Infants
(2016)
Rhythm in music and speech can be characterized by a constellation of several acoustic cues. Individually, these cues have different effects on rhythmic perception: sequences of sounds alternating in duration are perceived as short-long pairs (weak-strong/iambic pattern), whereas sequences of sounds alternating in intensity or pitch are perceived as loud-soft, or high-low pairs (strong-weak/trochaic pattern). This perceptual bias—called the Iambic-Trochaic Law (ITL)–has been claimed to be an universal property of the auditory system applying in both the music and the language domains. Recent studies have shown that language experience can modulate the effects of the ITL on rhythmic perception of both speech and non-speech sequences in adults, and of non-speech sequences in 7.5-month-old infants. The goal of the present study was to explore whether language experience also modulates infants’ grouping of speech. To do so, we presented sequences of syllables to monolingual French- and German-learning 7.5-month-olds. Using the Headturn Preference Procedure (HPP), we examined whether they were able to perceive a rhythmic structure in sequences of syllables that alternated in duration, pitch, or intensity. Our findings show that both French- and German-learning infants perceived a rhythmic structure when it was cued by duration or pitch but not intensity. Our findings also show differences in how these infants use duration and pitch cues to group syllable sequences, suggesting that pitch cues were the easier ones to use. Moreover, performance did not differ across languages, failing to reveal early language effects on rhythmic perception. These results contribute to our understanding of the origin of rhythmic perception and perceptual mechanisms shared across music and speech, which may bootstrap language acquisition.
Pivots revisited
(2021)
The term "pivot" usually refers to two overlapping syntactic units such that the completion of the first unit simultaneously launches the second. In addition, pivots are generally said to be characterized by the smooth prosodic integration of their syntactic parts. This prosodic integration is typically achieved by prosodic-phonetic matching of the pivot components. As research on such turns in a range of languages has illustrated, speakers routinely deploy pivots so as to be able to continue past a point of possible turn completion, in the service of implementing some additional or revised action. This article seeks to build on, and complement, earlier research by exploring two issues in more detail as follows: (1) what exactly do pivotal turn extensions accomplish on the action dimension, and (2) what role does prosodic-phonetic packaging play in this? We will show that pivot constructions not only exhibit various degrees of prosodic-phonetic (non-)integration, i.e., differently strong cesuras, but that they can be ordered on a continuum, and that this cline maps onto the relationship of the actions accomplished by the components of the pivot construction. While tighter prosodic-phonetic integration, i.e., weak(er) cesuring, co-occurs with post-pivot actions whose relationship to that of the pre-pivot tends to be rather retrospective in character, looser prosodic-phonetic integration, i.e., strong(er) cesuring, is associated with a more prospective orientation of the post-pivot's action. These observations also raise more general questions with regard to the analysis of action.
“Chunking” spoken language
(2021)
In this introductory paper to the special issue on “Weak cesuras in talk-in-interaction”, we aim to guide the reader into current work on the “chunking” of naturally occurring talk. It is conducted in the methodological frameworks of Conversation Analysis and Interactional Linguistics – two approaches that consider the interactional aspect of humans talking with each other to be a crucial starting point for its analysis. In doing so, we will (1) lay out the background of this special issue (what is problematic about “chunking” talk-in-interaction, the characteristics of the methodological approach chosen by the contributors, the cesura model), (2) highlight what can be gained from such a revised understanding of “chunking” in talk-in-interaction by referring to previous work with this model as well as the findings of the contributions to this special issue, and (3) indicate further directions such work could take starting from papers in this special issue. We hope to induce a fruitful exchange on the phenomena discussed, across methodological divides.
“Chunking” spoken language
(2021)
In this introductory paper to the special issue on “Weak cesuras in talk-in-interaction”, we aim to guide the reader into current work on the “chunking” of naturally occurring talk. It is conducted in the methodological frameworks of Conversation Analysis and Interactional Linguistics – two approaches that consider the interactional aspect of humans talking with each other to be a crucial starting point for its analysis. In doing so, we will (1) lay out the background of this special issue (what is problematic about “chunking” talk-in-interaction, the characteristics of the methodological approach chosen by the contributors, the cesura model), (2) highlight what can be gained from such a revised understanding of “chunking” in talk-in-interaction by referring to previous work with this model as well as the findings of the contributions to this special issue, and (3) indicate further directions such work could take starting from papers in this special issue. We hope to induce a fruitful exchange on the phenomena discussed, across methodological divides.
Infants start learning the prosodic properties of their native language before 12 months, as shown by the emergence of a trochaic bias in English-learning infants between 6 and 9 months (Jusczyk et al., 1993), and in German-learning infants between 4 and 6 months (Huhle et al., 2009, 2014), while French-learning infants do not show a bias at 6 months (Hohle et al., 2009). This language-specific emergence of a trochaic bias is supported by the fact that English and German are languages with trochaic predominance in their lexicons, while French is a language with phrase-final lengthening but lacking lexical stress. We explored the emergence of a trochaic bias in bilingual French/German infants, to study whether the developmental trajectory would be similar to monolingual infants and whether amount of relative exposure to the two languages has an impact on the emergence of the bias. Accordingly, we replicated Hohle et al. (2009) with 24 bilingual 6-month-olds learning French and German simultaneously. All infants had been exposed to both languages for 30 to 70% of the time from birth. Using the Head Preference Procedure, infants were presented with two lists of stimuli, one made up of several occurrences of the pseudoword /GAba/ with word-initial stress (trochaic pattern), the second one made up of several occurrences of the pseudoword /gaBA/ with word-final stress (iambic pattern). The stimuli were recorded by a native German female speaker. Results revealed that these French/German bilingual 6-month olds have a trochaic bias (as evidenced by a preference to listen to the trochaic pattern). Hence, their listening preference is comparable to that of monolingual German-learning 6-month-olds, but differs from that of monolingual French-learning 6-month-olds who did not show any preference (Noble et al., 2009). Moreover, the size of the trochaic bias in the bilingual infants was not correlated with their amount of exposure to German. The present results thus establish that the development of a trochaic bias in simultaneous bilinguals is not delayed compared to monolingual German-learning infants (Hohle et al., 2009) and is rather independent of the amount of exposure to German relative to French.
Infants start learning the prosodic properties of their native language before 12 months, as shown by the emergence of a trochaic bias in English-learning infants between 6 and 9 months (Jusczyk et al., 1993), and in German-learning infants between 4 and 6 months (Huhle et al., 2009, 2014), while French-learning infants do not show a bias at 6 months (Hohle et al., 2009). This language-specific emergence of a trochaic bias is supported by the fact that English and German are languages with trochaic predominance in their lexicons, while French is a language with phrase-final lengthening but lacking lexical stress. We explored the emergence of a trochaic bias in bilingual French/German infants, to study whether the developmental trajectory would be similar to monolingual infants and whether amount of relative exposure to the two languages has an impact on the emergence of the bias. Accordingly, we replicated Hohle et al. (2009) with 24 bilingual 6-month-olds learning French and German simultaneously. All infants had been exposed to both languages for 30 to 70% of the time from birth. Using the Head Preference Procedure, infants were presented with two lists of stimuli, one made up of several occurrences of the pseudoword /GAba/ with word-initial stress (trochaic pattern), the second one made up of several occurrences of the pseudoword /gaBA/ with word-final stress (iambic pattern). The stimuli were recorded by a native German female speaker. Results revealed that these French/German bilingual 6-month olds have a trochaic bias (as evidenced by a preference to listen to the trochaic pattern). Hence, their listening preference is comparable to that of monolingual German-learning 6-month-olds, but differs from that of monolingual French-learning 6-month-olds who did not show any preference (Noble et al., 2009). Moreover, the size of the trochaic bias in the bilingual infants was not correlated with their amount of exposure to German. The present results thus establish that the development of a trochaic bias in simultaneous bilinguals is not delayed compared to monolingual German-learning infants (Hohle et al., 2009) and is rather independent of the amount of exposure to German relative to French.
Language and music share many rhythmic properties, such as variations in intensity and duration leading to repeating patterns. Perception of rhythmic properties may rely on cognitive networks that are shared between the two domains. If so, then variability in speech rhythm perception may relate to individual differences in musicality. To examine this possibility, the present study focuses on rhythmic grouping, which is assumed to be guided by a domain-general principle, the Iambic/Trochaic law, stating that sounds alternating in intensity are grouped as strong-weak, and sounds alternating in duration are grouped as weak-strong. German listeners completed a grouping task: They heard streams of syllables alternating in intensity, duration, or neither, and had to indicate whether they perceived a strong-weak or weak-strong pattern. Moreover, their music perception abilities were measured, and they filled out a questionnaire reporting their productive musical experience. Results showed that better musical rhythm perception - ability was associated with more consistent rhythmic grouping of speech, while melody perception - ability and productive musical experience were not. This suggests shared cognitive procedures in the perception of rhythm in music and speech. Also, the results highlight the relevance of - considering individual differences in musicality when aiming to explain variability in prosody perception.
Rhythm perception is assumed to be guided by a domain-general auditory principle, the Iambic/Trochaic Law, stating that sounds varying in intensity are grouped as strong-weak, and sounds varying in duration are grouped as weak-strong. Recently, Bhatara et al. (2013) showed that rhythmic grouping is influenced by native language experience, French listeners having weaker grouping preferences than German listeners. This study explores whether L2 knowledge and musical experience also affect rhythmic grouping. In a grouping task, French late learners of German listened to sequences of coarticulated syllables varying in either intensity or duration. Data on their language and musical experience were obtained by a questionnaire. Mixed-effect model comparisons showed influences of musical experience as well as L2 input quality and quantity on grouping preferences. These results imply that adult French listeners' sensitivity to rhythm can be enhanced through L2 and musical experience.
This study investigated Dutch-speaking four- to five-year-olds’ use of word order and prosody in distinguishing focus types (broad focus, narrow focus, and contrastive narrow focus) via an interactive answer-reconstruction game. We have found an overall preference for the unmarked word order SVO and no evidence for the use of OVS to distinguish focus types. But the children used pitch and duration in the subject-nouns to distinguish focus types in SVO sentences. These findings show that Dutch-speaking four- to five-year-olds differ from their German- and Finnish-speaking peers, who show evidence of varying choice of word order to mark specific focus types, and use prosody to distinguish focus types in subject and object nouns in both SVO and OVS sentences. These comparisons suggest that typological differences in the relative importance between word order and prosody can lead to differences in children’s use of word order and prosody in unmarked and marked word orders. A more equal role of word order and prosody in the ambient language can stimulate more extensive use of prosody in the marked word order, whereas a more limited role of word order can restrict the use of prosody in the unmarked word order.
Aging in speech production is a multidimensional process. Biological, cognitive, social, and communicative factors can change over time, stay relatively stable, or may even compensate for each other. In this longitudinal work, we focus on stability and change at the laryngeal and supralaryngeal levels in the discourse particle euh produced by 10 older French-speaking females at two times, 10 years apart. Recognizing the multiple discourse roles of euh, we divided out occurrences according to utterance position. We quantified the frequency of euh, and evaluated acoustic changes in formants, fundamental frequency, and voice quality across time and utterance position. Results showed that euh frequency was stable with age. The only acoustic measure that revealed an age effect was harmonics-to-noise ratio, showing less noise at older ages. Other measures mostly varied with utterance position, sometimes in interaction with age. Some voice quality changes could reflect laryngeal adjustments that provide for airflow conservation utterance-finally. The data suggest that aging effects may be evident in some prosodic positions (e.g., utterance-final position), but not others (utterance-initial position). Thus, it is essential to consider the interactions among these factors in future work and not assume that vocal aging is evident throughout the signal.
This dissertation is about factors that contribute to the surface forms of tones in connected speech in Akan. Akan is an African tone language, which is spoken in Ghana. It has two level tones (low and high), automatic and non-automatic downstep. Downstep is the major factor that influences the surface forms of tones. The thesis shows that downstep is caused by declination. It is argued that declination is an intonational property of Akan, which serves to signal coherence. A phonological representation using a high and a low register tone, associating to the left and right edge of an intonational phrase (IP), respectively, is proposed. Declination/downstep is modelled using a (phonetic) pitch implementation algorithm (Liberman & Pierrehumbert, 1984). An innovative application of the algorithm is presented, which naturally captures the relation between declination and downstep in Akan. Another important factor is the prosodic manifestation of sentence level pragmatic meanings, such as sentence mode and focus. Regarding the former, the thesis shows that a post-lexical low tone, which associates with the right edge of an IP, signals interrogativity. Additionally, lexical tones in Yes – No questions are realized in a higher pitch register, which does not lead to a reduction of declination. It is claimed that the higher register is not part of the phonological representation in Akan, but that it emerges at the phonetic level to compensate for the ‘unnatural’ form of the question morpheme and to satisfy the Frequency code (Gussenhoven, 2002; 2004). An extension of Rialland’s (2007) typology in terms of a new category called “low tense” question prosody is proposed. Concerning focus marking, it is argued that the use of the morpho-syntactic focus marking strategy is related to extra grammatical factors, such as hearer expectation, discourse expectability (Zimmermann, 2007) and emphasis (Hartmann, 2008). If a speaker of Akan wants to highlight a particular element in a sentence, in-situ, i.e. by means of prosody, the default prosodic structure is modified in such a way that the focused element forms its own phonological phrase (pP). If it is already contained in a pP, the boundary deliminating the focused element is enhanced (Féry, 2012). This restructuring/enhancement is accompanied by an interruption of the otherwise continuous melody due to insertion of a pause and/or a glottal stop. Beside declination and intonation, raising of H tones applies in Akan. H raising is analyzed as a local anticipatory planning effect, employed at the phonetic level, which enhances the perceptual distance between low and high tones. Low tones are raised, if they are wedged between two high tones. L raising is argued to be a local carryover effect (co-articulation). Further, it is demonstrated that global anticipatory raising takes place. It is shown that Akan speakers anticipate the length of an IP. Preplanning (anticipatory raising) is argued to be an important process at the level of pitch implementation. It serves to ensure that declination can be maintained throughout the IP, which prevents pitch resetting.
The melody of an Akan sentence is largely determined by the choice of words. The inventory of post-lexical tones is small. It consists of post-lexical register tones, which trigger declination and post-lexical intonational tones, which signal sentence type. The overall melodic shape is falling. At the local level, H raising and L raising occur. At the global level, initial low and high tones are realized higher if they occur in a long and/or complex sentence. This dissertation shows that many factors, which emerge at different levels of the tone production process, contribute to the surface form of tones in Akan.
One of the central questions in psycholinguistic is understanding whether and how prosodic phrase boundaries are used to resolve syntactic ambiguities in sentence processing. The present work aimed to answer both, first, the effects of φ- and ι-boundaries on syntactic ambiguity resolution, and second, how the prosodic correlates of the auditory input are taken for the phonetic-phonology mapping in order to attain a meaningful sentence interpretation.
With regard to the first aim, we investigated locally syntactic ambiguities involving either φ- or ι-phrase boundaries in German and the structural preference that listeners have, based on the prosodic content. The experiments described in this work show that German listeners exploit both types of prosodic phrase boundaries to resolve local syntactic ambiguities, that however, their disambiguation altered by the presence or absence of prosodic cues correlated with the corresponding boundary. Specifically, the perception data revealed that the phonetically measured prosodic correlates of each prosodic boundary such as pitch accents, boundary tones, deaccentuation and durational properties do not contribute to ambiguity resolution in equal measure. Rather, it is the case that listeners rely primarily on prefinal lengthening as a correlate of phrasing in the vicinity of φ-phrase boundaries, while at the level of the ι-phrase boundary, boundary tones serve as phrasal cues. This way the results of the present work take account of the as yet missing information on individual contributions of prosodic correlates on listeners’ disambiguation of syntactically ambiguous sentences in German. It further implies that the question of how German listeners resolve syntactic ambiguities cannot simply be attributed to the presence or absence of prosodic correlates. The interpretation of the phrasal structure rather depends on a more general picture of cohesion between prosodic correlates and prosodic boundary sizes.
With respect to the second aim, the processing models proposed in the present work describe a specific phonetics-phonology mapping in the vicinity of both phrase boundaries. It is assumed that auditory sentence processing proceeds in several successively organized steps, during which listeners transform overt phonetic forms into language specific abstract surface forms. This process is referred to as phonetics-phonology mapping in the present work. Perceptual evidence resulting from the experiments of the present work suggest that the phonetics-phonology mapping is guided by the above mentioned boundary related prosodic correlates. The resulting abstract phonological structure is subjected to the syntax-prosody mapping, in turn. The outcome of the presented perception experiments are modulated in an Optimality-Theoretic framework. The offered OT-models are grounded on the assumption that single prosodic correlates are used by listeners as a signal to syntax in sentence processing. This is in line with studies arguing that the prosodic phrase structure determines the syntactic parse (Cutler et al., 1997; Warren et al., 1995; Pynte & Prieur, 1996; Snedeker & Trueswell, 2003; Kjelgaard & Speer, 1999), to name just a few.
Information structure has been one of the central topics of recent linguistic research. This review discusses a wide range of current approaches with particular reference to African languages, as these have been playing a crucial role in advancing our knowledge about the diversity of and recurring patterns in both meaning and form of information structural notions. We focus on cross-linguistic functional frameworks, the investigation of prosody, formal syntactic theories, and relevant effects of semantic interpretation. Information structure is a thriving research domain that promises to yield important advances in our general understanding of human language.
Prosody is a rich source of information that heavily supports spoken language comprehension. In particular, prosodic phrase boundaries divide the continuous speech stream into chunks reflecting the semantic and syntactic structure of an utterance. This chunking or prosodic phrasing plays a critical role in both spoken language processing and language acquisition. Aiming at a better understanding of the underlying processing mechanisms and their acquisition, the present work investigates factors that influence prosodic phrase boundary perception in adults and infants. Using the event-related potential (ERP) technique, three experimental studies examined the role of prosodic context (i.e., phrase length) in German phrase boundary perception and of the main prosodic boundary cues, namely pitch change, final lengthening, and pause. With regard to the boundary cues, the dissertation focused on the questions which cues or cue combination are essential for the perception of a prosodic boundary and on whether and how this cue weighting develops during infancy.
Using ERPs is advantageous because the technique captures the immediate impact of (linguistic) information during on-line processing. Moreover, as it can be applied independently of specific task demands or an overt response performance, it can be used with both infants and adults. ERPs are particularly suitable to study the time course and underlying mechanisms of boundary perception, because a specific ERP component, the Closure Positive Shift (CPS) is well established as neuro-physiological indicator of prosodic boundary perception in adults.
The results of the three experimental studies first underpin that the prosodic context plays an immediate role in the processing of prosodic boundary information. Moreover, the second study reveals that adult listeners perceive a prosodic boundary also on the basis of a sub-set of the boundary cues available in the speech signal. Both ERP and simultaneously collected behavioral data (i.e., prosodic judgements) suggest that the combination of pitch change and final lengthening triggers boundary perception; however, when presented as single cues, neither pitch change nor final lengthening were sufficient. Finally, testing six- and eight-month-old infants shows that the early sensitivity for prosodic information is reflected in a brain response resembling the adult CPS. For both age groups, brain responses to prosodic boundaries cued by pitch change and final lengthening revealed a positivity that can be interpreted as a CPS-like infant ERP component. In contrast, but comparable to the adults’ response pattern, pitch change as a single cue does not provoke an infant CPS. These results show that infant phrase boundary perception is not exclusively based on pause detection and hint at an early ability to exploit subtle, relational prosodic cues in speech perception.
This study examines the role of pitch and final lengthening in German intonation phrase boundary (IPB) perception. Since a prosody-related event-related potential (ERP) component termed Closure Positive Shift reflects the processing of major prosodic boundaries, we combined ERP and behavioural measures (i.e. a prosodic judgement task) to systematically test the impact of sole and combined cue occurrences on IPB perception. In two experiments we investigated whether adult listeners perceived an IPB in acoustically manipulated speech material that contained none, one, or two of the prosodic boundary cues. Both ERP and behavioural results suggest that pitch and final lengthening cues have to occur in combination to trigger IPB perception. Hence, the combination of behavioural and electrophysiological measures provides a comprehensive insight into prosodic boundary cue perception in German and leads to an argument in favour of interrelated cues from the frequency (i.e. pitch change) and the time (i.e. final lengthening) domain.
Prosodic information is crucial for spoken language comprehension and especially for syntactic parsing, because prosodic cues guide the hearer's syntactic analysis. The time course and mechanisms of this interplay of prosody and syntax are not yet well-understood. In particular, there is an ongoing debate whether local prosodic cues are taken into account automatically or whether they are processed in relation to the global prosodic context in which they appear. The present study explores whether the perception of a prosodic boundary is affected by its position within an utterance. In an event-related potential (PRP) study we tested if the brain response evoked by the prosodic boundary differs when the boundary occurs early in a list of three names connected by conjunctions (i.e., after the first name) as compared to later in the utterance (i.e., after the second name). A closure positive shift (CPS)-marking the processing of a prosodic phrase boundary-was elicited for stimuli with a late boundary, but not for stimuli with an early boundary. This result is further evidence for an immediate integration of prosodic information into the parsing of an utterance. In addition, it shows that the processing of prosodic boundary cues depends on the previously processed information from the preceding prosodic context.
Prosody by phase
(2004)
Japanese wh-questions always exhibit focus intonation (FI). Furthermore, the domain of FI exhibits a correspondence to the wh-scope. I propose that this phonology-semantics correspondence is a result of the cyclic computation of FI, which is explained under the notion of Multiple Spell-Out in the recent Minimalist framework. The proposed analysis makes two predictions: (1) embedding of an FI into another is possible; (2) (overt) movement of a wh-phrase to a phase edge position causes a mismatch between FI and wh-scope. Both predictions are tested experimentally, and shown to be borne out.
In this paper we review the current state of research on the issue of discourse structure (DS)/information structure (IS) interface. This field has received a lot of attention from discourse semanticists and pragmatists, and has made substantial progress in recent years. In this paper we summarize the relevant studies. In addition, we look at the issue of DS/ISinteraction at a different level - that of phonetics. It is known that both information structure and discourse structure can be realized prosodically, but the issue of phonetic interaction between the prosodic devices they employ has hardly ever been discussed in this context. We think that a proper consideration of this aspect of DS/IS-interaction would enrich our understanding of the phenomenon, and hence we formulate some related research-programmatic positions.