@article{BollAvetisyanBhataraHoehle2017, author = {Boll-Avetisyan, Natalie and Bhatara, Anjali and H{\"o}hle, Barbara}, title = {Effects of musicality on the perception of rhythmic structure in speech}, series = {Laboratory phonology}, volume = {8}, journal = {Laboratory phonology}, number = {1}, publisher = {Ubiquity Press}, address = {London}, issn = {1868-6346}, doi = {10.5334/labphon.91}, pages = {16}, year = {2017}, abstract = {Language and music share many rhythmic properties, such as variations in intensity and duration leading to repeating patterns. Perception of rhythmic properties may rely on cognitive networks that are shared between the two domains. If so, then variability in speech rhythm perception may relate to individual differences in musicality. To examine this possibility, the present study focuses on rhythmic grouping, which is assumed to be guided by a domain-general principle, the Iambic/Trochaic law, stating that sounds alternating in intensity are grouped as strong-weak, and sounds alternating in duration are grouped as weak-strong. German listeners completed a grouping task: They heard streams of syllables alternating in intensity, duration, or neither, and had to indicate whether they perceived a strong-weak or weak-strong pattern. Moreover, their music perception abilities were measured, and they filled out a questionnaire reporting their productive musical experience. Results showed that better musical rhythm perception - ability was associated with more consistent rhythmic grouping of speech, while melody perception - ability and productive musical experience were not. This suggests shared cognitive procedures in the perception of rhythm in music and speech. Also, the results highlight the relevance of - considering individual differences in musicality when aiming to explain variability in prosody perception.}, language = {en} } @article{GafosLieshout2020, author = {Gafos, Adamantios I. and Lieshout, Pascal H. H. M. van}, title = {Models and theories of speech production}, series = {Frontiers in psychology}, volume = {11}, journal = {Frontiers in psychology}, publisher = {Frontiers Research Foundation}, address = {Lausanne}, issn = {1664-1078}, doi = {10.3389/fpsyg.2020.01238}, pages = {4}, year = {2020}, language = {en} } @article{MokariGafosWilliams2021, author = {Mokari, Payam Ghaffarvand and Gafos, Adamantios I. and Williams, Daniel}, title = {Perceptuomotor compatibility effects in vowels}, series = {JASA Express Letters}, volume = {1}, journal = {JASA Express Letters}, number = {1}, publisher = {American Institute of Physics}, address = {Melville}, issn = {2691-1191}, doi = {10.1121/10.0003039}, pages = {8}, year = {2021}, abstract = {In a cue-distractor task, speakers' response times (RTs) were found to speed up when they perceived a distractor syllable whose vowel was identical to the vowel in the syllable they were preparing to utter. At a more fine-grained level, subphonemic congruency between response and distractor-defined by higher number of shared phonological features or higher acoustic proximity-was also found to be predictive of RT modulations. Furthermore, the findings indicate that perception of vowel stimuli embedded in syllables gives rise to robust and more consistent perceptuomotor compatibility effects (compared to isolated vowels) across different response-distractor vowel pairs.}, language = {en} } @misc{OffredeJacobiReberniketal.2020, author = {Offrede, Tom F. and Jacobi, Jidde and Rebernik, Teja and de Jong, Lisanne and Keulen, Stefanie and Veenstra, Pauline and Noiray, Aude and Wieling, Martijn}, title = {The impact of alcohol on L1 versus L2}, series = {Zweitver{\"o}ffentlichungen der Universit{\"a}t Potsdam : Humanwissenschaftliche Reihe}, journal = {Zweitver{\"o}ffentlichungen der Universit{\"a}t Potsdam : Humanwissenschaftliche Reihe}, number = {3}, issn = {1866-8364}, doi = {10.25932/publishup-54095}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-540955}, pages = {14}, year = {2020}, abstract = {Alcohol intoxication is known to affect many aspects of human behavior and cognition; one of such affected systems is articulation during speech production. Although much research has revealed that alcohol negatively impacts pronunciation in a first language (L1), there is only initial evidence suggesting a potential beneficial effect of inebriation on articulation in a non-native language (L2). The aim of this study was thus to compare the effect of alcohol consumption on pronunciation in an L1 and an L2. Participants who had ingested different amounts of alcohol provided speech samples in their L1 (Dutch) and L2 (English), and native speakers of each language subsequently rated the pronunciation of these samples on their intelligibility (for the L1) and accent nativelikeness (for the L2). These data were analyzed with generalized additive mixed modeling. Participants' blood alcohol concentration indeed negatively affected pronunciation in L1, but it produced no significant effect on the L2 accent ratings. The expected negative impact of alcohol on L1 articulation can be explained by reduction in fine motor control. We present two hypotheses to account for the absence of any effects of intoxication on L2 pronunciation: (1) there may be a reduction in L1 interference on L2 speech due to decreased motor control or (2) alcohol may produce a differential effect on each of the two linguistic subsystems.}, language = {en} } @article{OffredeJacobiReberniketal.2020, author = {Offrede, Tom F. and Jacobi, Jidde and Rebernik, Teja and de Jong, Lisanne and Keulen, Stefanie and Veenstra, Pauline and Noiray, Aude and Wieling, Martijn}, title = {The impact of alcohol on L1 versus L2}, series = {Language and Speech}, volume = {64}, journal = {Language and Speech}, number = {3}, publisher = {SAGE Publications}, address = {Thousand Oaks}, issn = {1756-6053}, doi = {10.1177/0023830920953169}, pages = {681 -- 692}, year = {2020}, abstract = {Alcohol intoxication is known to affect many aspects of human behavior and cognition; one of such affected systems is articulation during speech production. Although much research has revealed that alcohol negatively impacts pronunciation in a first language (L1), there is only initial evidence suggesting a potential beneficial effect of inebriation on articulation in a non-native language (L2). The aim of this study was thus to compare the effect of alcohol consumption on pronunciation in an L1 and an L2. Participants who had ingested different amounts of alcohol provided speech samples in their L1 (Dutch) and L2 (English), and native speakers of each language subsequently rated the pronunciation of these samples on their intelligibility (for the L1) and accent nativelikeness (for the L2). These data were analyzed with generalized additive mixed modeling. Participants' blood alcohol concentration indeed negatively affected pronunciation in L1, but it produced no significant effect on the L2 accent ratings. The expected negative impact of alcohol on L1 articulation can be explained by reduction in fine motor control. We present two hypotheses to account for the absence of any effects of intoxication on L2 pronunciation: (1) there may be a reduction in L1 interference on L2 speech due to decreased motor control or (2) alcohol may produce a differential effect on each of the two linguistic subsystems.}, language = {en} } @phdthesis{LopezGambino2023, author = {L{\´o}pez Gambino, Maria Soledad}, title = {Time Buying in Task-Oriented Spoken Dialogue Systems}, doi = {10.25932/publishup-59280}, url = {http://nbn-resolving.de/urn:nbn:de:kobv:517-opus4-592806}, school = {Universit{\"a}t Potsdam}, pages = {148}, year = {2023}, abstract = {This dissertation focuses on the handling of time in dialogue. Specifically, it investigates how humans bridge time, or "buy time", when they are expected to convey information that is not yet available to them (e.g. a travel agent searching for a flight in a long list while the customer is on the line, waiting). It also explores the feasibility of modeling such time-bridging behavior in spoken dialogue systems, and it examines how endowing such systems with more human-like time-bridging capabilities may affect humans' perception of them. The relevance of time-bridging in human-human dialogue seems to stem largely from a need to avoid lengthy pauses, as these may cause both confusion and discomfort among the participants of a conversation (Levinson, 1983; Lundholm Fors, 2015). However, this avoidance of prolonged silence is at odds with the incremental nature of speech production in dialogue (Schlangen and Skantze, 2011): Speakers often start to verbalize their contribution before it is fully formulated, and sometimes even before they possess the information they need to provide, which may result in them running out of content mid-turn. In this work, we elicit conversational data from humans, to learn how they avoid being silent while they search for information to convey to their interlocutor. We identify commonalities in the types of resources employed by different speakers, and we propose a classification scheme. We explore ways of modeling human time-buying behavior computationally, and we evaluate the effect on human listeners of embedding this behavior in a spoken dialogue system. Our results suggest that a system using conversational speech to bridge time while searching for information to convey (as humans do) can provide a better experience in several respects than one which remains silent for a long period of time. However, not all speech serves this purpose equally: Our experiments also show that a system whose time-buying behavior is more varied (i.e. which exploits several categories from the classification scheme we developed and samples them based on information from human data) can prevent overestimation of waiting time when compared, for example, with a system that repeatedly asks the interlocutor to wait (even if these requests for waiting are phrased differently each time). Finally, this research shows that it is possible to model human time-buying behavior on a relatively small corpus, and that a system using such a model can be preferred by participants over one employing a simpler strategy, such as randomly choosing utterances to produce during the wait —even when the utterances used by both strategies are the same.}, language = {en} }