Refine
Has Fulltext
- yes (180) (remove)
Year of publication
- 2023 (180) (remove)
Document Type
- Doctoral Thesis (180) (remove)
Is part of the Bibliography
- yes (180)
Keywords
- climate change (9)
- Klimawandel (8)
- machine learning (6)
- Modellierung (3)
- Reflexion (3)
- körperliche Fitness (3)
- maschinelles Lernen (3)
- physical fitness (3)
- reinforcement learning (3)
- Anden (2)
Institute
- Institut für Biochemie und Biologie (26)
- Institut für Geowissenschaften (24)
- Extern (23)
- Institut für Physik und Astronomie (21)
- Institut für Chemie (18)
- Hasso-Plattner-Institut für Digital Engineering GmbH (17)
- Institut für Umweltwissenschaften und Geographie (13)
- Department Psychologie (6)
- Department Sport- und Gesundheitswissenschaften (6)
- Institut für Ernährungswissenschaft (6)
- Department Erziehungswissenschaft (5)
- Department Linguistik (4)
- Institut für Informatik und Computational Science (4)
- Institut für Mathematik (4)
- Digital Engineering Fakultät (3)
- Fachgruppe Betriebswirtschaftslehre (3)
- Fachgruppe Politik- & Verwaltungswissenschaft (3)
- Fakultät für Gesundheitswissenschaften (2)
- Historisches Institut (2)
- Wirtschaftswissenschaften (2)
- Applied Computational Linguistics (1)
- Department Grundschulpädagogik (1)
- Department Musik und Kunst (1)
- Fachgruppe Soziologie (1)
- Fachgruppe Volkswirtschaftslehre (1)
- Institut für Germanistik (1)
- Institut für Künste und Medien (1)
- Institut für Romanistik (1)
- Language Acquisition (1)
- Patholinguistics/Neurocognition of Language (1)
- Phonology & Phonetics (1)
- Potsdam Institute for Climate Impact Research (PIK) e. V. (1)
- Strukturbereich Kognitionswissenschaften (1)
Die vorliegende Studie beschäftigt sich mit dem nach einer Strukturveränderung in der Sekundarstufe I entstandenen Schulmodell der Neuen Mittelschule. Untersucht wird, ob sich durch dieses Schulmodell und der damit intendierten neuen Lehr-, Lern- und Prüfungskultur Zusammenhänge zwischen gemessenen mathematischen Kompetenzen der Schüler und den durch Lehrer vergebenen Jahresnoten feststellen lassen.
Die Literaturrecherche macht deutlich, dass die Kritik an der Monokultur des leh-rerzentrierten Unterrichts zwar zu einer neuen Lehr-, Lern- und Prüfungskultur führt, deren Inhalte sind aber recht unterschiedlich, komplex und nicht eindeutig definiert. In der NMS soll die Leistungsbewertung als Lernhilfe fungieren, aber auch verlässliche Aussagen über die Leistung der Schüler treffen. Zur Wirkung der neuen Lernkultur in der NMS gibt es ebenso keine empirischen Befunde wie über die Wirkung der Leistungsbewertung.
An der empirischen Untersuchung nehmen 79 Schüler der sechsten Schulstufe aus drei Neuen Mittelschulen (dicht besiedelte, mittel besiedelte, dünn besiedelte Gemeinde) in Niederösterreich teil. In jeder Schule werden zwei Klassen untersucht. Dabei werden der Kompetenzstand in Mathematik, Schülerzentriertheit sowie Sozial- und Leistungsdruck aus Sicht der Schüler gemeinsam mit der Jah-resnote erhoben.
Für die Studie wird ein Pfadmodell entwickelt und mit einer Pfadanalyse ausge-wertet. Dabei zeigen sich zwar Zusammenhänge zwischen den gemessenen Kompetenzen in Mathematik und den Jahresnoten. Diese Jahresnoten besitzen über die Klasse bzw. die Schule hinaus aber nur eine bedingte Aussagekraft über die erbrachten Leistungen.
With the recent growth of sensors, cloud computing handles the data processing of many applications. Processing some of this data on the cloud raises, however, many concerns regarding, e.g., privacy, latency, or single points of failure. Alternatively, thanks to the development of embedded systems, smart wireless devices can share their computation capacity, creating a local wireless cloud for in-network processing. In this context, the processing of an application is divided into smaller jobs so that a device can run one or more jobs.
The contribution of this thesis to this scenario is divided into three parts. In part one, I focus on wireless aspects, such as power control and interference management, for deciding which jobs to run on which node and how to route data between nodes. Hence, I formulate optimization problems and develop heuristic and meta-heuristic algorithms to allocate wireless and computation resources. Additionally, to deal with multiple applications competing for these resources, I develop a reinforcement learning (RL) admission controller to decide which application should be admitted. Next, I look into acoustic applications to improve wireless throughput by using microphone clock synchronization to synchronize wireless transmissions.
In the second part, I jointly work with colleagues from the acoustic processing field to optimize both network and application (i.e., acoustic) qualities. My contribution focuses on the network part, where I study the relation between acoustic and network qualities when selecting a subset of microphones for collecting audio data or selecting a subset of optional jobs for processing these data; too many microphones or too many jobs can lessen quality by unnecessary delays. Hence, I develop RL solutions to select the subset of microphones under network constraints when the speaker is moving while still providing good acoustic quality. Furthermore, I show that autonomous vehicles carrying microphones improve the acoustic qualities of different applications. Accordingly, I develop RL solutions (single and multi-agent ones) for controlling these vehicles.
In the third part, I close the gap between theory and practice. I describe the features of my open-source framework used as a proof of concept for wireless in-network processing. Next, I demonstrate how to run some algorithms developed by colleagues from acoustic processing using my framework. I also use the framework for studying in-network delays (wireless and processing) using different distributions of jobs and network topologies.
Hybrid nanomaterials offer the combination of individual properties of different types of nanoparticles. Some strategies for the development of new nanostructures in larger scale rely on the self-assembly of nanoparticles as a bottom-up approach. The use of templates provides ordered assemblies in defined patterns. In a typical soft-template, nanoparticles and other surface-active agents are incorporated into non-miscible liquids. The resulting self-organized dispersions will mediate nanoparticle interactions to control the subsequent self-assembly. Especially interactions between nanoparticles of very different dispersibility and functionality can be directed at a liquid-liquid interface.
In this project, water-in-oil microemulsions were formulated from quasi-ternary mixtures with Aerosol-OT as surfactant. Oleyl-capped superparamagnetic iron oxide and/or silver nanoparticles were incorporated in the continuous organic phase, while polyethyleneimine-stabilized gold nanoparticles were confined in the dispersed water droplets. Each type of nanoparticle can modulate the surfactant film and the inter-droplet interactions in diverse ways, and their combination causes synergistic effects. Interfacial assemblies of nanoparticles resulted after phase-separation. On one hand, from a biphasic Winsor type II system at low surfactant concentration, drop-casting of the upper phase afforded thin films of ordered nanoparticles in filament-like networks. Detailed characterization proved that this templated assembly over a surface is based on the controlled clustering of nanoparticles and the elongation of the microemulsion droplets. This process offers versatility to use different nanoparticle compositions by keeping the surface functionalization, in different solvents and over different surfaces. On the other hand, a magnetic heterocoagulate was formed at higher surfactant concentration, whose phase-transfer from oleic acid to water was possible with another auxiliary surfactant in ethanol-water mixture. When the original components were initially mixed under heating, defined oil-in-water, magnetic-responsive nanostructures were obtained, consisting on water-dispersible nanoparticle domains embedded by a matrix-shell of oil-dispersible nanoparticles.
Herein, two different approaches were demonstrated to form diverse hybrid nanostructures from reverse microemulsions as self-organized dispersions of the same components. This shows that microemulsions are versatile soft-templates not only for the synthesis of nanoparticles, but also for their self-assembly, which suggest new approaches towards the production of new sophisticated nanomaterials in larger scale.
Volcanoes are one of the Earth’s most dynamic zones and responsible for many changes in our planet. Volcano seismology aims to provide an understanding of the physical processes in volcanic systems and anticipate the style and timing of eruptions by analyzing the seismic records. Volcanic tremor signals are usually observed in the seismic records before or during volcanic eruptions. Their analysis contributes to evaluate the evolving volcanic activity and potentially predict eruptions. Years of continuous seismic monitoring now provide useful information for operational eruption forecasting. The continuously growing amount of seismic recordings, however, poses a challenge for analysis, information extraction, and interpretation, to support timely decision making during volcanic crises. Furthermore, the complexity of eruption processes and precursory activities makes the analysis challenging.
A challenge in studying seismic signals of volcanic origin is the coexistence of transient signal swarms and long-lasting volcanic tremor signals. Separating transient events from volcanic tremors can, therefore, contribute to improving our understanding of the underlying physical processes. Some similar issues (data reduction, source separation, extraction, and classification) are addressed in the context of music information retrieval (MIR). The signal characteristics of acoustic and seismic recordings comprise a number of similarities. This thesis is going beyond classical signal analysis techniques usually employed in seismology by exploiting similarities of seismic and acoustic signals and building the information retrieval strategy on the expertise developed in the field of MIR.
First, inspired by the idea of harmonic–percussive separation (HPS) in musical signal processing, I have developed a method to extract harmonic volcanic tremor signals and to detect transient events from seismic recordings. This provides a clean tremor signal suitable for tremor investigation along with a characteristic function suitable for earthquake detection. Second, using HPS algorithms, I have developed a noise reduction technique for seismic signals. This method is especially useful for denoising ocean bottom seismometers, which are highly contaminated by noise. The advantage of this method compared to other denoising techniques is that it doesn’t introduce distortion to the broadband earthquake waveforms, which makes it reliable for different applications in passive seismological analysis. Third, to address the challenge of extracting information from high-dimensional data and investigating the complex eruptive phases, I have developed an advanced machine learning model that results in a comprehensive signal processing scheme for volcanic tremors. Using this method seismic signatures of major eruptive phases can be automatically detected. This helps to provide a chronology of the volcanic system. Also, this model is capable to detect weak precursory volcanic tremors prior to the eruption, which could be used as an indicator of imminent eruptive activity. The extracted patterns of seismicity and their temporal variations finally provide an explanation for the transition mechanism between eruptive phases.
In the last century, several astronomical measurements have supported that a significant percentage (about 22%) of the total mass of the Universe, on galactic and extragalactic scales, is composed of a mysterious ”dark” matter (DM). DM does not interact with the electromagnetic force; in other words it does not reflect, absorb or emit light. It is possible that DM particles are weakly interacting massive particles (WIMPs) that can annihilate (or decay) into Standard Model (SM) particles, and modern very- high-energy (VHE; > 100 GeV) instruments such as imaging atmospheric Cherenkov telescopes (IACTs) can play an important role in constraining the main properties of such DM particles, by detecting these products. One of the most privileged targets where to look for DM signal are dwarf spheroidal galaxies (dSphs), as they are expected to be high DM-dominated objects with a clean, gas-free environment. Some dSphs could be considered as extended sources, considering the angular resolution of IACTs; their angu- lar resolution is adequate to detect extended emission from dSphs. For this reason, we performed an extended-source analysis, by taking into account in the unbinned maximum likelihood estimation both the energy and the angular extension dependency of observed events. The goal was to set more constrained upper limits on the velocity-averaged cross-section annihilation of WIMPs with VERITAS data. VERITAS is an array of four IACTs, able to detect γ-ray photons ranging between 100 GeV and 30 TeV. The results of this extended analysis were compared against the traditional spectral analysis. We found that a 2D analysis may lead to more constrained results, depending on the DM mass, channel, and source. Moreover, in this thesis, the results of a multi-instrument project are presented too. Its goal was to combine already published 20 dSphs data from five different experiments, such as Fermi-LAT, MAGIC, H.E.S.S., VERITAS and HAWC, in order to set upper limits on the WIMP annihilation cross-section in the widest mass range ever reported.
Mit dem Alter kann eine Zunahme leichtgradiger Entzündungsprozesse beobachtet werden, von denen angenommen wird, dass sie den typischen, altersbedingten Verlust an Muskelmasse, -kraft und -funktion „befeuern“. Diese als Inflammaging bezeichneten Prozesse können auf ein komplexes Zusammenspiel aus einem dysfunktionalen (viszeralen) Fettgewebe, einer Dysbiose und damit einhergehender mikrobiellen Translokation und geringeren Abwehrfähigkeit sowie einer insgesamt zunehmenden Immunseneszenz zurückgeführt werden. In Summa begünstigt ein pro-inflammatorisches Milieu metabolische Störungen und chronische, altersassoziierte Erkrankungen, die das Entzündungsgeschehen aufrechterhalten oder vorantreiben. Neben einem essenziellen Bewegungsmangel trägt auch eine westlich geprägte, industrialisierte Ernährungsweise zum Entzündungsgeschehen und zur Entwicklung chronischer Erkrankungen bei. Daher liegt die Vermutung nahe, dem Entzündungsgeschehen mit ausreichend Bewegung und einer anti-inflammatorischen Ernährung entgegenzuwirken. In dieser Hinsicht werden insbesondere Omega-3-Fettsäuren (Omega-3) mit anti-inflammatorischen Eigenschaften verbunden. Obwohl ein Zusammenhang zwischen dem ernährungsbedingten Inflammationspotenzial bzw. der Zufuhr von Omega-3 und dem Inflammationsprofil bereits untersucht wurde, fehlen bislang Untersuchungen insbesondere bei älteren Erwachsenen, die den Link zwischen dem Inflammationspotenzial der Ernährung und Sarkopenie-relevanten Muskelparametern herstellen.
Aufgrund des Proteinmehrbedarfs zum Erhalt der funktionellen Muskulatur im Alter wurde bereits eine Vielzahl an Sport- und Ernährungsinterventionen durchgeführt, die eine Verbesserung des Muskelstatus mit Hilfe von strukturiertem Krafttraining und einer proteinreichen Ernährung zeigen. Es gibt zudem Hinweise, dass Omega-3 auch die Proteinsynthese verstärken könnten. Unklar ist jedoch, inwiefern eine anti-inflammatorische Ernährung mit Fokus auf Omega-3 sowohl die Entzündungsprozesse als auch den Muskelproteinmetabolismus und die neuromuskuläre Funktionalität im Alter günstig unterstützen kann. Dies vor allem im Hinblick auf die Muskelleistung, die eng mit der Sturzneigung und der Autonomie im Alltag verknüpft ist, aber in Interventionsstudien mit älteren Erwachsenen bisher wenig Berücksichtigung erhielt. Darüber hinaus werden häufig progressive Trainingselemente genutzt, die nach Studienabschluss oftmals wenig Anschluss im Lebensalltag der Betroffenen finden und somit wenig nachhaltig sind. Ziel dieser Arbeit war demnach die Evaluierung einer proteinreichen und zusätzlich mit Omega-3 supplementierten Ernährung in Kombination mit einem wöchentlichen Vibrationstraining und altersgemäßen Bewegungsprogramm auf Inflammation und neuromuskuläre Funktion bei älteren, selbständig lebenden Erwachsenen.
Hierzu wurden zunächst mögliche Zusammenhänge zwischen dem ernährungsbedingten Inflammationspotenzial, ermittelt anhand des Dietary Inflammatory Index, und dem Muskelstatus sowie dem Inflammationsprofil im Alter eruiert. Dazu dienten die Ausgangswerte von älteren, selbständig lebenden Erwachsenen einer postprandialen Interventionsstudie (POST-Studie), die im Querschnitt analysiert wurden. Die Ergebnisse bestätigten, dass eine pro-inflammatorische Ernährung sich einerseits in einem stärkeren Entzündungsgeschehen widerspiegelt und andererseits mit Sarkopenie-relevanten Parametern, wie einer geringeren Muskelmasse und Gehgeschwindigkeit, ungünstig assoziiert ist. Darüber hinaus zeigten sich diese Zusammenhänge auch in Bezug auf die Handgreifkraft bei den inaktiven, älteren Erwachsenen der Studie.
Anschließend wurde in einer explorativ ausgerichteten Pilot-Interventionsstudie (AIDA-Studie) in einem dreiarmigen Design untersucht, inwieweit sich eine Supplementierung mit Omega-3 unter Voraussetzung einer optimierten Proteinzufuhr und altersgemäßen Sportintervention mit Vibrationstraining auf die neuromuskuläre Funktion und Inflammation bei selbständig lebenden, älteren Erwachsenen auswirkt. Nach acht Wochen Intervention zeigte sich, dass eine mit Omega-3 supplementierte, proteinreiche Ernährung die Muskelleistung insbesondere bei den älteren Männern steigerte. Während sich die Kontrollgruppe nach acht Wochen Sportintervention nicht verbesserte, bestätigte sich zusätzlich eine Verbesserung der Beinkraft und der Testzeit beim Stuhl-Aufsteh-Test der älteren Erwachsenen mit einer proteinreichen Ernährung in Kombination mit der Sportintervention.
Darüber hinaus wurde deutlich, dass die zusätzliche Omega-3-Supplementierung insbesondere bei den Männern eine Reduktion der pro-inflammatorischen Zytokine im Serum zur Folge hatte. Allerdings spiegelten sich diese Beobachtungen nicht auf Genexpressionsebene in mononukleären Immunzellen oder in der LPS-induzierten Sekretion der Zytokine und Chemokine in Vollblutzellkulturen wider. Dies erfordert weitere Untersuchungen.
Most machine learning methods provide only point estimates when being queried to predict on new data. This is problematic when the data is corrupted by noise, e.g. from imperfect measurements, or when the queried data point is very different to the data that the machine learning model has been trained with. Probabilistic modelling in machine learning naturally equips predictions with corresponding uncertainty estimates which allows a practitioner to incorporate information about measurement noise into the modelling process and to know when not to trust the predictions. A well-understood, flexible probabilistic framework is provided by Gaussian processes that are ideal as building blocks of probabilistic models. They lend themself naturally to the problem of regression, i.e., being given a set of inputs and corresponding observations and then predicting likely observations for new unseen inputs, and can also be adapted to many more machine learning tasks. However, exactly inferring the optimal parameters of such a Gaussian process model (in a computationally tractable manner) is only possible for regression tasks in small data regimes. Otherwise, approximate inference methods are needed, the most prominent of which is variational inference.
In this dissertation we study models that are composed of Gaussian processes embedded in other models in order to make those more flexible and/or probabilistic. The first example are deep Gaussian processes which can be thought of as a small network of Gaussian processes and which can be employed for flexible regression. The second model class that we study are Gaussian process state-space models. These can be used for time-series modelling, i.e., the task of being given a stream of data ordered by time and then predicting future observations. For both model classes the state-of-the-art approaches offer a trade-off between expressive models and computational properties (e.g. speed or convergence properties) and mostly employ variational inference. Our goal is to improve inference in both models by first getting a deep understanding of the existing methods and then, based on this, to design better inference methods. We achieve this by either exploring the existing trade-offs or by providing general improvements applicable to multiple methods.
We first provide an extensive background, introducing Gaussian processes and their sparse (approximate and efficient) variants. We continue with a description of the models under consideration in this thesis, deep Gaussian processes and Gaussian process state-space models, including detailed derivations and a theoretical comparison of existing methods.
Then we start analysing deep Gaussian processes more closely: Trading off the properties (good optimisation versus expressivity) of state-of-the-art methods in this field, we propose a new variational inference based approach. We then demonstrate experimentally that our new algorithm leads to better calibrated uncertainty estimates than existing methods.
Next, we turn our attention to Gaussian process state-space models, where we closely analyse the theoretical properties of existing methods.The understanding gained in this process leads us to propose a new inference scheme for general Gaussian process state-space models that incorporates effects on multiple time scales. This method is more efficient than previous approaches for long timeseries and outperforms its comparison partners on data sets in which effects on multiple time scales (fast and slowly varying dynamics) are present.
Finally, we propose a new inference approach for Gaussian process state-space models that trades off the properties of state-of-the-art methods in this field. By combining variational inference with another approximate inference method, the Laplace approximation, we design an efficient algorithm that outperforms its comparison partners since it achieves better calibrated uncertainties.
This thesis explores the variation in coreference patterns across language modes (i.e., spoken and written) and text genres. The significance of research on variation in language use has been emphasized in a number of linguistic studies. For instance, Biber and Conrad [2009] state that “register/genre variation is a fundamental aspect of human language” and “Given the ubiquity of register/genre variation, an understanding of how linguistic features are used in patterned ways across text varieties is of central importance for both the description of particular languages and the development of cross-linguistic theories of language use.”[p.23]
We examine the variation across genres with the primary goal of contributing to the body of knowledge on the description of language use in English. On the computational side, we believe that incorporating linguistic knowledge into learning-based systems can boost the performance of automatic natural language processing systems, particularly for non-standard texts. Therefore, in addition to their descriptive value, the linguistic findings we provide in this study may prove to be helpful for improving the performance of automatic coreference resolution, which is essential for a good text understanding and beneficial for several downstream NLP applications, including machine translation and text summarization.
In particular, we study a genre of texts that is formed of conversational interactions on the well-known social media platform Twitter. Two factors motivate us: First, Twitter conversations are realized in written form but resemble spoken communication [Scheffler, 2017], and therefore they form an atypical genre for the written mode. Second, while Twitter texts are a complicated genre for automatic coreference resolution, due to their widespread use in the digital sphere, at the same time they are highly relevant for applications that seek to extract information or sentiments from users’ messages. Thus, we are interested in discovering more about the linguistic and computational aspects of coreference in Twitter conversations. We first created a corpus of such conversations for this purpose and annotated it for coreference. We are interested in not only the coreference patterns but the overall discourse behavior of Twitter conversations. To address this, in addition to the coreference relations, we also annotated the coherence relations on the corpus we compiled. The corpus is available online in a newly developed form that allows for separating the tweets from their annotations.
This study consists of three empirical analyses where we independently apply corpus-based, psycholinguistic and computational approaches for the investigation of variation in coreference patterns in a complementary manner. (1) We first make a descriptive analysis of variation across genres through a corpus-based study. We investigate the linguistic aspects of nominal coreference in Twitter conversations and we determine how this genre relates to other text genres in spoken and written modes. In addition to the variation across genres, studying the differences in spoken-written modes is also in focus of linguistic research since from Woolbert [1922]. (2) In order to investigate whether the language mode alone has any effect on coreference patterns, we carry out a crowdsourced experiment and analyze the patterns in the same genre for both spoken and written modes. (3) Finally, we explore the potentials of domain adaptation of automatic coreference resolution (ACR) for the conversational Twitter data. In order to answer the question of how the genre of Twitter conversations relates to other genres in spoken and written modes with respect to coreference patterns, we employ a state-of-the-art neural ACR model [Lee et al., 2018] to examine whether ACR on Twitter conversations will benefit from mode-based separation in out-of-domain training data.
Individuals with aphasia vary in the speed and accuracy they perform sentence comprehension tasks. Previous results indicate that the performance patterns of individuals with aphasia vary between tasks (e.g., Caplan, DeDe, & Michaud, 2006; Caplan, Michaud, & Hufford, 2013a). Similarly, it has been found that the comprehension performance of individuals with aphasia varies between homogeneous test sentences within and between sessions (e.g., McNeil, Hageman, & Matthews, 2005). These studies ascribed the variability in the performance of individuals with aphasia to random noise. This conclusion would be in line with an influential theory on sentence comprehension in aphasia, the resource reduction hypothesis (Caplan, 2012). However, previous studies did not directly compare variability in language-impaired and language-unimpaired adults. Thus, it is still unclear how the variability in sentence comprehension differs between individuals with and without aphasia. Furthermore, the previous studies were exclusively carried out in English. Therefore, the findings on variability in sentence processing in English still need to be replicated in a different language.
This dissertation aims to give a systematic overview of the patterns of variability in sentence comprehension performance in aphasia in German and, based on this overview, to put the resource reduction hypothesis to the test. In order to reach the first aim, variability was considered on three different dimensions (persons, measures, and occasions) following the classification by Hultsch, Strauss, Hunter, and MacDonald (2011). At the dimension of persons, the thesis compared the performance of individuals with aphasia and language-unimpaired adults. At the dimension of measures, this work explored the performance across different sentence comprehension tasks (object manipulation, sentence-picture matching). Finally, at the dimension of occasions, this work compared the performance in each task between two test sessions. Several methods were combined to study variability to gain a large and diverse database. In addition to the offline comprehension tasks, the self-paced-listening paradigm and the visual world eye-tracking paradigm were used in this work.
The findings are in line with the previous results. As in the previous studies, variability in sentence comprehension in individuals with aphasia emerged between test sessions and between tasks. Additionally, it was possible to characterize the variability further using hierarchical Bayesian models. For individuals with aphasia, it was shown that both between-task and between-session variability are unsystematic. In contrast to that, language-unimpaired individuals exhibited systematic differences between measures and between sessions. However, these systematic differences occurred only in the offline tasks. Hence, variability in sentence comprehension differed between language-impaired and language-unimpaired adults, and this difference could be narrowed down to the offline measures.
Based on this overview of the patterns of variability, the resource reduction hypothesis was evaluated. According to the hypothesis, the variability in the performance of individuals with aphasia can be ascribed to random fluctuations in the resources available for sentence processing. Given that the performance of the individuals with aphasia varied unsystematically, the results support the resource reduction hypothesis. Furthermore, the thesis proposes that the differences in variability between language-impaired and language-unimpaired adults can also be explained by the resource reduction hypothesis. More specifically, it is suggested that the systematic changes in the performance of language-unimpaired adults are due to decreasing fluctuations in available processing resources. In parallel, the unsystematic variability in the performance of individuals with aphasia could be due to constant fluctuations in available processing resources. In conclusion, the systematic investigation of variability contributes to a better understanding of language processing in aphasia and thus enriches aphasia research.
Air pollution has been a persistent global problem in the past several hundred years. While some industrialized nations have shown improvements in their air quality through stricter regulation, others have experienced declines as they rapidly industrialize. The WHO’s 2021 update of their recommended air pollution limit values reflects the substantial impacts on human health of pollutants such as NO2 and O3, as recent epidemiological evidence suggests substantial long-term health impacts of air pollution even at low concentrations. Alongside developments in our understanding of air pollution's health impacts, the new technology of low-cost sensors (LCS) has been taken up by both academia and industry as a new method for measuring air pollution. Due primarily to their lower cost and smaller size, they can be used in a variety of different applications, including in the development of higher resolution measurement networks, in source identification, and in measurements of air pollution exposure. While significant efforts have been made to accurately calibrate LCS with reference instrumentation and various statistical models, accuracy and precision remain limited by variable sensor sensitivity. Furthermore, standard procedures for calibration still do not exist and most proprietary calibration algorithms are black-box, inaccessible to the public. This work seeks to expand the knowledge base on LCS in several different ways: 1) by developing an open-source calibration methodology; 2) by deploying LCS at high spatial resolution in urban environments to test their capability in measuring microscale changes in urban air pollution; 3) by connecting LCS deployments with the implementation of local mobility policies to provide policy advice on resultant changes in air quality.
In a first step, it was found that LCS can be consistently calibrated with good performance against reference instrumentation using seven general steps: 1) assessing raw data distribution, 2) cleaning data, 3) flagging data, 4) model selection and tuning, 5) model validation, 6) exporting final predictions, and 7) calculating associated uncertainty. By emphasizing the need for consistent reporting of details at each step, most crucially on model selection, validation, and performance, this work pushed forward with the effort towards standardization of calibration methodologies. In addition, with the open-source publication of code and data for the seven-step methodology, advances were made towards reforming the largely black-box nature of LCS calibrations.
With a transparent and reliable calibration methodology established, LCS were then deployed in various street canyons between 2017 and 2020. Using two types of LCS, metal oxide (MOS) and electrochemical (EC), their performance in capturing expected patterns of urban NO2 and O3 pollution was evaluated. Results showed that calibrated concentrations from MOS and EC sensors matched general diurnal patterns in NO2 and O3 pollution measured using reference instruments. While MOS proved to be unreliable for discerning differences among measured locations within the urban environment, the concentrations measured with calibrated EC sensors matched expectations from modelling studies on NO2 and O3 pollution distribution in street canyons. As such, it was concluded that LCS are appropriate for measuring urban air quality, including for assisting urban-scale air pollution model development, and can reveal new insights into air pollution in urban environments.
To achieve the last goal of this work, two measurement campaigns were conducted in connection with the implementation of three mobility policies in Berlin. The first involved the construction of a pop-up bike lane on Kottbusser Damm in response to the COVID-19 pandemic, the second surrounded the temporary implementation of a community space on Böckhstrasse, and the last was focused on the closure of a portion of Friedrichstrasse to all motorized traffic. In all cases, measurements of NO2 were collected before and after the measure was implemented to assess changes in air quality resultant from these policies. Results from the Kottbusser Damm experiment showed that the bike-lane reduced NO2 concentrations that cyclists were exposed to by 22 ± 19%. On Friedrichstrasse, the street closure reduced NO2 concentrations to the level of the urban background without worsening the air quality on side streets. These valuable results were communicated swiftly to partners in the city administration responsible for evaluating the policies’ success and future, highlighting the ability of LCS to provide policy-relevant results.
As a new technology, much is still to be learned about LCS and their value to academic research in the atmospheric sciences. Nevertheless, this work has advanced the state of the art in several ways. First, it contributed a novel open-source calibration methodology that can be used by a LCS end-users for various air pollutants. Second, it strengthened the evidence base on the reliability of LCS for measuring urban air quality, finding through novel deployments in street canyons that LCS can be used at high spatial resolution to understand microscale air pollution dynamics. Last, it is the first of its kind to connect LCS measurements directly with mobility policies to understand their influences on local air quality, resulting in policy-relevant findings valuable for decisionmakers. It serves as an example of the potential for LCS to expand our understanding of air pollution at various scales, as well as their ability to serve as valuable tools in transdisciplinary research.