Refine
Has Fulltext
- yes (136) (remove)
Year of publication
- 2016 (136) (remove)
Document Type
- Doctoral Thesis (136) (remove)
Is part of the Bibliography
- yes (136)
Keywords
- Klimawandel (4)
- climate change (4)
- Blickbewegungen (3)
- earthquake (3)
- Aggression (2)
- Aphasie (2)
- Berlin (2)
- Bodenfeuchte (2)
- Centrosom (2)
- Deutschland (2)
Institute
- Institut für Geowissenschaften (26)
- Institut für Chemie (18)
- Institut für Biochemie und Biologie (13)
- Institut für Physik und Astronomie (12)
- Department Linguistik (8)
- Department Sport- und Gesundheitswissenschaften (7)
- Institut für Ernährungswissenschaft (7)
- Institut für Mathematik (7)
- Wirtschaftswissenschaften (7)
- Department Psychologie (6)
Die Projektierung und Abwicklung sowie die statische und dynamische Analyse von Geschäftsprozessen im Bereich des Verwaltens und Regierens auf kommunaler, Länder- wie auch Bundesebene mit Hilfe von Informations- und Kommunikationstechniken beschäftigen Politiker und Strategen für Informationstechnologie ebenso wie die Öffentlichkeit seit Langem.
Der hieraus entstandene Begriff E-Government wurde in der Folge aus den unterschiedlichsten technischen, politischen und semantischen Blickrichtungen beleuchtet.
Die vorliegende Arbeit konzentriert sich dabei auf zwei Schwerpunktthemen:
• Das erste Schwerpunktthema behandelt den Entwurf eines hierarchischen Architekturmodells, für welches sieben hierarchische Schichten identifiziert werden können. Diese erscheinen notwendig, aber auch hinreichend, um den allgemeinen Fall zu beschreiben.
Den Hintergrund hierfür liefert die langjährige Prozess- und Verwaltungserfahrung als Leiter der EDV-Abteilung der Stadtverwaltung Landshut, eine kreisfreie Stadt mit rund 69.000 Einwohnern im Nordosten von München. Sie steht als Repräsentant für viele Verwaltungsvorgänge in der Bundesrepublik Deutschland und ist dennoch als Analyseobjekt in der Gesamtkomplexität und Prozessquantität überschaubar.
Somit können aus der Analyse sämtlicher Kernabläufe statische und dynamische Strukturen extrahiert und abstrakt modelliert werden.
Die Schwerpunkte liegen in der Darstellung der vorhandenen Bedienabläufe in einer Kommune. Die Transformation der Bedienanforderung in einem hierarchischen System, die Darstellung der Kontroll- und der Operationszustände in allen Schichten wie auch die Strategie der Fehlererkennung und Fehlerbehebung schaffen eine transparente Basis für umfassende Restrukturierungen und Optimierungen.
Für die Modellierung wurde FMC-eCS eingesetzt, eine am Hasso-Plattner-Institut für Softwaresystemtechnik GmbH (HPI) im Fachgebiet Kommunikationssysteme entwickelte Methodik zur Modellierung zustandsdiskreter Systeme unter Berücksichtigung möglicher Inkonsistenzen (Betreuer: Prof. Dr.-Ing. Werner Zorn [ZW07a, ZW07b]).
• Das zweite Schwerpunktthema widmet sich der quantitativen Modellierung und Optimierung von E-Government-Bediensystemen, welche am Beispiel des Bürgerbüros der Stadt Landshut im Zeitraum 2008 bis 2015 durchgeführt wurden. Dies erfolgt auf Basis einer kontinuierlichen Betriebsdatenerfassung mit aufwendiger Vorverarbeitung zur Extrahierung mathematisch beschreibbarer Wahrscheinlichkeitsverteilungen.
Der hieraus entwickelte Dienstplan wurde hinsichtlich der erzielbaren Optimierungen im dauerhaften Echteinsatz verifiziert.
[ZW07a] Zorn, Werner: «FMC-QE A New Approach in Quantitative Modeling», Vortrag anlässlich: MSV'07- The 2007 International Conference on Modeling, Simulation and Visualization Methods WorldComp2007, Las Vegas, 28.6.2007.
[ZW07b] Zorn, Werner: «FMC-QE, A New Approach in Quantitative Modeling», Veröffentlichung, Hasso-Plattner-Institut für Softwaresystemtechnik an der Universität Potsdam, 28.6.2007.
Discourse production is crucial for communicative success and is in the core of aphasia assessment and treatment. Coherence differentiates discourse from a series of utterances/sentences; it is internal unity and connectedness, and, as such, perhaps the most inherent property of discourse. It is unclear whether people with aphasia, who experience various language production difficulties, preserve the ability to produce coherent discourse. A more general question of how coherence is established and represented linguistically has been addressed in the literature, yet remains unanswered. This dissertation presents an investigation of discourse production in aphasia and the linguistic mechanisms of establishing coherence.
Since the 1960ies, Germany has been host to a large Turkish immigrant community. While migrant communities often shift to the majority language over the course of time, Turkish is a very vital minority language in Germany and bilingualism in this community is an obvious fact which has been subject to several studies. The main focus usually is on German, the second language (L2) of these speakers (e.g. Hinnenkamp 2000, Keim 2001, Auer 2003, Cindark & Aslan (2004), Kern & Selting 2006, Selting 2009, Kern 2013). Research on the Turkish spoken by Turkish bilinguals has also attracted attention although to a lesser extend mainly in the framework of so called heritage language research (cf. Polinski 2011). Bilingual Turkish has been investigated under the perspective of code-switching and codemixing (e.g. Kallmeyer & Keim 2003, Keim 2003, 2004, Keim & Cindark 2003, Hinnenkamp 2003, 2005, 2008, Dirim & Auer 2004), and with respect to changes in the morphologic, the syntactic and the orthographic system (e.g. Rehbein & Karakoç 2004, Schroeder 2007). Attention to the changes in the prosodic system of bilingual Turkish on the other side has been exceptional so far (Queen 2001, 2006).
With the present dissertation, I provide a study on contact induced linguistic changes on the prosodic level in the Turkish heritage language of adult early German-Turkish bilinguals. It describes structural changes in the L1 Turkish intonation of yes/no questions of a representative sample of bilingual Turkish speakers. All speakers share a similar sociolinguistic background. All acquired Turkish as their first language from their families and the majority language German as an early L2 at latest in the kinder garden by the age of 3.
A study of changes in bilingual varieties requires a previous cross-linguistic comparison of both of the involved languages in language contact in order to draw conclusions on the contact-induced language change in delimitation to language-internal development.
While German is one of the best investigated languages with respect to its prosodic system, research on Turkish intonational phonology is not as progressed. To this effect, the analysis of bilingual Turkish, as elicited for the present dissertation, is preceded by an experimental study on monolingual Turkish. In this regard an additional experiment with 11 monolingual university students of non-linguistic subjects was conducted at the Ege University in Izmir in 2013. On these grounds the present dissertation additionally contributes new insights with respect to Turkish intonational phonology and typology. The results of the contrastive analysis of German and Turkish bring to light that the prosodic systems of both languages differ with respect to the use of prosodic cues in the marking of information structure (IS) and sentence type. Whereas German distinguishes in the prosodic marking between explicit categories for focus and givenness, Turkish uses only one prosodic cue to mark IS. Furthermore it is shown that Turkish in contrast to German does not use a prosodic correlate to mark yes/no questions, but a morphological question marker.
To elicit Turkish yes/no questions in a bilingual context which differ with respect to their information structure in a further step the methodology of Xu (1999) to elicit in-situ focus on different constituents was adapted in the experimental study. A data set of 400 Turkish yes/no questions of 20 bilingual Turkish speakers was compiled at the Zentrum für Allgemeine Sprachwissenschaft (ZAS) in Berlin and at the University of Potsdam in 2013. The prosodic structure of the yes/no questions was phonologically and phonetically analyzed with respect to changes in the f0 contour according to IS modifications and the use of prosodic cues to indicate sentence type.
The results of the analyses contribute surprising observations to the research of bilingual prosody. Studies on bilingual language change and language acquisition have repeatedly shown that the use of prosodic features that are considered as marked by means of lower and implicational use across and within a language cause difficulties in language contact and second language acquisition. Especially, they are not expected to pass from one language to another through language contact. However, this structurally determined expectation on language development is refuted by the results of the present study. Functionally related prosody, such as the cues to indicate IS, are transferred from German L2 to the Turkish L1 of German-Turkish bilingual speakers. This astonishing observation provides the base for an approach to language change centered on functional motivation. Based on Matras’ (2007, 2010) assumption of functionality in language change, Paradis’ (1993, 2004, 2008) approach of Language Activation and the Subsystem Theory and the Theory of Language as a Dynamic System (Heredina & Jessner 2002), it will be shown that prosodic features which are absent in one of the languages of bilingual speech communities are transferred from the respective language to the other when they contribute to the contextualization of a pragmatic concept which is not expressed by other linguistic means in the target language. To this effect language interaction is based on language activation and inhibition mechanisms dealing with differences in the implicit pragmatic knowledge between bilinguals and monolinguals. The motivator for this process of language change is the contextualization of the message itself and not the structure of the respective feature on the surface. It is shown that structural consideration may influence language change but that bilingual language change does not depend on structural restrictions nor does the structure cause a change. The conclusions drawn on the basis of empirical facts can especially contribute to a better understanding of the processes of bilingual language development as it combines methodologies and theoretical aspects of different linguistic subfields.
This doctoral thesis seeks to elaborate how Wittgenstein’s very sparse writings on ethics and ethical thought, together with his later work on the more general problem of normativity and his approach to philosophical problems as a whole, can be applied to contemporary meta-ethical debates about the nature of moral thought and language and the sources of moral obligation. I begin with a discussion of Wittgenstein’s early “Lecture on Ethics”, distinguishing the thesis of a strict fact/value dichotomy that Wittgenstein defends there from the related thesis that all ethical discourse is essentially and intentionally nonsensical, an attempt to go beyond the limits of sense. The first chapter discusses and defends Wittgenstein’s argument that moral valuation always goes beyond any ascertaining of fact; the second chapter seeks to draw out the valuable insights from Wittgenstein’s (early) insistence that value discourse is nonsensical while also arguing that this thesis is ultimately untenable and also incompatible with later Wittgensteinian understanding of language. On the basis of this discussion I then take up the writings of the American philosopher Cora Diamond, who has worked out an ethical approach in a very closely Wittgensteinian spirit, and show how this approach shares many of the valuable insights of the moral expressivism and constructivism of contemporary authors such as Blackburn and Korsgaard while suggesting a way to avoid some of the problems and limitations of their approaches. Subsequently I turn to a criticism of the attempts by Lovibond and McDowell to enlist Wittgenstein in the support for a non-naturalist moral realism. A concluding chapter treats the ways that a broadly Wittgensteinian conception expands the subject of metaethics itself by questioning the primacy of discursive argument in moral thought and of moral propositions as the basic units of moral belief.
We do magnetohydrodynamic (MHD) simulations of local box models of turbulent Interstellar Medium (ISM) and analyse the process of amplification and saturation of mean magnetic fields with methods of mean field dynamo theory. It is shown that the process of saturation of mean fields can be partially described by the prolonged diffusion time scales in presence of the dynamically significant magnetic fields. However, the outward wind also plays an essential role in the saturation in higher SN rate case. Algebraic expressions for the back reaction of the magnetic field onto the turbulent transport coefficients are derived, which allow a complete description of the nonlinear dynamo. We also present the effects of dynamically significant mean fields on the ISM configuration and pressure distribution. We further add the cosmic ray component in the simulations and investigate the kinematic growth of mean fields with a dynamo perspective.
Faktor Mensch
(2016)
Eine Krisensituation ist eine Umbruchssituation. Sie kann als Chance, als Herausforderung sowie im Sinne einer SCHUMPETER’SCHEN „schöpferischen Zerstörung“ als Ausgangspunkt von bedeutsamen Veränderungen und neuen Entwicklungsmöglichkeiten begriffen werden. Die Krisensituation der 1970er und 1980er Jahre bildet dahingehend durch ihre Einzigartigkeit einen herausragenden wirtschaftshistorischen Untersuchungsgegenstand. Für die westeuropäischen Staaten waren es nicht nur Jahre einer, nach der außergewöhnlich langen Boomzeit der „Wirtschaftswunderjahre“, problematischen sozialen und wirtschaftliche Situation. Es war auch eine Zeit eines bedeutungsvollen und beschleunigten Wandels. Die Gleichzeitigkeit vieler Veränderungen, das zeitliche Zusammentreffen von konjunkturellen und strukturellen Problemlagen sowie die äußerst kritische Situation an den westeuropäischen Arbeitsmärkten bildete eine multiple Krisensituation. Deren Auswirkungen waren weitreichender, als es von den Zeitgenossen erahnt werden konnte. Es gab nicht nur Änderungen im (wirtschafts-)politischen Makrogefüge vieler Volkswirtschaften, auch bedeutende Einflüsse waren auf der Mikroebene feststellbar. Marktorientierte Unternehmen mussten sich auf die neue Situation einstellen und im Rahmen einer Neupositionierung von Betriebsstrategien, organisatorischen Umgestaltungen und einer stärkeren Ressourcenorientierung betriebswirtschaftlich handeln. Das schien letztlich zu einer stärkeren Beachtung und Entwicklung der in den Unternehmen vorhandenen humanen und sozialen Ressourcen zu führen.
Diese Arbeit stellt die Hypothese auf, dass umfassende organisatorische Veränderungen und strategische Neupositionierungen, insbesondere die effektivere Nutzung sowie der intensive Auf- und Ausbau betriebsinterner Personalressourcen Unternehmen maßgeblich halfen, die Krisensituation der 1970er und 1980er Jahre besser und nachhaltiger überwinden zu können. Anders als die bisherige wirtschaftshistorische Forschungsliteratur nimmt diese Dissertation nicht die makroökonomisch Perspektive in den Fokus, sondern untersucht die Hypothese anhand mehrerer Unternehmensfallstudien. Ausgewählt sind drei Großunternehmen der westeuropäischen Elektroindustrie. Diese Arbeit liefert mit dieser Untersuchung einen weiteren Baustein zur wirtschaftshistorischen Annäherung an die 1970er und 1980er Jahre und leistet ebenso einen Beitrag zur Fortschreibung der Firmengeschichte der drei Unternehmen.
Die Elektrobranche fand bisher wirtschaftshistorisch nur wenig Beachtung, dennoch ist sie ein gutes Beispiel für die umfassende Veränderungssituation jener Jahre. Entsprechende Sekundärquellen sind für diesen Zeitraum für die drei Unternehmen kaum vorhanden. Aus diesem Grund bildet eine Vielzahl von Archivalien das Fundament dieser Arbeit. Sie werden als Primärquellen aus den jeweiligen Unternehmensarchiven als Basis der Fallanalyse herangezogen. Mit Hilfe zahlreicher Dokumente des betrieblichen „Alltagsgeschäfts“, wie beispielsweise Daten des betrieblichen Personal- und Rechnungswesens, Protokolle von Sitzungen der Arbeitsnehmervertreter, des Aufsichtsrats oder des Vorstands sowie interne Strategiepapiere und Statistiken, wird nicht nur der anfangs aufgestellten Hypothese nachgegangen, sondern auch mehrerer sich aus ihr ergebende Fragenkomplexe. Im Rahmen derer wird untersucht, wie die Unternehmen – die Beschäftigten, die Leitungsebenen und die Aufsichtsräte – auf die Krisensituation reagierten, ob sie ggf. ihrerseits Einfluss darauf zu nehmen versuchten und welche betriebswirtschaftlichen Schlüsse daraus gezogen wurden. Es wird hinterfragt, ob eine stärkere Ressourcenorientierung wirklich eine neue Strategieperspektive bot und es diesbezüglich zu einer stärkeren Beachtung humaner und sozialer Ressourcen im Unternehmen kam. Diese Arbeit untersucht, ob gezielt zur Krisenüberwindung in diese Ressourcen investiert wurde und diese Investitionen halfen, die Krisensituation erfolgreich zu überstehen und nachhaltig den Unternehmenserfolg zu sichern.
The Milky Way is only one out of billions of galaxies in the universe. However, it is a special galaxy because it allows to explore the main mechanisms involved in its evolution and formation history by unpicking the system star-by-star. Especially, the chemical fingerprints of its stars provide clues and evidence of past events in the Galaxy’s lifetime. These information help not only to decipher the current structure and building blocks of the Milky Way, but to learn more about the general formation process of galaxies.
In the past decade a multitude of stellar spectroscopic Galactic surveys have scanned millions of stars far beyond the rim of the solar neighbourhood. The obtained spectroscopic information provide unprecedented insights to the chemo-dynamics of the Milky Way. In addition analytic models and numerical simulations of the Milky Way provide necessary descriptions and predictions suited for comparison with observations in order to decode the physical properties that underlie the complex system of the Galaxy.
In the thesis various approaches are taken to connect modern theoretical modelling of galaxy formation and evolution with observations from Galactic stellar surveys. With its focus on the chemo-kinematics of the Galactic disk this work aims to determine new observational constraints on the formation of the Milky Way providing also proper comparisons with two different models. These are the population synthesis model TRILEGAL based on analytical distribution functions, which aims to simulate the number and distribution of stars in the Milky Way and its different components, and a hybrid model (MCM) that combines an N-body simulation of a Milky Way like galaxy in the cosmological framework with a semi-analytic chemical evolution model for the Milky Way. The major observational data sets in use come from two surveys, namely the “Radial Velocity Experiment” (RAVE) and the “Sloan Extension for Galactic Understanding and Exploration” (SEGUE).
In the first approach the chemo-kinematic properties of the thin and thick disk of the Galaxy as traced by a selection of about 20000 SEGUE G-dwarf stars are directly compared to the predictions by the MCM model. As a necessary condition for this, SEGUE's selection function and its survey volume are evaluated in detail to correct the spectroscopic observations for their survey specific selection biases. Also, based on a Bayesian method spectro-photometric distances with uncertainties below 15% are computed for the selection of SEGUE G-dwarfs that are studied up to a distance of 3 kpc from the Sun.
For the second approach two synthetic versions of the SEGUE survey are generated based on the above models. The obtained synthetic stellar catalogues are then used to create mock samples best resembling the compiled sample of observed SEGUE G-dwarfs. Generally, mock samples are not only ideal to compare predictions from various models. They also allow validation of the models' quality and improvement as with this work could be especially achieved for TRILEGAL. While TRILEGAL reproduces the statistical properties of the thin and thick disk as seen in the observations, the MCM model has shown to be more suitable in reproducing many chemo-kinematic correlations as revealed by the SEGUE stars. However, evidence has been found that the MCM model may be missing a stellar component with the properties of the thick disk that the observations clearly show. While the SEGUE stars do indicate a thin-thick dichotomy of the stellar Galactic disk in agreement with other spectroscopic stellar studies, no sign for a distinct metal-poor disk is seen in the MCM model.
Usually stellar spectroscopic surveys are limited to a certain volume around the Sun covering different regions of the Galaxy’s disk. This often prevents to obtain a global view on the chemo-dynamics of the Galactic disk. Hence, a suitable combination of stellar samples from independent surveys is not only useful for the verification of results but it also helps to complete the picture of the Milky Way. Therefore, the thesis closes with a comparison of the SEGUE G-dwarfs and a sample of RAVE giants. The comparison reveals that the chemo-kinematic relations agree in disk regions where the samples of both surveys show a similar number of stars. For those parts of the survey volumes where one of the surveys lacks statistics they beautifully complement each other. This demonstrates that the comparison of theoretical models on the one side, and the combined observational data gathered by multiple surveys on the other side, are key ingredients to understand and disentangle the structure and formation history of the Milky Way.
Die vorliegende explorative empirische Untersuchung muslimischer Frauengruppen leistet einen Beitrag zur Erforschung des religiösen Wandels im religiösen Feld in Deutschland. Zum einen werden damit erstmals qualitative Daten zu religiösen Gruppen muslimischer Frauen erhoben. Zum anderen liefern die analysierten Anlässe der Gruppengründung und die Gruppenziele Einblicke in die relevanten Themen des religiös-muslimischen Engagements im Zeitverlauf. Gemäß der explorativen Konzeption interessiert sich diese Studie insbesondere für die Vielfalt muslimischer Frauengruppen in Deutschland. Es wurde gefragt, welche Selbstbeschreibungen als muslimische Frauengruppen sich derzeit erkennen lassen? Dazu wurden thematische Leitfadeninterviews mit Ansprechpartnerinnen muslimischer Frauengruppen im religiösen Feld (2006-2011) durchgeführt.
Die Gründungen der zwölf untersuchten muslimischen Frauengruppen lagen im Zeitraum von 1978 bis 2009. Dies umfasst im Hinblick auf das muslimisch-religiöse Engagement Phasen mit unterschiedlichen Schwerpunkten, die von der Verfestigung der religiösen Strukturen über den Kampf um rechtliche Gleichstellung als religiöse Minderheit bis zu einer Auseinandersetzung mit der staatlichen Islampolitik reichen. Die ältesten der untersuchten Gruppen reflektieren in ihrem historischen Verlauf den Aufbau der religiösen Strukturen, indem sie zunächst Räume für sich in den Gemeinden schufen. Diese füllten sie in Form von religiösen Bildungsprozessen und zwar indem sie einander an ihren Kenntnissen teilhaben ließen und gemeinsam die religiöse Quelle erschlossen. Andere Gruppen schlossen sich zusammen, um von den Kenntnissen religiöser Expertinnen zu profitieren, wieder andere etablierten Angebote, die über den eigenen Gruppenzusammenhang hinausreichten. Mit dieser Ausrichtung der Weltgestaltung ging auch die Gründung einer Organisation, d.h. ein Wandel der Sozialform einher.
Die Ergebnisse konstatieren eine Kontingenz hinsichtlich der Selbstzuordnung muslimische Frauengruppe. Es handelt sich um historisch spezifische Selbstzuschreibungen, die Ausdruck eines religiösen Wandels im muslimisch-religiösen Feld initiiert von muslimischen Frauen sind. Zentrales Ergebnis ist hier, dass die Gruppen zwar hinsichtlich ihrer Formen heterogen sind, allerdings eine Verbindungslinie in ihren Kernideen als Frauengruppen- und Organisationen besteht. Es zeigt sich durch alle Phasen des muslimisch-religiösen Engagements im religiösen Feld ein hohes Interesse an religiösen Bildungsthemen seitens muslimischer Frauen. Diese sind verbunden mit der Auseinandersetzung mit dem religiösen Geschlechterverhältnis.
Die Aufmerksamkeit, die das religiöse Geschlechterverhältnis im Kontext des Institutionalisierungs- und Partizipationsprozess des Islam im politischen Feld derzeit besitzt, kann einerseits als spezifisch gelten. Andererseits zeigen historische Kontextualisierungen mit der religiösen Frauenbewegung im 19. Jahrhundert, dass auch hier über religiöse Geschlechterbilder Partizipationsfragen verhandelt wurden.
Die Ergebnisse dieser Studie belegen die Relevanz von religiösen Gruppen innerhalb religiöser Wandlungsprozesse. Weiterhin liefern sie neue Erkenntnisse hinsichtlich des Verhältnisses von religiöser Individualisierung und Gruppenbindung: Muslimische Frauen vergemeinschaften sich aus religiösen Bildungszwecken innerhalb von religiösen Gruppen und behandeln dabei Themen ihre weibliche religiöse Identität und die religiöse Lebensführung als Frau betreffend und dies stärkt ihre individuelle religiöse Bindung.
In the current paradigm of cosmology, the formation of large-scale structures is mainly driven by non-radiating dark matter, making up the dominant part of the matter budget of the Universe. Cosmological observations however, rely on the detection of luminous galaxies, which are biased tracers of the underlying dark matter. In this thesis I present cosmological reconstructions of both, the dark matter density field that forms the cosmic web, and cosmic velocities, for which both aspects of my work are delved into, the theoretical formalism and the results of its applications to cosmological simulations and also to a galaxy redshift survey.The foundation of our method is relying on a statistical approach, in which a given galaxy catalogue is interpreted as a biased realization of the underlying dark matter density field. The inference is computationally performed on a mesh grid by sampling from a probability density function, which describes the joint posterior distribution of matter density and the three dimensional velocity field. The statistical background of our method is described in Chapter ”Implementation of argo”, where the introduction in sampling methods is given, paying special attention to Markov Chain Monte-Carlo techniques. In Chapter ”Phase-Space Reconstructions with N-body Simulations”, I introduce and implement a novel biasing scheme to relate the galaxy number density to the underlying dark matter, which I decompose into a deterministic part, described by a non-linear and scale-dependent analytic expression, and a stochastic part, by presenting a negative binomial (NB) likelihood function that models deviations from Poissonity. Both bias components had already been studied theoretically, but were so far never tested in a reconstruction algorithm. I test these new contributions againstN-body simulations to quantify improvements and show that, compared to state-of-the-art methods, the stochastic bias is inevitable at wave numbers of k≥0.15h Mpc^−1 in the power spectrum in order to obtain unbiased results from the reconstructions. In the second part of Chapter ”Phase-Space Reconstructions with N-body Simulations” I describe and validate our approach to infer the three dimensional cosmic velocity field jointly with the dark matter density. I use linear perturbation theory for the large-scale bulk flows and a dispersion term to model virialized galaxy motions, showing that our method is accurately recovering the real-space positions of the redshift-space distorted galaxies. I analyze the results with the isotropic and also the two-dimensional power spectrum.Finally, in Chapter ”Phase-space Reconstructions with Galaxy Redshift Surveys”, I show how I combine all findings and results and apply the method to the CMASS (for Constant (stellar) Mass) galaxy catalogue of the Baryon Oscillation Spectroscopic Survey (BOSS). I describe how our method is accounting for the observational selection effects inside our reconstruction algorithm. Also, I demonstrate that the renormalization of the prior distribution function is mandatory to account for higher order contributions in the structure formation model, and finally a redshift-dependent bias factor is theoretically motivated and implemented into our method. The various refinements yield unbiased results of the dark matter until scales of k≤0.2 h Mpc^−1in the power spectrum and isotropize the galaxy catalogue down to distances of r∼20h^−1 Mpc in the correlation function. We further test the results of our cosmic velocity field reconstruction by comparing them to a synthetic mock galaxy catalogue, finding a strong correlation between the mock and the reconstructed velocities. The applications of both, the density field without redshift-space distortions, and the velocity reconstructions, are very broad and can be used for improved analyses of the baryonic acoustic oscillations, environmental studies of the cosmic web, the kinematic Sunyaev-Zel’dovic or integrated Sachs-Wolfe effect.
In the past decades, development cooperation (DC) led by conventional bi- and multilateral donors has been joined by a large number of small, private or public-private donors. This pluralism of actors raises questions as to whether or not these new donors are able to implement projects more or less effectively than their conventional counterparts. In contrast to their predecessors, the new donors have committed themselves to be more pragmatic, innovative and flexible in their development cooperation measures. However, they are also criticized for weakening the function of local civil society and have the reputation of being an intransparent and often controversial alternative to public services. With additional financial resources and their new approach to development, the new donors have been described in the literature as playing a controversial role in transforming development cooperation. This dissertation compares the effectiveness of initiatives by new and conventional donors with regard to the provision of public goods and services to the poor in the water and sanitation sector in India.
India is an emerging country but it is experiencing high poverty rates and poor water supply in predominantly rural areas. It lends itself for analyzing this research theme as it is currently being confronted by a large number of actors and approaches that aim to find solutions for these challenges .
In the theoretical framework of this dissertation, four governance configurations are derived from the interaction of varying actor types with regard to hierarchical and non-hierarchical steering of their interactions. These four governance configurations differ in decision-making responsibilities, accountability and delegation of tasks or direction of information flow. The assumption on actor relationships and steering is supplemented by possible alternative explanations in the empirical investigation, such as resource availability, the inheritance of structures and institutions from previous projects in a project context, gaining acceptance through beneficiaries (local legitimacy) as a door opener, and asymmetries of power in the project context.
Case study evidence from seven projects reveals that the actors' relationship is important for successful project delivery. Additionally, the results show that there is a systematic difference between conventional and new donors. Projects led by conventional donors were consistently more successful, due to an actor relationship that placed the responsibility in the hands of the recipient actors and benefited from the trust and reputation of a long-term cooperation. The trust and reputation of conventional donors always went along with a back-up from federal level and trickled down as reputation also at local level implementation. Furthermore, charismatic leaders, as well as the acquired structures and institutions of predecessor projects, also proved to be a positive influencing factor for successful project implementation.
Despite the mixed results of the seven case studies, central recommendations for action can be derived for the various actors involved in development cooperation. For example, new donors could fulfill a supplementary function with conventional donors by developing innovative project approaches through pilot studies and then implementing them as a supplement to the projects of conventional donors on the ground. In return, conventional donors would have to make room the new donors by integrating their approaches into already programs in order to promote donor harmonization. It is also important to identify and occupy niches for activities and to promote harmonization among donors on state and federal sides.
The empirical results demonstrate the need for a harmonization strategy of different donor types in order to prevent duplication, over-experimentation and the failure of development programs. A transformation to successful and sustainable development cooperation can only be achieved through more coordination processes and national self-responsibility.
In der vorliegenden Arbeit wird die planetare Grenzschicht in Ny-Ålesund, Spitzbergen, sowohl bezüglich kleinskaliger („mikrometeorologischer“) Effekte als auch in ihrer Kopplung mit der Synoptik untersucht. Dazu werden verschiedene Beobachtungsdaten aus der Säule und in Bodennähe zusammengezogen und bewertet. Die so gewonnenen Datensätze werden dann zur Validierung eines nicht-hydrostatischen, regionalen Klimamodells genutzt. Weiterhin werden orographisch bedingte Einflüsse, die Untergrundbeschaffenheit und die lokale Heterogenität der Unterlage untersucht. Hierzu werden meteorologische Größen, wie die Variabilität der Temperatur und insbesondere die jährliche Windverteilung in Bodennähe untersucht und es erfolgt ein Vergleich von in-situ gemessenen turbulenten Flüssen von den Eddy-Kovarianz-Messkomplexen bei Ny-Ålesund und im Bayelva-Tal unter demselben Aspekt. Es zeigt sich, dass der Eddy-Kovarianz-Messkomplex im Bayelva-Tal sehr stark durch eine orographisch bedingte Kanalisierung der Strömung beeinflusst ist und sich nicht für Vergleiche mit regionalen Klimamodellen mit horizontalen Auflösungen von <1km eignet. Die hohe Bodenfeuchte im Bayelva-Tal führt zudem zu einem deutlich kleineren Bowen-Verhältnis, als es für diese Region zu erwarten ist. Der Eddy-Kovarianz-Messkomplex bei Ny-Ålesund erweist sich hingegen als geeigneter für solche Modellvergleiche, aufgrund der typischen, küstennahen Windverteilung und des repräsentativen Footprints. Letzteres wird durch die Bestimmung der Footprint-Klimatologie des Jahres 2013 mit einem aktuellen Footprint-Modell erarbeitet.
Weiterhin wird die Auswirkung von (Anti-) Zyklonen über den Archipel auf die zeitliche Variabilität der lokalen Grenzschichteigenschaften untersucht und bewertet. Dazu wird ein Zyklonen-Detektions-Algorithmus auf ERA-Interim-Reanalysedatensätze angewendet, wodurch die Häufigkeit von nahezu ideal konzentrischen Hoch- und die Tiefdruckgebieten für drei Jahre bestimmt wird. Aus dieser Verteilung werden insgesamt drei interessante Zeiträume zu verschiedenen Jahreszeiten ausgewählt und im Rahmen von Prozessstudien die lokalen bodennahen meteorologischen Messungen, der turbulente Austausch an der Oberfläche und die Grenzschichtdynamik in der Säule untersucht. Die zeitliche Variabilität der dynamischen Grenzschichtstabilität in der Säule wird anhand von zeitlich hoch aufgelösten vertikalen Profilen der Bulk-Richardson-Zahl aus Kompositprofilen aus Fernerkundungsinstrumenten (Radiometer, Wind-LIDAR) sowie Mastdaten (BSRN-Mast) untersucht und die Grenzschichthöhe ermittelt. Aus diesen Analysen ergibt sich eine deutliche Abhängigkeit der thermischen Stabilität beim Durchzug von Fronten, eine damit einhergehende erhebliche Abhängigkeit der Grenzschichtdynamik und der Grenzschichthöhe sowie des turbulenten Austauschs von der zeitlichen Variabilität der Windgeschwindigkeit in der Säule.
Auf Grundlage der Standortanalysen und Prozessstudien erfolgt ein Vergleich der bodennahen Messungen und den Beobachtungen aus der Säule, sowohl von den genannten Fernerkundungsinstrumenten als auch von In-situ-Messungen (Radiosonden) für den Zeitraum einer Radiosondierungskampagne mit dem nicht-hydrostatischen, regionalen Klimamodel WRF (ARW). Auf Grundlage der Fragestellung, inwieweit aktuelle Schemata die Grenzschichtcharakteristika in orographisch stark gegliedertem Gelände in der Arktis reproduzieren können, werden zwei Grenzschichtparametrisierungsschemata mit verschiedenen Ordnungen der Schließung validiert. Hierzu wird die zeitliche Variabilität der Temperatur, der Feuchte und des Windfeldes in der Säule bis 2000m in den Simulationen mit den Beobachtungsdaten vergleichen. Es wird gezeigt, dass durch Modifikation der Initialwertfelder eine sehr gute Übereinstimmung zwischen den Simulationen und den Beobachtungen bereits bei einer horizontalen Auflösung von 1km erreicht werden kann und die Wahl des Grenzschichtschemas nur untergeordneten Einfluss hat. Hieraus werden Ansätze der Weiterentwicklung der Parametrisierungen, aber auch Empfehlungen bezüglich der Initialwertfelder, wie der Landmaske und der Orographie, vorgeschlagen.
Background: Low back pain (LBP) is one of the world wide leading causes of limited activity and disability. Impaired motor control has been found to be one of the possible factors related to the development or persistence of LBP. In particularly, motor control strategies seemed to be altered in situations requiring reactive responses of the trunk counteracting sudden external forces. However, muscular responses were mostly assessed in (quasi) static testing situations under simplified laboratory conditions. Comprehensive investigations in motor control strategies during dynamic everyday situations are lacking. The present research project aimed to investigate muscular compensation strategies following unexpected gait perturbations in people with and without LBP. A novel treadmill stumbling protocol was tested for its validity and reliability to provoke muscular reflex responses at the trunk and the lower extremities (study 1). Thereafter, motor control strategies in response to sudden perturbations were compared between people with LBP and asymptomatic controls (CTRL) (study 2). In accordance with more recent concepts of motor adaptation to pain, it was hypothesized that pain may have profound consequences on motor control strategies in LBP. Therefore, it was investigated whether differences in compensation strategies were either consisting of changes local to the painful area at the trunk, or also being present in remote areas such as at the lower extremities.
Methods: All investigations were performed on a custom build split-belt treadmill simulating trip-like events by unexpected rapid deceleration impulses (amplitude: 2 m/s; duration: 100 ms; 200 ms after heel contact) at 1m/s baseline velocity. A total number of 5 (study 1) and 15 (study 2) right sided perturbations were applied during walking trials. Muscular activities were assessed by surface electromyography (EMG), recorded at 12 trunk muscles and 10 (study 1) respectively 5 (study 2) leg muscles. EMG latencies of muscle onset [ms] were retrieved by a semi-automatic detection method. EMG amplitudes (root mean square (RMS)) were assessed within 200 ms post perturbation, normalized to full strides prior to any perturbation [RMS%]. Latency and amplitude investigations were performed for each muscle individually, as well as for pooled data of muscles grouped by location. Characteristic pain intensity scores (CPIS; 0-100 points, von Korff) based on mean intensity ratings reported for current, worst and average pain over the last three months were used to allocate participants into LBP (≥30 points) or CTRL (≤10 points). Test-retest reproducibility between measurements was determined by a compilation of measures of reliability. Differences in muscular activities between LBP and CTRL were analysed descriptively for individual muscles; differences based on grouped muscles were statistically tested by using a multivariate analysis of variance (MANOVA, α =0.05).
Results: Thirteen individuals were included into the analysis of study 1. EMG latencies revealed reflex muscle activities following the perturbation (mean: 89 ms). Respective EMG amplitudes were on average 5-fold of those assessed in unperturbed strides, though being characterized by a high inter-subject variability. Test-retest reliability of muscle latencies showed a high reproducibility, both for muscles at the trunk and legs. In contrast, reproducibility of amplitudes was only weak to moderate for individual muscles, but increased when being assessed as a location specific outcome summary of grouped muscles. Seventy-six individuals were eligible for data analysis in study 2. Group allocation according to CPIS resulted in n=25 for LBP and n=29 for CTRL. Descriptive analysis of activity onsets revealed longer delays for all muscles within LBP compared to CTRL (trunk muscles: mean 10 ms; leg muscles: mean 3 ms). Onset latencies of grouped muscles revealed statistically significant differences between LBP and CTRL for right (p=0.009) and left (p=0.007) abdominal muscle groups. EMG amplitude analysis showed a high variability in activation levels between individuals, independent of group assignment or location. Statistical testing of grouped muscles indicated no significant difference in amplitudes between LBP and CTRL.
Discussion: The present research project could show that perturbed treadmill walking is suitable to provoke comprehensive reflex responses at the trunk and lower extremities, both in terms of sudden onsets and amplitudes of reflex activity. Moreover, it could demonstrate that sudden loadings under dynamic conditions provoke an altered reflex timing of muscles surrounding the trunk in people with LBP compared to CTRL. In line with previous investigations, compensation strategies seemed to be deployed in a task specific manner, with differences between LBP and CTRL being evident predominately at ventral sides. No muscular alterations exceeding the trunk could be found when being assessed under the automated task of locomotion. While rehabilitation programs tailored towards LBP are still under debate, it is tempting to urge the implementation of dynamic sudden loading incidents of the trunk to enhance motor control and thereby to improve spinal protection. Moreover, in respect to the consistently observed task specificity of muscular compensation strategies, such a rehabilitation program should be rich in variety.
Prevalence of Achilles tendinopathy increases with age, leading to a weaker tendon with predisposition to rupture. Previous studies, investigating Achilles tendon (AT) properties, are restricted to standardized isometric conditions. Knowledge regarding the influence of age and pa-thology on AT response under functional tasks remains limited. Therefore, the aim of the thesis was to investigate the influence of age and pathology on AT properties during a single-leg vertical jump.
Healthy children, asymptomatic adults and patients with Achilles tendinopathy participated. Ultrasonography was used to assess AT-length, AT-cross-sectional area and AT-elongation. The reliability of the methodology used was evaluated both Intra- and inter-rater at rest and at maximal isometric plantar-flexion contraction and was further implemented to investigate tendon properties during functional task. During the functional task a single-leg vertical jump on a force plate was performed while simultaneously AT elongation and vertical ground reaction forces were recorded. AT compliance [mm/N] (elongation/force) and AT strain [%] (elongation/length) were calculated. Differences between groups were evaluated with respect to age (children vs. adults) and pathology (asymptomatic adults vs. patients).
Good to excellent reliability with low levels of variability was achieved in the assessment of AT properties. During the jumps AT elongation was found to be statistical significant higher in children. However, no statistical significant difference was found for force among the groups. AT compliance and strain were found to be statistical significant higher only in children. No significant differences were found between asymptomatic adults and patients with tendinopathy.
The methodology used to assess AT properties is reliable, allowing its implementation into further investigations. Higher AT-compliance in children might be considered as a protective factor against load-related injuries. During functional task, when higher forces are acting on the AT, tendinopathy does not result in a weaker tendon.
Difficulties with object relative clauses (ORC), as compared to subject relative clauses (SR), are widely attested across different languages, both in adults and in children. This SR-ORC asymmetry is reduced, or even eliminated, when the embedded constituent in the ORC is a pronoun, rather than a lexical noun phrase. The studies included in this thesis were designed to explore under what circumstances the pronoun facilitation occurs; whether all pronouns have the same effect; whether SRs are also affected by embedded pronouns; whether children perform like adults on such structures; and whether performance is related to cognitive abilities such as memory or grammatical knowledge. Several theoretical approaches that explain the pronoun facilitation in relative clauses are evaluated. The experimental data have been collected in three languages–German, Italian and Hebrew–stemming from both children and adults.
In the German study (Chapter 2), ORCs with embedded 1st- or 3rd-person pronouns are compared to ORCs with an embedded lexical noun phrase. Eye-movement data from 5-year-old children show that the 1st-person pronoun facilitates processing, but not the 3rd-person pronoun. Moreover, children’s performance is modulated by additive effects of their memory and grammatical skills. In the Italian study (Chapter 3), the 1st-person pronoun advantage over the 3rd-person pronoun is tested in ORCs and SRs that display a similar word order. Eye-movement data from 5-year-olds and adult controls and reading times data from adults are pitted against the outcome of a corpus analysis, showing that the 1st-/3rd-person pronoun asymmetry emerges in the two relative clause types to an equal extent. In the Hebrew study (Chapter 4), the goal is to test the effect of a special kind of pronoun–a non-referential arbitrary subject pronoun–on ORC comprehension, in the light of potential confounds in previous studies that used this pronoun. Data from a referent-identification task with 4- to 5-year-olds indicate that, when the experimental material is controlled, the non-referential pronoun does not necessarily facilitate ORC comprehension. Importantly, however, children have even more difficulties when the embedded constituent is a referential pronoun. The non-referentiality / referentiality asymmetry is emphasized by the relation between children’s performance on the experimental task and their memory skills.
Together, the data presented in this thesis indicate that sentence processing is not only driven by structural (or syntactic) factors, but also by discourse-related ones, like pronouns’ referential properties or their discourse accessibility mechanism, which is defined as the level of ease or difficulty with which referents of pronouns are identified and retrieved from the discourse model. Although independent in essence, these structural and discourse factors can in some cases interact in a way that affects sentence processing. Moreover, both types of factors appear to be strongly related to memory. The data also support the idea that, from early on, children are sensitive to the same factors that affect adults’ sentence processing, and that the processing strategies of both populations are qualitatively similar.
In sum, this thesis suggests that a comprehensive theory of human sentence processing needs to account for effects that are due to both structural and discourse-related factors, which operate as a function of memory capacity.
Intracontinental deformation usually is a result of tectonic forces associated with distant plate collisions. In general, the evolution of mountain ranges and basins in this environment is strongly controlled by the distribution and geometries of preexisting structures. Thus, predictive models usually fail in forecasting the deformation evolution in these kinds of settings. Detailed information on each range and basin-fill is vital to comprehend the evolution of intracontinental mountain belts and basins. In this dissertation, I have investigated the complex Cenozoic tectonic evolution of the western Tien Shan in Central Asia, which is one of the most active intracontinental ranges in the world. The work presented here combines a broad array of datasets, including thermo- and geochronology, paleoenvironmental interpretations, sediment provenance and subsurface interpretations in order to track changes in tectonic deformation. Most of the identified changes are connected and can be related to regional-scale processes that governed the evolution of the western Tien Shan.
The NW-SE trending Talas-Fergana fault (TFF) separates the western from the central Tien Shan and constitutes a world-class example of the influence of preexisting anisotropies on the subsequent structural development of a contractile orogen. While to the east most of ranges and basins have a sub-parallel E-W trend, the triangular-shaped Fergana basin forms a substantial feature in the western Tien Shan morphology with ranges on all three sides. In this thesis, I present 55 new thermochronologic ages (apatite fission track and zircon (U-Th)/He)) used to constrain exhumation histories of several mountain ranges in the western Tien Shan. At the same time, I analyzed the Fergana basin-fill looking for progressive changes in sedimentary paleoenvironments, source areas and stratal geometrical configurations in the subsurface and outcrops.
The data presented in this thesis suggests that low cooling rates (<1°C Myr-1), calm depositional environments, and low depositional rates (<10 m Myr-1) were widely distributed across the western Tien Shan, describing a quiescent tectonic period throughout the Paleogene. Increased cooling rates in the late Cenozoic occurred diachronously and with variable magnitudes in different ranges. This rapid cooling stage is interpreted to represent increased erosion caused by active deformation and constrains the onset of Cenozoic deformation in the western Tien Shan. Time-temperature histories derived from the northwestern Tien Shan samples show an increase in cooling rates by ~25 Ma. This event is correlated with a synchronous pulse
iv
in the South Tien Shan. I suggest that strike-slip motion along the TFF commenced at the Oligo-Miocene boundary, facilitating CCW rotation of the Fergana basin and enabling exhumation of the linked horsetail splays. Higher depositional rates (~150 m Myr-1) in the Oligo-Miocene section (Massaget Fm.) of the Fergana basin suggest synchronous deformation in the surrounding ranges. The central Alai Range also experienced rapid cooling around this time, suggesting that the onset of intramontane basin fragmentation and isolation is coeval. These results point to deformation starting simultaneously in the late Oligocene – early Miocene in geographically distant mountain ranges. I suggest that these early uplifts are controlled by reactivated structures (like the TFF), which are probably the frictionally weakest and most-suitably oriented for accommodating and transferring N-S horizontal shortening along the western Tien Shan.
Afterwards, in the late Miocene (~10 Ma), a period of renewed rapid cooling affected the Tien Shan and most mountain ranges and inherited structures started to actively deform. This episode is widely distributed and an increase in exhumation is interpreted in most of the sampled ranges. Moreover, the Pliocene section in the basin subsurface shows the higher depositional rates (>180 m Myr-1) and higher energy facies. The deformation and exhumation increase further contributed to intramontane basin partitioning. Overall, the interpretation is that the Tien Shan and much of Central Asia suffered a global increase in the rate of horizontal crustal shortening. Previously, stress transfer along the rigid Tarim block or Pamir indentation has been proposed to account for Himalayan hinterland deformation. However, the extent of the episode requires a different and broader geodynamic driver.
Precision horticulture encompasses site- or tree-specific management in fruit plantations. Of decisive importance is spatially resolved data (this means data from each tree) from the production site, since it may enable customized and, therefore, resource-efficient production measures.
The present thesis involves an examination of the apparent electrical conductivity of the soil (ECa), the plant water status spatially measured by means of the crop water stress index (CWSI), and the fruit quality (e.g. fruit size) for Prunus domestica L. (plums) and Citrus x aurantium, Syn. Citrus paradisi (grapefruit). The goals of the present work were i) characterization of the 3D distribution of the apparent electrical conductivity of the soil and variability of the plant’s water status; ii) investigation of the interaction between ECa, CWSI, and fruit quality; and iii) an approach for delineating management zones with respect to managing trees individually.
To that end, the main investigations took place in the plum orchard. This plantation got a slope of 3° grade on Pleistocene and post-Pleistocene substrates in a semi-humid climate (Potsdam, Germany) and encloses an area of 0.37 ha with 156 trees of the cultivar ˈTophit Plusˈ on a Wavit rootstock. The plantation was laid in 2009 with annual and biannual trees spaced 4 m distance along the irrigation system and 5 m between the rows. The trees were watered three times a week with a drip irrigation system positioned 50 cm above ground level providing 1.6 l per tree per event. With the help of geoelectric measurements, the apparent electrical conductivity of the upper soil (0.25 m) was measured for each tree with an electrode spacing of 0.5 m (4-point light hp). In this manner, the plantation was spatially charted with respect to the soil’s ECa. Additionally, tomography measurements were performed for 3D mapping of the soil ECa and spot checks of drilled cores with a profile of up to 1 m. The vegetative, generative, and fruit quality data were collected for each tree. The instantaneous plant water status was comprehensively determined in spot checks with the established Scholander method for water potential analysis (Scholander pressure bomb) as well as thermal imaging. An infrared camera was used for the thermal imaging (ThermaCam SC 500), mounted on a tractor 3.3 m above ground level. The thermal images (320 x 240 px) of the canopy surface were taken with an aperture of 45° and a geometric resolution of 8.54 x 6.41 mm. With the aid of the canopy temperature readings from the thermal images, cross-checked with manual temperature measurements of a dry and a wet reference leaf, the crop water stress index (CWSI) was calculated. Adjustments in CWSI for measurements in a semi-humid climate were developed, whereas the collection of reference temperatures was automatically collected from thermal images.
The bonitur data were transformed with the help of a variance stabilization process into a normal distribution. The statistical analyses as well as the automatic evaluation routine were performed with several scripts in MATLAB® (R2010b and R2016a) and a free program (spatialtoolbox). The hot spot analysis served to check whether an observed pattern is statistically significant. The method was evaluated with an established k-mean analysis. To test the hot-spot analysis by comparison, data from a grapefruit plantation (Adana, Turkey) was collected, including soil ECa, trunk circumference, and yield data. The plantation had 179 trees on a soil of type Xerofkuvent with clay and clay-loamy texture. The examination of the interaction between the critical values from the soil and plant water status information and the vegetative and generative plant growth variables was performed with the application from ANOVA.
The study indicates that the variability of the soil and plant information in fruit production is high, even considering small orchards. It was further indicated that the spatial patterns found in the soil ECa stayed constant through the years (r = 0.88 in 2011-2012 and r = 0.71 in 2012-2013). It was also demonstrated that CWSI determination may also be possible in semi-humid climate. A correlation (r = - 0.65, p < 0.0001) with the established method of leaf water potential analysis was found. The interaction between the ECa from various depths and the plant variables produced a highly significant connection with the topsoil in which the irrigation system was to be found. A correlation between yield and ECatopsoil of r = 0.52 was determined. By using the hot-spot analysis, extreme values in the spatial data could be determined. These extremes served to divide the zones (cold-spot, random, hot-spot). The random zone showed the highest correlation to the plant variables.
In summary it may be said that the cumulative water use efficiency (WUEc) was enhanced with high crop load. While the CWSI had no effect on fruit quality, the interaction of CWSI and WUEc even outweighed the impact of soil ECa on fruit quality in the production system with irrigation. In the plum orchard, irrigation was relevant for obtaining high quality produce even in the semi-humid climate.
Complex networks are ubiquitous in nature and society. They appear in vastly different domains, for instance as social networks, biological interactions or communication networks. Yet in spite of their different origins, these networks share many structural characteristics. For instance, their degree distribution typically follows a power law. This means that the fraction of vertices of degree k is proportional to k^(−β) for some constant β; making these networks highly inhomogeneous. Furthermore, they also typically have high clustering, meaning that links between two nodes are more likely to appear if they have a neighbor in common.
To mathematically study the behavior of such networks, they are often modeled as random graphs. Many of the popular models like inhomogeneous random graphs or Preferential Attachment excel at producing a power law degree distribution. Clustering, on the other hand, is in these models either not present or artificially enforced.
Hyperbolic random graphs bridge this gap by assuming an underlying geometry to the graph: Each vertex is assigned coordinates in the hyperbolic plane, and two vertices are connected if they are nearby. Clustering then emerges as a natural consequence: Two nodes joined by an edge are close by and therefore have many neighbors in common. On the other hand, the exponential expansion of space in the hyperbolic plane naturally produces a power law degree sequence. Due to the hyperbolic geometry, however, rigorous mathematical treatment of this model can quickly become mathematically challenging.
In this thesis, we improve upon the understanding of hyperbolic random graphs by studying its structural and algorithmical properties. Our main contribution is threefold. First, we analyze the emergence of cliques in this model. We find that whenever the power law exponent β is 2 < β < 3, there exists a clique of polynomial size in n. On the other hand, for β >= 3, the size of the largest clique is logarithmic; which severely contrasts previous models with a constant size clique in this case. We also provide efficient algorithms for finding cliques if the hyperbolic node coordinates are known. Second, we analyze the diameter, i. e., the longest shortest path in the graph. We find
that it is of order O(polylog(n)) if 2 < β < 3 and O(logn) if β > 3. To complement
these findings, we also show that the diameter is of order at least Ω(logn). Third, we provide an algorithm for embedding a real-world graph into the hyperbolic plane using only its graph structure. To ensure good quality of the embedding, we perform extensive computational experiments on generated hyperbolic random graphs. Further, as a proof of concept, we embed the Amazon product recommendation network and observe that products from the same category are mapped close together.
During the last decade, high intensity interval training (HIIT) has been used as an alternative to endurance (END) exercise, since it requires less time to produce similar physiological adaptations. Previous literature has focused on HIIT changes in aerobic metabolism and cardiorespiratory fitness, however, there are currently no studies focusing on its neuromuscular adaptations.
Therefore, this thesis aimed to compare the neuromuscular adaptations of both HIIT and END after a two-week training intervention, by using a novel technology called high-density surface electromyography (HDEMG) motor unit decomposition. This project consisted in two experiments, where healthy young men were recruited (aged between 18 to 35 years). In experiment one, the reliability of HDEMG motor unit variables (mean discharge rate, peak-to-peak amplitude, conduction velocity and discharge rate variability) was tested (Study 1), a new method to track the same motor units longitudinally was proposed (Study 2), and the level of low (<5Hz) and high (>5Hz) frequency motor unit coherence between vastus medialis (VM) and lateralis (VL) knee extensor muscles was measured (Study 4). In experiment two, a two-week HIIT and END intervention was conducted where cardiorespiratory fitness parameters (e.g. peak oxygen uptake) and motor unit variables from the VM and VL muscles were assessed pre and post intervention (Study 3).
The results showed that HDEMG is reliable to monitor changes in motor unit activity and also allows the tracking of the same motor units across different testing sessions. As expected, both HIIT and END improved cardiorespiratory fitness parameters similarly. However, the neuromuscular adaptations of both types of training differed after the intervention, with HIIT showing a significant increase in knee extensor muscle strength that was accompanied by increased VM and VL motor unit discharge rates and HDEMG amplitude at the highest force levels [(50 and 70% of the maximum voluntary contraction force (MVC)], while END training induced a marked increase in time to task failure at lower force levels (30% MVC), without any influence on HDEMG amplitude and discharge rates. Additionally, the results showed that VM and VL muscles share most of their synaptic input since they present a large amount of low and high frequency motor unit coherence, which can explain the findings of the training intervention where both muscles showed similar changes in HDEMG amplitude and discharge rates.
Taken together, the findings of the current thesis show that despite similar improvements in cardiopulmonary fitness, HIIT and END induced opposite adjustments in motor unit behavior. These results suggest that HIIT and END show specific neuromuscular adaptations, possibly related to their differences in exercise load intensity and training volume.
This work reports about new high-resolution imaging and spectroscopic observations of solar type III radio bursts at low radio frequencies in the range from 30 to 80 MHz. Solar type III radio bursts are understood as result of the beam-plasma interaction of electron beams in the corona. The Sun provides a unique opportunity to study these plasma processes of an active star. Its activity appears in eruptive events like flares, coronal mass ejections and radio bursts which are all accompanied by enhanced radio emission. Therefore solar radio emission carries important information about plasma processes associated with the Sun’s activity. Moreover, the Sun’s atmosphere is a unique plasma laboratory with plasma processes under conditions not found in terrestrial laboratories. Because of the Sun’s proximity to Earth, it can be studied in greater detail than any other star but new knowledge about the Sun can be transfer to them. This “solar stellar connection” is important for the understanding of processes on other stars.
The novel radio interferometer LOFAR provides imaging and spectroscopic capabilities to study these processes at low frequencies. Here it was used for solar observations.
LOFAR, the characteristics of its solar data and the processing and analysis of the latter with the Solar Imaging Pipeline and Solar Data Center are described. The Solar Imaging Pipeline is the central software that allows using LOFAR for solar observations. So its development was necessary for the analysis of solar LOFAR data and realized here. Moreover a new density model with heat conduction and Alfvén waves was developed that provides the distance of radio bursts to the Sun from dynamic radio spectra.
Its application to the dynamic spectrum of a type III burst observed on March 16, 2016 by LOFAR shows a nonuniform radial propagation velocity of the radio emission. The analysis of an imaging observation of type III bursts on June 23, 2012 resolves a burst as bright, compact region localized in the corona propagating in radial direction along magnetic field lines with an average velocity of 0.23c. A nonuniform propagation velocity is revealed. A new beam model is presented that explains the nonuniform motion of the radio source as a propagation effect of an electron ensemble with a spread velocity distribution and rules out a monoenergetic electron distribution. The coronal electron number density is derived in the region from 1.5 to 2.5 R☉ and fitted with the newly developed density model. It determines the plasma density for the interplanetary space between Sun and Earth. The values correspond to a 1.25- and 5-fold Newkirk model for harmonic and fundamental emission, respectively. In comparison to data from other radio instruments the LOFAR data shows a high sensitivity and resolution in space, time and frequency.
The new results from LOFAR’s high resolution imaging spectroscopy are consistent with current theories of solar type III radio bursts and demonstrate its capability to track fast moving radio sources in the corona. LOFAR solar data is found to be a valuable source for solar radio physics and opens a new window for studying plasma processes associated with highly energetic electrons in the solar corona.
Proteins are natural polypeptides produced by cells; they can be found in both animals and plants, and possess a variety of functions. One of these functions is to provide structural support to the surrounding cells and tissues. For example, collagen (which is found in skin, cartilage, tendons and bones) and keratin (which is found in hair and nails) are structural proteins. When a tissue is damaged, however, the supporting matrix formed by structural proteins cannot always spontaneously regenerate. Tailor-made synthetic polypeptides can be used to help heal and restore tissue formation.
Synthetic polypeptides are typically synthesized by the so-called ring opening polymerization (ROP) of α-amino acid N-carboxyanhydrides (NCA). Such synthetic polypeptides are generally non-sequence-controlled and thus less complex than proteins. As such, synthetic polypeptides are rarely as efficient as proteins in their ability to self-assemble and form hierarchical or structural supramolecular assemblies in water, and thus, often require rational designing. In this doctoral work, two types of amino acids, γ-benzyl-L/D-glutamate (BLG / BDG) and allylglycine (AG), were selected to synthesize a series of (co)polypeptides of different compositions and molar masses.
A new and versatile synthetic route to prepare polypeptides was developed, and its mechanism and kinetics were investigated. The polypeptide properties were thoroughly studied and new materials were developed from them. In particular, these polypeptides were able to aggregate (or self-assemble) in solution into microscopic fibres, very similar to those formed by collagen. By doing so, they formed robust physical networks and organogels which could be processed into high water-content, pH-responsive hydrogels. Particles with highly regular and chiral spiral morphologies were also obtained by emulsifying these polypeptides. Such polypeptides and the materials derived from them are, therefore, promising candidates for biomedical applications.
Extreme hydro-meteorological events, such as severe droughts or heavy rainstorms, constitute primary manifestations of climate variability and exert a critical impact on the natural environment and human society. This is particularly true for high-mountain areas, such as the eastern flank of the southern Central Andes of NW Argentina, a region impacted by deep convection processes that form the basis of extreme events, often resulting in floods, a variety of mass movements, and hillslope processes. This region is characterized by pronounced E-W gradients in topography, precipitation, and vegetation cover, spanning low to medium-elevation, humid and densely vegetated areas to high-elevation, arid and sparsely vegetated environments. This strong E-W gradient is mirrored by differences in the efficiency of surface processes, which mobilize and transport large amounts of sediment through the fluvial system, from the steep hillslopes to the intermontane basins and further to the foreland. In a highly sensitive high-mountain environment like this, even small changes in the spatiotemporal distribution, magnitude and rates of extreme events may strongly impact environmental conditions, anthropogenic activity, and the well-being of mountain communities and beyond. However, although the NW Argentine Andes comprise the catchments for the La Plata river that traverses one of the most populated and economically relevant areas of South America, there are only few detailed investigations of climate variability and extreme hydro-meteorological events.
In this thesis, I focus on deciphering the spatiotemporal variability of rainfall and river discharge, with particular emphasis on extreme hydro-meteorological events in the subtropical southern Central Andes of NW Argentina during the past seven decades. I employ various methods to assess and quantify statistically significant trend patterns of rainfall and river discharge, integrating high-quality daily time series from gauging stations (40 rainfall and 8 river discharge stations) with gridded datasets (CPC-uni and TRMM 3B42 V7), for the period between 1940 and 2015. Evidence for a general intensification of the hydrological cycle at intermediate elevations (~ 0.5 – 3 km asl) at the eastern flank of the southern Central Andes is found both from rainfall and river-discharge time-series analysis during the period from 1940 to 2015. This intensification is associated with the increase of the annual total amount of rainfall and the mean annual discharge. However, most pronounced trends are found at high percentiles, i.e. extreme hydro-meteorological events, particularly during the wet season from December to February.An important outcome of my studies is the recognition of a rapid increase in the amount of river discharge during the period between 1971 and 1977, most likely linked to the 1976-77 global climate shift, which is associated with the North Pacific Ocean sea surface temperature variability. Interestingly, after this rapid increase, both rainfall and river discharge decreased at low and intermediate elevations along the eastern flank of the Andes. In contrast, during the same time interval, at high elevations, extensive areas on the arid Puna de Atacama plateau have recorded increasing annual rainfall totals. This has been associated with more intense extreme hydro-meteorological events from 1979 to 2014. This part of the study reveals that low-, intermediate, and high-elevation sectors in the Andes of NW Argentina respond differently to changing climate conditions.
Possible forcing mechanisms of the pronounced hydro-meteorological variability observed in the study area are also investigated. For the period between 1940 and 2015, I analyzed modes of oscillation of river discharge from small to medium drainage basins (102 to 104 km2), located on the eastern flank of the orogen. First, I decomposed the relevant monthly time series using the Hilbert-Huang Transform, which is particularly appropriate for non-stationary time series that result from non-linear natural processes. I observed that in the study region discharge variability can be described by five quasi-periodic oscillatory modes on timescales varying from 1 to ~20 years. Secondly, I tested the link between river-discharge variations and large-scale climate modes of variability, using different climate indices, such as the BEST ENSO (Bivariate El Niño-Southern Oscillation Time-series) index. This analysis reveals that, although most of the variance on the annual timescale is associated with the South American Monsoon System, a relatively large part of river-discharge variability is linked to Pacific Ocean variability (PDO phases) at multi-decadal timescales (~20 years). To a lesser degree, river discharge variability is also linked to the Tropical South Atlantic (TSA) sea surface temperature anomaly at multi-annual timescales (~2-5 years).
Taken together, these findings exemplify the high degree of sensitivity of high-mountain environments with respect to climatic variability and change. This is particularly true for the topographic transitions between the humid, low-moderate elevations and the semi-arid to arid highlands of the southern Central Andes. Even subtle changes in the hydro-meteorological regime of these areas of the mountain belt react with major impacts on erosional hillslope processes and generate mass movements that fundamentally impact the transport capacity of mountain streams. Despite more severe storms in these areas, the fluvial system is characterized by pronounced variability of the stream power on different timescales, leading to cycles of sediment aggradation, the loss of agriculturally used land and severe impacts on infrastructure.
Solar-like stars maintain their magnetic fields thanks to a dynamo mechanism. The Babcock-Leighton dynamo is one possible dynamo that has the particularity to require magnetic flux tubes. Magnetic flux tubes are assumed to form at the bottom of the convective zone and rise buoyantly to the surface. A delayed dynamo model has been suggested, where the delay accounts for the rise time of the magnetic flux tubes; a time, that has been ignored by former studies.
The present thesis aims to study the applicability of the flux tube/Babcock-Leighton dynamo to other stars. To do so, we attempt to constrain the rise time of magnetic flux tubes thanks to the first fully compressible MHD simulations of rising magnetic flux tubes in stratified rotating spherical shells.
Such simulations are limited to an unrealistic parameter space, therefore, a scaling relation is required to scale the results to realistic physical regimes. We extended earlier works on 2D scaling relations and derived a general scaling law valid for both 2D and 3D. We then carried out two large series of numerical experiments and verified that the scaling law we have derived indeed applies to the fully non-linear case. It allowed us to extract a constraint for the rise time of magnetic flux tubes that is valid for any solar-like star. We finally introduced this constraint to a delayed dynamo model.
By carrying out simulations of a mean-field, delayed, flux tube/Babcock-Leighton dynamo, we were able to identify a new dynamo regime resulting from the delay. This regime requires delays about an entire cycle and exhibits subequipartition magnetic activity. Revealing this new regime shows that even for long delays the flux tube/Babcock-Leighton dynamo can still deliver non-decaying solutions and remains a good candidate for a wide range of solar-like stars.
Diese Arbeit befasst sich mit der Herstellung und Charakterisierung von thermoresponsiven Filmen auf Goldelektroden durch Fixierung eines bereits synthetisierten thermoresponsiven Polymers. Als Basis für die Entwicklung der responsiven Grenzfläche dienten drei unterschiedliche Copolymere (Polymere I, II und III) aus der Gruppe der thermisch schaltbaren Poly(oligo(ethylenglykol)methacrylate).
Die turbidimetrischen Messungen der Copolymere in Lösungen haben gezeigt, dass der Trübungspunkt vom pH-Wert, der Gegenwart von Salzen sowie von der Ionenstärke der Lösung abhängig ist. Nach der Charakterisierung der Polymere in Lösung wurden Experimente der kovalenten Kopplung der Polymere I bis III an die Oberfläche der Gold-Elektroden durchgeführt. Während bei Polymeren I und II die Ankopplung auf einer Amidverbrückung basierte, wurde bei Polymer III als alternative Methode zur Immobilisierung eine photoinduzierte Anbindung unter gleichzeitiger Vernetzung gewählt. Der Nachweis der erfolgreichen Ankopplung erfolgte bei allen Polymeren elektrochemisch mittels Cyclovoltammetrie und Impedanzspektroskopie in K3/4[Fe(CN)6]-Lösungen. Wie die Ellipsometrie-Messungen zeigten, waren die erhaltenen Polymer-Filme unterschiedlich dick. Die Ankopplung über Amidverbrückung lieferte dünne Filme (10 – 15 nm), während der photovernetzte Film deutlich dicker war (70-80 nm) und die darunter liegende Oberfläche relativ gut isolierte.
Elektrochemische Temperaturexperimente an Polymer-modifizierten Oberflächen in Lösungen in Gegenwart von K3/4[Fe(CN)6] zeigten, dass auch die immobilisierten Polymere I bis III responsives Temperaturverhalten zeigen. Bei Elektroden mit den immobilisierten Polymeren I und II ist der Temperaturverlauf der Parameterwerte diskontinuierlich – ab einem kritischen Punkt (37 °C für Polymer I und 45 °C für Polymer II) wird zunächst langsame Zunahme der Peakströme wird deutlich schneller. Das Temperaturverhalten von Polymer III ist dagegen bis 50 °C kontinuierlich, der Peakstrom sinkt hier durchgehend.
Weiterhin wurde mit den auf Polymeren II und III basierten Elektroden deren Anwendung als responsive Matrix für Bioerkennungsreaktionen untersucht. Es wurde die Ankopplung von kleinen Biorezeptoren, TAG-Peptiden, an Polymer II- und Polymer III-modifizierten Elektroden durchgeführt. Das hydrophile FLAG-TAG-Peptid verändert das Temperaturverhalten des Polymer II-Films unwesentlich, da es die Hydrophilie des Netzwerkes nicht beeinflusst. Weiterhin wurde der Effekt der Ankopplung der ANTI-FLAG-TAG-Antikörper an FLAG-TAG-modifizierte Polymer II-Filme untersucht. Es konnte gezeigt werden, dass die Antikörper spezifisch an FLAG-TAG-modifiziertes Polymer II binden. Es wurde keine unspezifische Anbindung von ANTI-FLAG-TAG an Polymer II beobachtet. Die Temperaturexperimente haben gezeigt, dass die thermische Restrukturierung des Polymer II-FLAG-TAG-Filmes auch nach der Antikörper-Ankopplung noch stattfindet. Der Einfluss der ANTI-FLAG-TAG-Ankopplung ist gering, da der Unterschied in der Hydrophilie zwischen Polymer II und FLAG-TAG bzw. ANTI-FLAG-TAG zu gering ist.
Für die Untersuchungen mit Polymer III-Elektroden wurde neben dem hydrophilen FLAG-TAG-Peptid das deutlich hydrophobere HA-TAG-Peptid ausgewählt. Wie im Falle der Polymer II Elektrode beeinflusst das gekoppelte FLAG-TAG-Peptid das Temperaturverhalten des Polymer III-Netzwerkes nur geringfügig. Die gemessenen Stromwerte sind geringer als bei der Polymer III-Elektrode. Das Temperaturverhalten der FLAG-TAG-Elektrode ähnelt dem der reinen Polymer III-Elektrode – die Stromwerte sinken kontinuierlich bis die Temperatur von ca. 40 °C erreicht ist, bei der ein Plateau beobachtet wird. Offensichtlich verändert FLAG-TAG auch in diesem Fall nicht wesentlich die Hydrophilie des Polymer III-Netzwerkes. Das an Polymer III-Elektroden gekoppelte hydrophobe HA-TAG-Peptid beeinflusst dagegen im starken Maße den Quellzustand des Netzwerkes. Die Ströme für die HA-TAG-Elektroden sind deutlich geringer als die für die FLAG-TAG-Polymer III-Elektroden, was auf geringeren Wassergehalt und dickeren Film zurückzuführen ist. Bereits ab 30 °C erfolgt der Anstieg von Stromwerten, der bei Polymer III- bzw. bei Polymer III-FLAG-TAG-Elektroden nicht beobachtet werden kann. Das gekoppelte hydrophobe HA-TAG-Peptid verdrängt Wasser aus dem Polymer III-Netzwerk, was in der Stauchung des Films bereits bei Raumtemperatur resultiert. Dies führt dazu, dass der Film im Laufe des Temperaturanstieges kaum noch komprimiert. Die Stromwerte steigen in diesem Fall entsprechend des Anstiegs der temperaturabhängigen Diffusion des Redoxpaares. Diese Untersuchungen zeigen, dass das HA-TAG-Peptid als Ankermolekül deutlich besser für eine potentielle Verwendung der Polymer III-Filme für sensorische Zwecke geeignet ist, da es sich deutlich in der Hydrophilie von Polymer III unterscheidet.
Numerous reports of relatively rapid climate changes over the past century make a clear case of the impact of aerosols and clouds, identified as sources of largest uncertainty in climate projections. Earth’s radiation balance is altered by aerosols depending on their size, morphology and chemical composition. Competing effects in the atmosphere can be further studied by investigating the evolution of aerosol microphysical properties, which are the focus of the present work.
The aerosol size distribution, the refractive index, and the single scattering albedo are commonly used such properties linked to aerosol type, and radiative forcing. Highly advanced lidars (light detection and ranging) have reduced aerosol monitoring and optical profiling into a routine process. Lidar data have been widely used to retrieve the size distribution through the inversion of the so-called Lorenz-Mie model (LMM). This model offers a reasonable treatment for spherically approximated particles, it no longer provides, though, a viable description for other naturally occurring arbitrarily shaped particles, such as dust particles. On the other hand, non-spherical geometries as simple as spheroids reproduce certain optical properties with enhanced accuracy. Motivated by this, we adapt the LMM to accommodate the spheroid-particle approximation introducing the notion of a two-dimensional (2D) shape-size distribution.
Inverting only a few optical data points to retrieve the shape-size distribution is classified as a non-linear ill-posed problem. A brief mathematical analysis is presented which reveals the inherent tendency towards highly oscillatory solutions, explores the available options for a generalized solution through regularization methods and quantifies the ill-posedness. The latter will improve our understanding on the main cause fomenting instability in the produced solution spaces. The new approach facilitates the exploitation of additional lidar data points from depolarization measurements, associated with particle non-sphericity. However, the generalization of LMM vastly increases the complexity of the problem. The underlying theory for the calculation of the involved optical cross sections (T-matrix theory) is computationally so costly, that would limit a retrieval analysis to an unpractical point. Moreover the discretization of the model equation by a 2D collocation method, proposed in this work, involves double integrations which are further time consuming. We overcome these difficulties by using precalculated databases and a sophisticated retrieval software (SphInX: Spheroidal Inversion eXperiments) especially developed for our purposes, capable of performing multiple-dataset inversions and producing a wide range of microphysical retrieval outputs.
Hybrid regularization in conjunction with minimization processes is used as a basis for our algorithms. Synthetic data retrievals are performed simulating various atmospheric scenarios in order to test the efficiency of different regularization methods. The gap in contemporary literature in providing full sets of uncertainties in a wide variety of numerical instances is of major concern here. For this, the most appropriate methods are identified through a thorough analysis on an overall-behavior basis regarding accuracy and stability. The general trend of the initial size distributions is captured in our numerical experiments and the reconstruction quality depends on data error level. Moreover, the need for more or less depolarization points is explored for the first time from the point of view of the microphysical retrieval. Finally, our approach is tested in various measurement cases giving further insight for future algorithm improvements.
Möglichkeiten der Mittelstandsförderung durch Vergaberechtsgestaltung und Vergaberechtspraxis
(2016)
Die Förderungswürdigkeit und die Förderungsfähigkeit mittelständischer Unternehmen ist ein gesamteuropäisches, wirtschaftspolitisches Anliegen. Hiervon zeugen zum einen zahlreiche Regelungen im Primär-, Sekundär-, Verfassungs- und einfachgesetzlichem Recht, zum anderen auch die Bedeutung der mittelständischen Unternehmen im wirtschaftlichen, gesellschaftlichen und sozialen Gefüge. So herrscht innerhalb der Europäischen Union nicht nur der Slogan „Vorfahrt für KMU“, sondern auch die im Frühjahr 2014 verabschiedeten Vergaberichtlinien legten ein besonderes Augenmerk auf die Förderung des Zugangs der KMU zum öffentlichen Beschaffungsmarkt. Denn gemessen am Steuerungs- und Lenkungspotenzial der Auftragsvergabe, deren Einfluss auf die Innovationstätigkeit der Wirtschaft sowie deren Auswirkungen auf die Wirtschafts- und Wettbewerbstätigkeit auf der einen Seite und dem gesamtwirtschaftlichen Stellenwert der mittelständischen Unternehmen auf der anderen Seite, sind mittelständische Unternehmen trotz zahlreicher europäischer und nationaler Initiativen im Vergabeverfahren unterrepräsentiert. Neben der undurchsichtigen Regelungsstruktur des deutschen Vergaberechts, unterliegen die mittelständischen Unternehmen vom Beginn bis zum Ende des Vergabeverfahrens besonderen Schwierigkeiten. Dieser Ausgangsbefund wurde zum Anlass genommen, um die Möglichkeiten der Mittelstandsförderung durch Vergaberechtsgestaltung und Vergaberechtspraxis erneut auf den Prüfstand zu stellen.
Infants' lexical processing is modulated by featural manipulations made to words, suggesting that early lexical representations are sufficiently specified to establish a match with the corresponding label. However, the precise degree of detail in early words requires further investigation due to equivocal findings. We studied this question by assessing children’s sensitivity to the degree of featural manipulation (Chapters 2 and 3), and sensitivity to the featural makeup of homorganic and heterorganic consonant clusters (Chapter 4). Gradient sensitivity on the one hand and sensitivity to homorganicity on the other hand would suggest that lexical processing makes use of sub-phonemic information, which in turn would indicate that early words contain sub-phonemic detail. The studies presented in this thesis assess children’s sensitivity to sub-phonemic detail using minimally demanding online paradigms suitable for infants: single-picture pupillometry and intermodal preferential looking. Such paradigms have the potential to uncover lexical knowledge that may be masked otherwise due to cognitive limitations. The study reported in Chapter 2 obtained a differential response in pupil dilation to the degree of featural manipulation, a result consistent with gradient sensitivity. The study reported in Chapter 3 obtained a differential response in proportion of looking time and pupil dilation to the degree of featural manipulation, a result again consistent with gradient sensitivity. The study reported in Chapter 4 obtained a differential response to the manipulation of homorganic and heterorganic consonant clusters, a result consistent with sensitivity to homorganicity. These results suggest that infants' lexical representations are not only specific, but also detailed to the extent that they contain sub-phonemic information.
Das Widerspenstige bändigen
(2016)
Dem Handeln von Lehrkräften wird in der schulischen Praxis wie in der wissenschaftlichen Literatur ein wesentlicher Einfluss auf die Qualität von schulischem Unterricht zugesprochen. Auch wenn umfangreiche normative Vorstellungen über ein gutes Lehr-Handeln bestehen, so gibt es wenig Erkenntnis darüber, welche Gründe Lehrkräfte für ihr pädagogisches Handeln haben. Das Handeln von Lehrkräften kann nur dann adäquat erfasst werden, wenn Bildung einerseits als Weitergabe von Kultur an die nachfolgende Generation und andererseits als eine vom sich bildenden Subjekt ausgehende Selbst- und Weltverständigung verstanden wird. Damit einhergehende Anforderungen an die Lehrkraft stehen notwendigerweise in Widerspruch zueinander; dies gilt besonders für eine Gesellschaft mit großer kultureller und sozialer Heterogenität. Bei der Suche nach Zusammenhängen zwischen Persönlichkeit, pädagogischem Wissen oder Kompetenzen und einem unterrichtlichen Handeln wird häufig von einer Bedingtheit dieses Handelns ausgegangen und dieses auf kognitive Aspekte und an externen Normen orientierte Merkmale verkürzt. Ertragreicher für eine Antwort auf die Frage nach den Begründungen sind wissenschaftliche Arbeiten, die Professionalität als eine Bezugnahme auf einen besonderen strukturellen Rahmen beschreiben, der durch Widersprüche geprägt ist und Entscheidungen zu den Spannungsfeldern pädagogischer Verhältnisse erfordert. Die subjektwissenschaftliche Lerntheorie bietet eine Basis für ein Verständnis eines Lernens in institutionellen Kontexten ausgehend von den Lerninteressen der Schülerinnen und Schüler. Lehren kann darauf bezugnehmend als Unterstützung von Selbst- und Weltverständigungsprozessen durch Wertschätzung, Verstehen und Angebote alternativer Bedeutungshorizonte verstanden werden. Das Handeln von Lehrkräften ist als sinngebende Bezugnahme auf daraus resultierende sowie institutionelle Anforderungen mittels gesellschaftlicher Bedeutungsstrukturen verstehbar. Das handelnde Subjekt erschließt sich selbst und die Welt mit Hilfe von Bedeutungen. Diese können verstanden werden als der Besonderheit der Biographie, der gesellschaftlichen Position sowie der Lebenslage geschuldete Reinterpretationen gesellschaftlicher Bedeutungsstrukturen. Im empirischen Verfahren können mittels eines Übergangs von sequentiellen zu komparativen Analysen Positionierungen als thematisch spezifische und über die konkrete Handlungssituation hinausreichende Bedeutungs-Begründungs-Zusammenhänge rekonstruiert werden. Daraus werden situationsunabhängige Strukturmomente des Gegenstands Lehren an beruflichen Schulen aber auch komplexe, situationsbezogene subjektive Bedeutungs-Begründungs-Muster abgeleitet. Als wesentliche strukturelle Merkmale lassen sich die Schlüsselkategorien ‚Deutungsmacht‘ und ‚instrumentelle pädagogische Beziehung‘ aus dem empirischen Material unter Zuhilfenahme weiterer theoretischer Folien entwickeln. Da Deutungsmacht auf Akzeptanz angewiesen ist und in instrumentellen Beziehungen eine kooperative Bezugnahme auf den Lehr-Lern-Gegenstand allenfalls punktuell erfolgt, können damit asymmetrische metastabile Arrangements zwischen einer Lehrkraft und Schülerinnen und Schülern verstanden werden. Als empirische Ausprägungen weist Deutungsmacht die Varianten ‚absoluter Anspruch‘, ‚Akzeptanz der Fragilität‘ und ‚Akzeptanz der Legitimität eines Infragestellens‘ auf. Bei der zweiten Schlüsselkategorie treten die Varianten ‚strukturelle Prägung‘, ‚unspezifischer allgemein-menschlicher Charakter‘ und ‚Außenprägung‘ der instrumentellen pädagogischen Beziehung auf. Die Bedeutungs-Begründungs-Musters weisen teilweise Inkonsistenzen und Übergänge in den Positionierungen bezogen auf die dargestellten Varianten auf. Nur bei einem Teil der Muster sind Bemühungen um Wertschätzung und Verstehen der Schülerinnen und Schüler plausibel ableitbar, gleiches gilt in Hinblick auf eine Offenheit für eine Revision der Muster. Die Muster, wie etwa ‚Durchsetzend-ertragendes Nachsteuern‘, ‚Direktiv-personalisierendes Praktizieren‘ oder ‚Regulierend-flexibles Managen‘ sind zu verstehen als Bewältigungsmodi der kontingenten pädagogischen (Konflikt-)Situationen, auf die sich die Fallschilderungen beziehen. Die jeweilige Lehrkraft hat dieses Muster in dem beschriebenen Fall genutzt, was allerdings keine Aussage darüber zulässt, auf welche Muster die Lehrkraft in anderen Fällen zugreifen würde. Die Ergebnisse der vorliegenden Arbeit eignen sich als eine heuristische bzw. theoretische Folie, die Lehrkräfte beim Erschließen ihres eigenen pädagogischen Handelns - etwa in einer als Fallberatung konzipierten Fortbildung - unterstützen kann. Möglich sind Anschlüsse an andere theoretische Ansätze zum Handeln von Lehrkräften aber auch deren veränderte Einordnung. Erweitert werden die Optionen, dieses Handeln über wissenschaftliche Zugänge zu erfassen.
Eye movements serve as a window into ongoing visual-cognitive processes and can thus be used to investigate how people perceive real-world scenes. A key issue for understanding eye-movement control during scene viewing is the roles of central and peripheral vision, which process information differently and are therefore specialized for different tasks (object identification and peripheral target selection respectively). Yet, rather little is known about the contributions of central and peripheral processing to gaze control and how they are coordinated within a fixation during scene viewing. Additionally, the factors determining fixation durations have long been neglected, as scene perception research has mainly been focused on the factors determining fixation locations. The present thesis aimed at increasing the knowledge on how central and peripheral vision contribute to spatial and, in particular, to temporal aspects of eye-movement control during scene viewing. In a series of five experiments, we varied processing difficulty in the central or the peripheral visual field by attenuating selective parts of the spatial-frequency spectrum within these regions. Furthermore, we developed a computational model on how foveal and peripheral processing might be coordinated for the control of fixation duration. The thesis provides three main findings. First, the experiments indicate that increasing processing demands in central or peripheral vision do not necessarily prolong fixation durations; instead, stimulus-independent timing is adapted when processing becomes too difficult. Second, peripheral vision seems to play a prominent role in the control of fixation durations, a notion also implemented in the computational model. The model assumes that foveal and peripheral processing proceed largely in parallel and independently during fixation, but can interact to modulate fixation duration. Thus, we propose that the variation in fixation durations can in part be accounted for by the interaction between central and peripheral processing. Third, the experiments indicate that saccadic behavior largely adapts to processing demands, with a bias of avoiding spatial-frequency filtered scene regions as saccade targets. We demonstrate that the observed saccade amplitude patterns reflect corresponding modulations of visual attention. The present work highlights the individual contributions and the interplay of central and peripheral vision for gaze control during scene viewing, particularly for the control of fixation duration. Our results entail new implications for computational models and for experimental research on scene perception.
Water scarcity, adaption on climate change, and risk assessment of droughts and floods are critical topics for science and society these days. Monitoring and modeling of the hydrological cycle are a prerequisite to understand and predict the consequences for weather and agriculture. As soil water storage plays a key role for partitioning of water fluxes between the atmosphere, biosphere, and lithosphere, measurement techniques are required to estimate soil moisture states from small to large scales.
The method of cosmic-ray neutron sensing (CRNS) promises to close the gap between point-scale and remote-sensing observations, as its footprint was reported to be 30 ha. However, the methodology is rather young and requires highly interdisciplinary research to understand and interpret the response of neutrons to soil moisture. In this work, the signal of nine detectors has been systematically compared, and correction approaches have been revised to account for meteorological and geomagnetic variations. Neutron transport simulations have been consulted to precisely characterize the sensitive footprint area, which turned out to be 6--18 ha, highly local, and temporally dynamic. These results have been experimentally confirmed by the significant influence of water bodies and dry roads. Furthermore, mobile measurements on agricultural fields and across different land use types were able to accurately capture the various soil moisture states. It has been further demonstrated that the corresponding spatial and temporal neutron data can be beneficial for mesoscale hydrological modeling. Finally, first tests with a gyrocopter have proven the concept of airborne neutron sensing, where increased footprints are able to overcome local effects.
This dissertation not only bridges the gap between scales of soil moisture measurements. It also establishes a close connection between the two worlds of observers and modelers, and further aims to combine the disciplines of particle physics, geophysics, and soil hydrology to thoroughly explore the potential and limits of the CRNS method.
Subcultures creating culture
(2016)
The purpose of this work is to apply the methods of textual semiotics to subcultures, in particular to the little known glam subculture. Subcultures have been the main research field of the Birmingham Centre for Contemporary Cultural Studies, known for its interdisciplinary approach, and for its focus on the creative aspects of subculture. Hebdige, in particular, introduced many semiotic elements in his work, as the aberrant decoding after Eco and the cultural creativity via bricolage after Lévi-Strauss. His definition of subculture as symbolic resistance has been criticized by the following post-subcultural researchers for its abstractness and lack of cohesion.
Semiotics eventually have been expelled from the set of tools used in sociology for the analysis of subcultures. Nowadays, the studies on subcultures have a strong ethnographic focus. Due to terminological proliferation and a descriptive approach, it is difficult to compare them on a common basis.
Textual semiotics, through the concept of semiosphere developed by Lotman, allows to go back to the intuitions of Hebdige, organizing the semiotic elements already present in his work into a wider system of interpretation. The semiosphere offers a coherent theoretical horizon as a basis for further analysis, and a new methodological perspective focusing on the cultural. In this thesis for the first time the work of Lotman is applied to the study of a subculture.
Dietary approaches contribute to the prevention and treatment of type 2 diabetes. High protein diets were shown to exert beneficial as well as adverse effects on metabolism. However, it is unclear whether the protein origin plays a role in these effects. The LeguAN study investigated in detail the effects of two high protein diets, either from plant or animal origin, in type 2 diabetic patients. Both diets contained 30 EN% protein, 40 EN% carbohydrates, and 30 EN% fat. Fiber content, glycemic index, and composition of dietary fats were similar in both diets. In comparison to previous dietary habits, the fat content was exchanged for protein, while the carbohydrate intake was not modified. Overall, both high protein diets led to improvements of glycemic control, insulin sensitivity, liver fat, and cardiovascular risk markers without remarkable differences between the protein types.
Fasting glucose together with indices of insulin resistance were ameliorated by both interventions to varying extents but without significant differences between protein types. The decline of HbA1c was more pronounced in the plant protein group, whereby the improvement of insulin sensitivity in the animal protein group. The high protein intake had only slight influence on postprandial metabolism seen for free fatty acids and indices of insulin secretion, sensitivity and degradation. Except for GIP release, ingestion of animal and plant meals did not provoke differential metabolic and hormonal responses despite diverse circulating amino acid levels.
The animal protein diets led to a selective increase of fat-free mass and decrease of total fat mass, which was not significantly different from the plant protein diet. Moreover, the high protein diets potently decreased liver fat content by 42% on average which was linked to significantly diminished lipogenesis, free fatty acids flux and lipolysis in adipose tissue. Moderate decline of circulating liver enzymes was induced by both interventions. The liver fat reduction was associated with improved glucose homeostasis and insulin sensitivity which underlines the protective effect of the diets.
Blood lipid profile improved in all subjects and was probably related to the lower fat intake. Reductions in uric acid and markers of inflammation further argued for metabolic benefits of both high protein diets. Systolic and diastolic blood pressure declined only in the PP group pointing a possible role of arginine.
Kidney function was not altered by high protein consumption over 6 weeks. The rapid decrease of serum creatinine in the PP group was noteworthy and should be further investigated. Protein type did not seem to play a role but long-term studies are warranted to fully elucidate safety of high protein regimen.
Varying the source of dietary proteins did not affect the mTOR pathway in adipose tissue and blood cells under neither acute nor chronic settings. Enhancement of whole-body insulin sensitivity suggested also no alteration of mTOR and no impairment of insulin sensitivity in skeletal muscle.
A remarkable outcome was the extensive reduction of FGF21, critical regulator of metabolic processes, by approximately 50% independently of protein type. Whether hepatic ER-stress, ammonia flux or rather macronutrient preferences is behind this paradoxical finding remains to be investigated in detail.
Unlike initial expectations and previous reports plant protein based diet had no clear advantage over animal proteins. The pronounced beneficial effect of animal protein on insulin homeostasis despite high BCAA and methionine intake was certainly unexpected assuming more complex metabolic adaptations occurring upon prolonged consumption. In addition, the reduced fat intake may have also contributed to the overall improvements in both groups.
Taking into account the above observed study results, a short-term diet containing 30 EN% protein (either from plant or animal origin), 40 EN% carbohydrates, and 30 EN% fat with lower SFA amount leads to metabolic improvements in diabetic patients, regardless of protein source.
Durch die Zunahme metabolischer Stoffwechselstörungen und Erkrankungen in der Weltbevölkerung wird in der Medizin und den Lebenswissenschaften vermehrt nach Präventionsstrategien und Ansatzpunkten gesucht, die die Gesundheit fördern, Erkrankungen verhindern helfen und damit auch die Gesamtlast auf die Gesundheitssysteme erleichtern. Ein Ansatzpunkt wird dabei in der Ernährung gesehen, da insbesondere der Konsum von gesättigten Fetten die Gesundheit nachträglich zu beeinflussen scheint. Dabei wird übersehen, dass in vielen Studien Hochfettdiäten nicht ausreichend von den Einflüssen einer zum Bedarf hyperkalorischen Energiezufuhr getrennt werden, sodass die Datenlage zu dem Einfluss von (gesättigten) Fetten auf den Metabolismus bei gleichbleibender Energieaufnahme noch immer unzureichend ist.
In der NUtriGenomic Analysis in Twins-Studie wurden 46 Zwillingspaare (34 monozygot, 12 dizygot) über einen Zeitraum von sechs Wochen mittels einer kohlenhydratreichen, fettarmen Diät nach Richtlinien der Deutschen Gesellschaft für Ernährung für ihr Ernährungsverhalten standardisiert, ehe sie zu einer kohlenhydratarmen, fettreichen Diät, die insbesondere gesättigte Fette enthielt, für weitere sechs Wochen wechselten. Beide Diäten waren dem individuellen Energiebedarf der Probanden angepasst, um so sowohl akut nach einerWoche als auch längerfristig nach sechs Wochen Änderungen des Metabolismus beobachten zu können, die sich in der vermehrten Aufnahme von (gesättigten) Fetten begründeten.
Die über die detaillierte Charakterisierung der Probanden an den klinischen Untersuchungstagen generierten Datensätze wurden mit statistischen und mathematischen Methoden (z.B. lineare gemischte Modellierung) analysiert, die der Größe der Datensätze und damit ihrem Informationsvolumen angepasst waren.
Es konnte gezeigt werden, dass die metabolisch gesunden und relativ jungen Probanden, die eine gute Compliance zeigten, im Hinblick auf ihren Glukosestoffwechsel adaptieren konnten, indem die Akutantwort nach einer Woche im Nüchterninsulin und dem Index für Insulinresistenz in den weiteren fünf Wochen ausgeglichen wurde.
Der Lipidstoffwechsel in Form der klassischen Marker wie Gesamtcholesterin, LDL und HDL war dagegen stärker beeinflusst und auch nach insgesamt sechs Wochen deutlich erhöht.
Letzteres unterstützt die Beobachtung im Transkriptom des weißen, subkutanen Fettgewebes, bei der eine Aktivierung der über die Toll-like receptors und das Inflammasom vermittelten subklinischen Inflammation beobachtet werden konnte.
Die auftretenden Veränderungen in Konzentration und Komposition des Plasmalipidoms zeigte ebenfalls nur eine teilweise und auf bestimmte Spezies begrenzte Gegenregulation.
Diesbezüglich kann also geschlussfolgert werden, dass auch die isokalorische Aufnahme von (gesättigten) Fetten zu Veränderungen im Metabolismus führt, wobei die Auswirkungen in weiteren (Langzeit-)Studien und Experimenten noch genauer untersucht werden müssen. Insbesondere wäre dabei ein längerer Zeitraum unter isokalorischen Bedingungen von Interesse und die Untersuchung von Probanden mit metabolischer Vorbelastung (z.B. Insulinresistenz).
Darüber hinaus konnte in NUGAT aber ebenfalls gezeigt werden, dass die Nutrigenetik und Nutrigenomik zwei nicht zu vernachlässigende Faktoren darstellen. So zeigten unter anderem die Konzentrationen einiger Lipidspezies eine starke Erblichkeit und Abhängigkeit der Diät.
Zudem legen die Ergebnisse nahe, dass laufende wie geplante Präventionsstrategien und medizinische Behandlungen deutlich stärker den Patienten als Individuum mit einbeziehen müssen, da die Datenanalyse interindividuelle Unterschiede identifizierte und Hinweise lieferte, dass einige Probanden die nachteiligen, metabolischen Auswirkungen einer Hochfettdiät besser ausgleichen konnten als andere.
The collision of bathymetric anomalies, such as oceanic spreading centers, at convergent plate margins can profoundly affect subduction dynamics, magmatism, and the structural and geomorphic evolution of the overriding plate. The Southern Patagonian Andes of South America are a prime example for sustained oceanic ridge collision and the successive formation and widening of an extensive asthenospheric slab window since the Middle Miocene. Several of the predicted upper-plate geologic manifestations of such deep-seated geodynamic processes have been studied in this region, but many topics remain highly debated. One of the main controversial topics is the interpretation of the regional low-temperature thermochronology exhumational record and its relationship with tectonic and/or climate-driven processes, ultimately manifested and recorded in the landscape evolution of the Patagonian Andes. The prominent along-strike variance in the topographic characteristics of the Andes, combined with coupled trends in low-temperature thermochronometer cooling ages have been interpreted in very contrasting ways, considering either purely climatic (i.e. glacial erosion) or geodynamic (slab-window related) controlling factors.
This thesis focuses on two main aspects of these controversial topics. First, based on field observations and bedrock low-temperature thermochronology data, the thesis addresses an existing research gap with respect to the neotectonic activity of the upper plate in response to ridge collision - a mechanism that has been shown to affect the upper plate topography and exhumational patterns in similar tectonic settings. Secondly, the qualitative interpretation of my new and existing thermochronological data from this region is extended by inverse thermal modelling to define thermal histories recorded in the data and evaluate the relative importance of surface vs. geodynamic factors and their possible relationship with the regional cooling record.
My research is centered on the Northern Patagonian Icefield (NPI) region of the Southern Patagonian Andes. This site is located inboard of the present-day location of the Chile Triple Junction - the juncture between the colliding Chile Rise spreading center and the Nazca and Antarctic Plates along the South American convergent margin. As such this study area represents the region of most recent oceanic-ridge collision and associated slab window formation. Importantly, this location also coincides with the abrupt rise in summit elevations and relief characteristics in the Southern Patagonian Andes. Field observations, based on geological, structural and geomorphic mapping, are combined with bedrock apatite (U-Th)/He and apatite fission track (AHe and AFT) cooling ages sampled along elevation transects across the orogen. This new data reveals the existence of hitherto unrecognized neotectonic deformation along the flanks of the range capped by the NPI.
This deformation is associated with the closely spaced oblique collision of successive oceanic-ridge segments in this region over the past 6 Ma. I interpret that this has caused a crustal-scale partitioning of deformation and the decoupling, margin-parallel migration, and localized uplift of a large crustal sliver (the NPI block) along the subduction margin. The location of this uplift coincides with a major increase of summit elevations and relief at the northern edge of the NPI massif. This mechanism is compatible with possible extensional processes along the topographically subdued trailing edge of the NPI block as documented by very recent and possibly still active normal faulting. Taken together, these findings suggest a major structural control on short-wavelength variations in topography in the Southern Patagonian Andes - the region affected by ridge collision and slab window formation.
The second research topic addressed here focuses on using my new and existing bedrock low-temperature cooling ages in forward and inverse thermal modeling. The data was implemented in the HeFTy and QTQt modeling platforms to constrain the late Cenozoic thermal history of the Southern Patagonian Andes in the region of the most recent upper-plate sectors of ridge collision. The data set combines AHe and AFT data from three elevation transects in the region of the Northern Patagonian Icefield. Previous similar studies claimed far-reaching thermal effects of the approaching ridge collision and slab window to affect patterns of Late Miocene reheating in the modelled thermal histories. In contrast, my results show that the currently available data can be explained with a simpler thermal history than previously proposed. Accordingly, a reheating event is not needed to reproduce the observations. Instead, the analyzed ensemble of modelled thermal histories defines a Late Miocene protracted cooling and Pliocene-to-recent stepwise exhumation. These findings agree with the geological record of this region. Specifically, this record indicates an Early Miocene phase of active mountain building associated with surface uplift and an active fold-and-thrust belt, followed by a period of stagnating deformation, peneplanation, and lack of synorogenic deposition in the Patagonian foreland. The subsequent period of stepwise exhumation likely resulted from a combination of pulsed glacial erosion and coeval neotectonic activity. The differences between the present and previously published interpretation of the cooling record can be reconciled with important inconsistencies of previously used model setup. These include mainly the insufficient convergence of the models and improper assumptions regarding the geothermal conditions in the region. This analysis puts a methodological emphasis on the prime importance of the model setup and the need for its thorough examination to evaluate the robustness of the final outcome.
Savannas cover a broad geographical range across continents and are a biome best described by a mix of herbaceous and woody plants. The former create a more or less continuous layer while the latter should be sparse enough to leave an open canopy. What has long intrigued ecologists is how these two competing plant life forms of vegetation coexist.
Initially attributed to resource competition, coexistence was considered the stable outcome of a root niche differentiation between trees and grasses. The importance of environmental factors became evident later, when data from moister environments demonstrated that tree cover was often lower than what the rainfall conditions would allow for. Our current understanding relies on the interaction of competition and disturbances in space and time. Hence, the influence of grazing and fire and the corresponding feedbacks they generate have been keenly investigated. Grazing removes grass cover, initiating a self-reinforcing process propagating tree cover expansion. This is known as the encroachment phenomenon. Fire, on the other hand, imposes a bottleneck on the tree population by halting the recruitment of young trees into adulthood. Since grasses fuel fires, a feedback linking grazing, grass cover, fire, and tree cover is created. In African savannas, which are the focus of this dissertation, these feedbacks play a major role in the dynamics.
The importance of these feedbacks came into sharp focus when the notion of alternative states began to be applied to savannas. Alternative states in ecology arise when different states of an ecosystem can occur under the same conditions. According to this an open savanna and a tree-dominated savanna can be classified as alternative states, since they can both occur under the same climatic conditions. The aforementioned feedbacks are critical in the creation of alternative states. The grass-fire feedback can preserve an open canopy as long as fire intensity and frequency remain above a certain threshold. Conversely, crossing a grazing threshold can force an open savanna to shift to a tree-dominated state. Critically, transitions between such alternative states can produce hysteresis, where a return to pre-transition conditions will not suffice to restore the ecosystem to its original state.
In the chapters that follow, I will cover aspects relating to the coexistence mechanisms and the role of feedbacks in tree-grass interactions. Coming back to the coexistence question, due to the overwhelming focus on competition and disturbance another important ecological process was neglected: facilitation. Therefore, in the first study within this dissertation I examine how facilitation can expand the tree-grass coexistence range into drier conditions. For the second study I focus on another aspect of savanna dynamics which remains underrepresented in the literature: the impacts of inter-annual rainfall variability upon savanna trees and the resilience of the savanna state. In the third and final study within this dissertation I approach the well-researched encroachment phenomenon from a new perspective: I search for an early warning indicator of the process to be used as a prevention tool for savanna conservation. In order to perform all this work I developed a mathematical ecohydrological model of Ordinary Differential Equations (ODEs) with three variables: soil moisture content, grass cover and tree cover.
Facilitation: Results showed that the removal of grass cover through grazing was detrimental to trees under arid conditions, contrary to expectation based on resource competition. The reason was that grasses preserved moisture in the soil through infiltration and shading, thus ameliorating the harsh conditions for trees in accordance with the Stress Gradient Hypothesis. The exclusion of grasses from the model further demonstrated this: tree cover was lower in the absence of grasses, indicating that the benefits of grass facilitation outweighed the costs of grass competition for trees. Thus, facilitation expanded the climatic range where savannas persisted into drier conditions.
Rainfall variability: By adjusting the model to current rainfall patterns in East Africa, I simulated conditions of increasing inter-annual rainfall variability for two distinct mean rainfall scenarios: semi-arid and mesic. Alternative states of tree-less grassland and tree-dominated savanna emerged in both cases. Increasing variability reduced semi-arid savanna tree cover to the point that at high variability the savanna state was eliminated, because variability intensified resource competition and strengthened the fire disturbance during high rainfall years. Mesic savannas, on the other hand, became more resilient along the variability gradient: increasing rainfall variability created more opportunities for the rapid growth of trees to overcome the fire disturbance, boosting the chances of savannas persisting and thus increasing mesic savanna resilience.
Preventing encroachment: The breakdown in the grass-fire feedback caused by heavy grazing promoted the expansion of woody cover. This could be irreversible due to the presence of alternative states of encroached and open savanna, which I found along a simulated grazing gradient. When I simulated different short term heavy grazing treatments followed by a reduction to the original grazing conditions, certain cases converged to the encroached state. Utilising woody cover changes only during the heavy grazing treatment, I developed an early warning indicator which identified these cases with a high risk of such hysteresis and successfully distinguished them from those with a low risk. Furthermore, after validating the indicator on encroachment data, I demonstrated that it appeared early enough for encroachment to be prevented through realistic grazing-reduction treatments.
Though this dissertation is rooted in the theory of savanna dynamics, its results can have significant applications in savanna conservation. Facilitation has only recently become a topic of interest within savanna literature. Given the threat of increasing droughts and a general anticipation of drier conditions in parts of Africa, insights stemming from this research may provide clues for preserving arid savannas. The impacts of rainfall variability on savannas have not yet been thoroughly studied, either. Conflicting results appear as a result of the lack of a robust theoretical understanding of plant interactions under variable conditions. . My work and other recent studies argue that such conditions may increase the importance of fast resource acquisition creating a ‘temporal niche’. Woody encroachment has been extensively studied as phenomenon, though not from the perspective of its early identification and prevention. The development of an encroachment forecasting tool, as the one presented in this work, could protect both the savanna biome and societies dependent upon it for (economic) survival. All studies which follow are bound by the attempt to broaden the horizons of savanna-related research in order to deal with extreme conditions and phenomena; be it through the enhancement of the coexistence debate or the study of an imminent external threat or the development of a management-oriented tool for the conservation of savannas.
Molekulare Charakterisierung des Centrosom-assoziierten Proteins CP91 in Dictyostelium discoideum
(2016)
Das Dictyostelium-Centrosom ist ein Modell für acentrioläre Centrosomen. Es besteht aus einer dreischichtigen Kernstruktur und ist von einer Corona umgeben, welche Nukleationskomplexe für Mikrotubuli beinhaltet. Die Verdoppelung der Kernstruktur wird einmal pro Zellzyklus am Übergang der G2 zur M-Phase gestartet. Durch eine Proteomanalyse isolierter Centrosomen konnte CP91 identifiziert werden, ein 91 kDa großes Coiled-Coil Protein, das in der centrosomalen Kernstruktur lokalisiert. GFP-CP91 zeigte fast keine Mobilität in FRAP-Experimenten während der Interphase, was darauf hindeutet, dass es sich bei CP91 um eine Strukturkomponente des Centrosoms handelt. In der Mitose hingegen dissoziieren das GFP-CP91 als auch das endogene CP91 ab und fehlen an den Spindelpolen von der späten Prophase bis zur Anaphase. Dieses Verhalten korreliert mit dem Verschwinden der zentralen Schicht der Kernstruktur zu Beginn der Centrosomenverdopplung. Somit ist CP91 mit großer Wahrscheinlichkeit ein Bestandteil dieser Schicht. CP91-Fragmente der N-terminalen bzw. C-terminalen Domäne (GFP-CP91 N-Terminus, GFP-CP91 C-Terminus) lokalisieren als GFP-Fusionsproteine exprimiert auch am Centrosom, zeigen aber nicht die gleiche mitotische Verteilung des Volllängenproteins. Das CP91-Fragment der zentralen Coiled-Coil Domäne (GFP-CP91cc) lokalisiert als GFP-Fusionsprotein exprimiert, als ein diffuser cytosolische Cluster, in der Nähe des Centrosoms. Es zeigt eine partiell ähnliche mitotische Verteilung wie das Volllängenprotein. Dies lässt eine regulatorische Domäne innerhalb der Coiled-Coil Domäne vermuten. Die Expression der GFP-Fusionsproteine unterdrückt die Expression des endogenen CP91 und bringt überzählige Centrosomen hervor. Dies war auch eine markante Eigenschaft nach der Unterexpression von CP91 durch RNAi. Zusätzlich zeigte sich in CP91-RNAi Zellen eine stark erhöhte Ploidie verursacht durch schwere Defekte in der Chromosomensegregation verbunden mit einer erhöhten Zellgröße und Defekten im Abschnürungsprozess während der Cytokinese. Die Unterexpression von CP91 durch RNAi hatte auch einen direkten Einfluss auf die Menge an den centrosomalen Proteinen CP39, CP55 und CEP192 und dem Centromerprotein Cenp68 in der Interphase. Die Ergebnisse deuten darauf hin, dass CP91 eine zentrale centrosomale Kernkomponente ist und für den Zusammenhalt der beiden äußeren Schichten der Kernstruktur benötigt wird. Zudem spielt CP91 eine wichtige Rolle für eine ordnungsgemäße Centrosomenbiogenese und, unabhängig davon, bei dem Abschnürungsprozess der Tochterzellen während der Cytokinese.
Background: Aggression is a severe behavioral problem that interferes with many developmental challenges individuals face in middle childhood and adolescence. Particularly in the peer and in the academic domain, aggression inhibits the individual from making important learning experiences that are predictive for a healthy transition into adulthood. Furthermore, the resulting developmental deficits have the propensity to feedback and to promote aggression at later developmental stages. The aim of the present PhD thesis was to investigate pathways and processes involved in the etiology of aggression by examining the interrelation between multiple developmental problems in the peer and in the academic domain. More specifically, the relevance of affiliation with deviant peers as a driving mechanism for the development of aggression, factors promoting the affiliation with deviant peers (social rejection; academic failure), and mechanisms by which affiliation with deviant peers leads to aggression (external locus of control) were investigated.
Method: The research questions were addressed by three studies. Three data waves were available for the first study, the second and third study were based on two data waves. The first study specified pathways to antisocial behavior by investigating the temporal interrelation between social rejection, academic failure, and affiliation with deviant peers in a sample of 1,657 male and female children and adolescents aged between 6 and 15 years. The second study examined the role of external control beliefs as a potential mediator in the link between affiliation with deviant peers and aggression in a sample of 1,466 children and adolescents in the age of 9 to 19 years, employing a half-longitudinal design. The third study aimed to expand the findings of Study 1 and Study 2 by examining the differential predictivity of combinations of developmental risks for different functions of aggression, using a sample of 1,479 participants in the age between 9 and 19 years. First, profiles of social rejection, academic failure, and affiliation with deviant peers were identified, using latent profile analysis. Second, prospective pathways between risk-profiles and reactive and proactive aggression were investigated, using latent path analysis.
Results: The first study revealed that antisocial behavior at T1 was associated with social rejection and academic failure at T2. Both mechanisms promoted affiliation with deviant peers at the same data wave, which predicted deviancy at T3. Furthermore, both an indirect pathway via social rejection and affiliation with deviant peers and an indirect pathway via academic failure and affiliation with deviant peers significantly mediated the link between antisocial behavior at the first and the third data wave. Additionally, the proposed pathways generalized across genders and different age groups. The second study yielded that external control beliefs significantly mediated the link between affiliation with deviant peers and aggression, with affiliation with deviant peers at T1 predicting external control beliefs at T2 and external control beliefs at T1 predicting aggressive behavior at T2. Again, the analyses provided no evidence for gender and age specific variations in the proposed pathways. In the third study, three distinct risk groups were identified, made up of a large non-risk group, with low scores on all risk measures, a group characterized by high scores on social rejection (SR group), and a group with the highest scores on measures of affiliation with deviant peers and academic failure (APAF group). Importantly, risk group membership was differentially associated with reactive and proactive aggression. Only membership in the SR group at T1 was associated with the development of reactive aggression at T2 and only membership in the APAF group at T1 predicted proactive aggression at T2. Additionally, proactive aggression at T1 predicted membership in the APAF group at T2, indicating a reciprocal relationship between both constructs.
Conclusion: The results demonstrated that aggression causes severe behavioral deficits in social and academic domains which promote future aggression by increasing individuals’ tendency to affiliate with deviant peers. The stimulation of external control beliefs provides an explanation for deviant peers’ effect on the progression and intensification of aggression. Finally, multiple developmental risks were shown to co-occur within individuals and to be differentially predictive of reactive and proactive aggression. The findings of this doctoral dissertation have possible implications for the conceptualization of prevention and intervention programs aimed to reduce aggression in middle childhood and adolescence.
A majority of studies documented a reduced ankle muscle activity, particularly of the peroneus longus muscle (PL), in patients with functional ankle instability (FI). It is considered valid that foot orthoses as well as sensorimotor training have a positive effect on ankle muscle activity in healthy individuals and those with lower limb overuse injuries or flat arched feet (reduced reaction time by sensorimotor exercises; increased ankle muscle amplitude by orthoses use). However, the acute- and long-term influence of foot orthoses on ankle muscle activity in individuals with FI is unknown.
AIMS: The present thesis addressed (1a) acute- and (1b) long-term effects of foot orthoses compared to sensorimotor training on ankle muscle activity in patients with FI. (2) Further, it was investigated if the orthosis intervention group demonstrate higher ankle muscle activity by additional short-term use of a measurement in-shoe orthosis (compared to short-term use of “shoe only”) after intervention. (3) As prerequisite, it was evaluated if ankle muscle activity can be tested reliably and (4) if this differs between healthy individuals and those with FI.
METHODS: Three intervention groups (orthosis group [OG], sensorimotor training group [SMTG], control group [CG]), consisting of both, healthy individuals and those with FI, underwent one longitudinal investigation (randomised controlled trial). Throughout 6 weeks of intervention, OG wore an in-shoe orthosis with a specific “PL stimulation module”, whereas SMTG conducted home-based exercises. CG served to measure test-retest reliability of ankle muscle activity (PL, M. tibialis anterior [TA] and M. gastrocnemius medialis [GM]). Pre- and post-intervention, ankle muscle activity (EMG amplitude) was recorded during “normal” unperturbed (NW) and perturbed walking (PW) on a split-belt treadmill (stimulus 200 ms post initial heel contact [IC]) as well as during side cutting (SC), each while wearing “shoes only” and additional measurement in-shoe orthoses (randomized order). Normalized RMS values (100% MVC, mean±SD) were calculated pre- (100-50 ms) and post (200-400 ms) - IC.
RESULTS: (3) Test-retest reliability showed a high range of values in healthy individuals and those with FI. (4) Compared to healthy individuals, patients with FI demonstrated lower PL pre-activity during SC, however higher PL pre-activity for NW and PW. (1a) Acute orthoses use did not influence ankle muscle activity. (1b) For most conditions, sensorimotor training was more effective in individuals with FI than long-term orthotic intervention (increased: PL and GM pre-activity and TA reflex-activity for NW, PL pre-activity and TA, PL and GM reflex-activity for SC, PL reflex-activity for PW). However, prolonged orthoses use was more beneficial in terms of an increase in GM pre-activity during SC. For some conditions, long-term orthoses intervention was as effective as sensorimotor training for individuals with FI (increased: PL pre-activity for PW, TA pre-activity for SC, PL and GM reflex-activity for NW). Prolonged orthoses use was also advantageous in healthy individuals (increased: PL and GM pre-activity for NW and PW, PL pre-activity for SC, TA and PL reflex-activity for NW, PL and GM reflex-activity for PW). (2) The orthosis intervention group did not present higher ankle muscle activity by the additional short-term use of a measurement in-shoe orthosis at re-test after intervention.
CONCLUSION: High variations of reproducibility reflect physiological variability in muscle activity during gait and therefore deemed acceptable. The main findings confirm the presence of sensorimotor long-term effects of specific foot orthoses in healthy individuals (primary preventive effect) and those with FI (therapeutic effect). Neuromuscular compensatory feedback- as well as anticipatory feedforward adaptation mechanism to prolonged orthoses use, specifically of the PL muscle, underpins the key role of PL in providing essential dynamic ankle joint stability. Due to its advantages over sensorimotor training (positive subjective feedback in terms of comfort, time-and-cost-effectiveness), long-term foot orthoses use can be recommended as an applicable therapy alternative in the treatment of FI. Long-term effect of foot orthoses in a population with FI must be validated in a larger sample size with longer follow-up periods to substantiate the generalizability of the existing outcomes.
Die vorgelegte Dissertation präsentiert wissenschaftliche Ergebnisse, die in der Zeit vom Dezember 2012 bis August 2016, erarbeitet wurden. Der zentrale Inhalt der Arbeit ist die Simulation von Röntgenabsorptionsprozessen von verschiedenen Systemen in kondensierter Phase. Genauer gesagt, werden Nahkantenabsorptions- (NEXAFS) sowie Röntgenphotoelektronenspektren (XPS) berechnet. In beiden Fällen wird ein Röntgenphoton von einem molekularen System absorbiert. Aufgrund der hohen Photonenenergie wird ein stark gebundenes kernnahes Elektron angeregt. Bei der XPS gelangt dieses mit einer zu messenden kinetischen Energie in Kontinuumszustände. In Abhängigkeit der eingestrahlten Photonenenergie und der kinetischen Energie des austreten Elektrons, kann die Bindungsenergie berechnet werden, welche die zentrale Größe der XPS ist. Im Falle der NEXAFS-Spektroskopie wird das kernnahe Elektron in unbesetzte gebundene Zustände angeregt. Die zentrale Größe ist die Absorption als Funktion der eingestrahlten Photonenenergie. Das erste Kapitel meiner Arbeit erörtert detailliert die experimentellen Methoden sowie die daraus gewonnenen charakteristischen Größen.
Die experimentellen Spektren zeigen oft viele Resonanzen, deren Interpretation aufgrund fehlender Referenzmaterialien schwierig ist. In solchen Fällen bietet es sich an, die Spektren mittels quantenchemischer Methoden zu simulieren. Der dafür erforderliche mathematisch-physikalische Methodenkatalog wird im zweiten Kapitel der Arbeit erörtert.
Das erste von mir untersuchte System ist Graphen. In experimentellen Arbeiten wurde die Oberfläche mittels Bromplasma modifiziert. Die im Anschluss gemessenen NEXAFS-Spektren unterscheiden sich maßgeblich von den Spektren der unbehandelten Oberfläche. Mithilfe periodischer DFT-Rechnungen wurden verschiedene Gitterdefekte sowie bromierte Systeme untersucht und die NEXAFS-Spektren simuliert. Mittels der Simulationen können die Beiträge verschiedener Anregungszentren analysiert werden. Die Berechnungen erlauben den Schluss, dass Gitterdefekte maßgeblich für die entstandenen Veränderungen verantwortlich sind.
Polyvinylalkohol (PVA) wurde als zweites System behandelt. Hierbei sollte untersucht werden, wie groß der Einfluss der Molekularbewegung auf die Verbreiterung der Peaks im XP-Spektrum ist. Des Weiteren wurde untersucht, wie groß der Einfluss von intermolekularen Wechselwirkungen auf die Peakpositionen und Peakverbreiterung ist. Für die Berechnung dieses Systems wurde eine Kombination aus molekulardynamischen und quantenchemischen Methoden verwendet. Als Strukturen dienten Oligomermodelle, die unter dem Einfluss eines (ab initio) Potentials propagiert wurden. Entlang der erstellten Trajektorie wurden Schnappschüsse der Geometrien extrahiert und für die Berechnung der XP-Spektren verwendet. Die Spektren werden bereits mithilfe klassischer Molekulardynamik sehr gut reproduziert. Die erhaltenen Peakbreiten sind verglichen mit dem Experiment allerdings zu klein. Die Hauptursache der Peakverbreiterung ist die Molekularbewegung. Intermolekulare Wechselwirkungen verschieben die Peakpositionen um 0.6 eV zu kleineren Anregungsenergien.
Im dritten Teil der Arbeit stehen die NEXAFS-Spektren von ionischen Flüssigkeiten (ILs) im Fokus. Die experimentell gefundenen Spektren zeigen eine komplexe Struktur mit vielen Resonanzen. In der Arbeit wurden zwei ILs untersucht. Als Geometrien verwenden wir Clustermodelle, die aus experimentellen Kristallstrukturen extrahiert wurden. Die berechneten Spektren erlauben es, die Resonanzen den Anregungszentren zuzuordnen. Außerdem kann eine erstmals gemessene Doppelresonanz simuliert und erklärt werden. Insgesamt kann die Interpretation der Spektren mithilfe der Simulation signifikant erweitert werden.
In allen Systemen wurde zur Berechnung des NEXAFS-Spektrums eine auf Dichtefunktionaltheorie basierende Methode verwendet (die sogenannte Transition-Potential Methode). Gängige wellenfunktionsbasierte Methoden, wie die Konfigurationswechselwirkung mit Einfachanregungen (CIS), zeigen eine starke Blauverschiebung, wenn als Referenz eine Hartree-Fock Slaterdeterminante verwendet wird. Wir zeigen, dass die Verwendung von kernnah-angeregten Determinanten sowohl das resultierende Spektrum als auch die Anregungsenergien deutlich verbessert. Des Weiteren werden auch Referenzen aus Dichtefunktionalrechnungen getestet. Zusätzlich werden auch Referenzen mit gebrochenen Besetzungszahlen für kernnahe Elektronen verwendet. In der Arbeit werden die Resultate der verschiedenen Referenzen miteinander verglichen. Es zeigt sich, dass Referenzen mit gebrochenen Besetzungszahlen das Spektrum nicht weiter verbessern. Der Einfluss der verwendeten Elektronenstrukturmethode ist eher gering.
Surface-enhanced Raman scattering (SERS) is a promising tool to obtain rich chemical information about analytes at trace levels. However, in order to perform selective experiments on individual molecules, two fundamental requirements have to be fulfilled. On the one hand, areas with high local field enhancement, so-called “hot spots”, have to be created by positioning the supporting metal surfaces in close proximity to each other. In most cases hot spots are formed in the gap between adjacent metal nanoparticles (NPs). On the other hand, the analyte has to be positioned directly in the hot spot in order to profit from the highest signal amplification. The use of DNA origami substrates provides both, the arrangement of AuNPs with nm precision as well as the ability to bind analyte molecules at predefined positions. Consequently, the present cumulative doctoral thesis aims at the development of a novel SERS substrate based on a DNA origami template. To this end, two DNA-functionalized gold nanoparticles (AuNPs) are attached to one DNA origami substrate resulting in the formation of a AuNP dimer and thus in a hot spot within the corresponding gap. The obtained structures are characterized by correlated atomic force microscopy (AFM) and SERS imaging which allows for the combination of structural and chemical information.
Initially, the proof-of principle is presented which demonstrates the potential of the novel approach. It is shown that the Raman signal of 15 nm AuNPs coated with dye-modified DNA
(dye: carboxytetramethylrhodamine (TAMRA)) is significantly higher for AuNP dimers arranged on a DNA origami platform in comparison to single AuNPs. Furthermore, by attaching single TAMRA molecules in the hot spot between two 5 nm AuNPs and optimizing the size of the AuNPs by electroless gold deposition, SERS experiments at the few-molecule level are presented. The initially used DNA origami-AuNPs design is further optimized in many respects. On the one hand, larger AuNPs up to a diameter of 60 nm are used which are additionally treated with a silver enhancement solution to obtain Au-Ag-core-shell NPs. On the other hand, the arrangement of both AuNPs is altered to improve the position of the dye molecule within the hot spot as well as to decrease the gap size between the two particles. With the optimized design the detection of single dye molecules (TAMRA and cyanine 3 (Cy3)) by means of SERS is demonstrated. Quantitatively, enhancement factors up to 10^10 are estimated which is sufficiently high to detect single dye molecules.
In the second part, the influence of graphene as an additional component of the SERS substrate is investigated. Graphene is a two-dimensional material with an outstanding combination of electronical, mechanical and optical properties. Here, it is demonstrated that
single layer graphene (SLG) replicates the shape of underlying non-modified DNA origami
substrates very well, which enables the monitoring of structural alterations by AFM imaging.
In this way, it is shown that graphene encapsulation significantly increases the structural
stability of bare DNA origami substrates towards mechanical force and prolonged exposure
to deionized water.
Furthermore, SLG is used to cover DNA origami substrates which are functionalized with a
40 nm AuNP dimer. In this way, a novel kind of hybrid material is created which exhibits
several advantages compared to the analogue non-covered SERS substrates. First, the fluorescence background of dye molecules that are located in between the AuNP surface and SLG is efficiently reduced. Second, the photobleaching rate of the incorporated dye molecules is decreased up to one order of magnitude. Third, due to the increased photostability of the investigated dye molecules, the performance of polarization-dependent series measurements on individual structures is enabled. This in turn reveals extensive information about the dye molecules in the hot spot as well as about the strain induced within the graphene lattice.
Although SLG can significantly influence the SERS substrate in the aforementioned ways, all
those effects are strongly related to the extent of contact with the underlying AuNP dimer.
Physikalische Hydrogele gewinnen derzeit als Zellsubstrate zunehmend an Interesse, da Viskoelastizität oder Stressrelaxation ein bedeutender Parameter in der Mechanotransduktion ist, der bisher vernachlässigt wurde. In dieser Arbeit wurden multi-funktionelle Polyurethane entworfen, die über einen neuartigen Gelierungsmechanismus physikalische Hydrogele bilden. In Wasser bilden die anionischen Polyurethane spontan Aggregate, welche durch elektrostatische Abstoßung in Lösung gehalten werden. Eine schnelle Gelierung kann von hier aus durch Ladungsabschirmung erreicht werden, wodurch die Aggregation voranschreitet und ein Netzwerk ausgebildet wird. Dies kann durch die Zugabe von verschiedenen Säuren oder Salzen geschehen, sodass sowohl saure (pH 4 - 5) als auch pH-neutrale Hydrogele erhalten werden können. Während konventionelle Hydrogele auf Polyurethan-Basis in der Regel durch toxische isocyanat-haltige Präpolymere hergestellt werden, eignet sich der hier beschriebene physikalische Gelierungsmechanismus für in situ Anwendungen in sensitiven Umgebungen. Sowohl Härte als auch Stressrelaxation der Hydrogele können unabhängig voneinander über einen breiten Bereich eingestellt werden. Darüberhinaus zeichnen sich die Hydrogele durch exzellente Stressregeneration aus.
Rapidly uplifting coastlines are frequently associated with convergent tectonic boundaries, like subduction zones, which are repeatedly breached by giant megathrust earthquakes. The coastal relief along tectonically active realms is shaped by the effect of sea-level variations and heterogeneous patterns of permanent tectonic deformation, which are accumulated through several cycles of megathrust earthquakes. However, the correlation between earthquake deformation patterns and the sustained long-term segmentation of forearcs, particularly in Chile, remains poorly understood. Furthermore, the methods used to estimate permanent deformation from geomorphic markers, like marine terraces, have remained qualitative and are based on unrepeatable methods. This contrasts with the increasing resolution of digital elevation models, such as Light Detection and Ranging (LiDAR) and high-resolution bathymetric surveys.
Throughout this thesis I study permanent deformation in a holistic manner: from the methods to assess deformation rates, to the processes involved in its accumulation. My research focuses particularly on two aspects: Developing methodologies to assess permanent deformation using marine terraces, and comparing permanent deformation with seismic cycle deformation patterns under different spatial scales along the M8.8 Maule earthquake (2010) rupture zone. Two methods are developed to determine deformation rates from wave-built and wave-cut terraces respectively. I selected an archetypal example of a wave-built terrace at Santa Maria Island studying its stratigraphy and recognizing sequences of reoccupation events tied with eleven radiocarbon sample ages (14C ages). I developed a method to link patterns of reoccupation with sea-level proxies by iterating relative sea level curves for a range of uplift rates. I find the best fit between relative sea-level and the stratigraphic patterns for an uplift rate of 1.5 +- 0.3 m/ka.
A Graphical User Interface named TerraceM® was developed in Matlab®. This novel software tool determines shoreline angles in wave-cut terraces under different geomorphic scenarios. To validate the methods, I select test sites in areas of available high-resolution LiDAR topography along the Maule earthquake rupture zone and in California, USA. The software allows determining the 3D location of the shoreline angle, which is a proxy for the estimation of permanent deformation rates. The method is based on linear interpolations to define the paleo platform and cliff on swath profiles. The shoreline angle is then located by intersecting these interpolations. The
accuracy and precision of TerraceM® was tested by comparing its results with previous assessments, and through an experiment with students in a computer lab setting at the University
of Potsdam.
I combined the methods developed to analyze wave-built and wave-cut terraces to assess regional patterns of permanent deformation along the (2010) Maule earthquake rupture. Wave-built terraces are tied using 12 Infra Red Stimulated luminescence ages (IRSL ages) and shoreline angles in wave-cut terraces are estimated from 170 aligned swath profiles. The comparison of coseismic slip, interseismic coupling, and permanent deformation, leads to three areas of high permanent uplift, terrace warping, and sharp fault offsets. These three areas correlate with regions of high slip and low coupling, as well as with the spatial limit of at least eight historical megathrust ruptures (M8-9.5). I propose that the zones of upwarping at Arauco and Topocalma reflect changes in frictional properties of the megathrust, which result in discrete boundaries for the propagation of mega earthquakes.
To explore the application of geomorphic markers and quantitative morphology in offshore areas I performed a local study of patterns of permanent deformation inferred from hitherto unrecognized drowned shorelines at the Arauco Bay, at the southern part of the (2010) Maule earthquake rupture zone. A multidisciplinary approach, including morphometry, sedimentology, paleontology, 3D morphoscopy, and a landscape Evolution Model is used to recognize, map, and assess local rates and patterns of permanent deformation in submarine environments. Permanent deformation patterns are then reproduced using elastic models to assess deformation rates of an active submarine splay fault defined as Santa Maria Fault System. The best fit suggests a reverse structure with a slip rate of 3.7 m/ka for the last 30 ka. The register of land level changes during the earthquake cycle at Santa Maria Island suggest that most of the deformation may be accrued through splay fault reactivation during mega earthquakes, like the (2010) Maule event. Considering a recurrence time of 150 to 200 years, as determined from historical and geological observations, slip between 0.3 and 0.7 m per event would be required to account for the 3.7 m/ka millennial slip rate. However, if the SMFS slips only every ~1000 years, representing a few megathrust earthquakes, then a slip of ~3.5 m per event would be required to account for the long- term rate. Such event would be equivalent to a magnitude ~6.7 earthquake capable to generate a local tsunami.
The results of this thesis provide novel and fundamental information regarding the amount of permanent deformation accrued in the crust, and the mechanisms responsible for this accumulation at millennial time-scales along the M8.8 Maule earthquake (2010) rupture zone. Furthermore, the results of this thesis highlight the application of quantitative geomorphology and the use of repeatable methods to determine permanent deformation, improve the accuracy of marine terrace assessments, and estimates of vertical deformation rates in tectonically active coastal areas. This is vital information for adequate coastal-hazard assessments and to anticipate realistic earthquake and tsunami scenarios.
Seit der Einführung von Antibiotika in die medizinische Behandlung von bakteriellen Infektionskrankheiten existiert ein Wettlauf zwischen der Evolution von Bakterienresistenzen und der Entwicklung wirksamer Antibiotika. Während bis in die 80er Jahre verstärkt an neuen Antibiotika geforscht wurde, gewinnen multiresistente Keime heute zunehmend die Oberhand. Um einzelne Pathogene erfolgreich nachzuweisen und zu bekämpfen, ist ein grundlegendes Wissen über den Erreger unumgänglich. Bakterielle Proteine, die bei einer Infektion vorrangig vom Immunsystem prozessiert und präsentiert werden, könnten für die Entwicklung von Impfstoffen oder gezielten Therapeutika nützlich sein. Auch für die Diagnostik wären diese immundominanten Proteine interessant. Allerdings herrscht ein Mangel an Wissen über spezifische Antigene vieler pathogener Bakterien, die eine eindeutige Diagnostik eines einzelnen Erregers erlauben würden.
Daher wurden in dieser Arbeit vier verschiedene Humanpathogene mittels Phage Display untersucht: Neisseria gonorrhoeae, Neisseria meningitidis, Borrelia burgdorferi und Clostridium difficile. Hierfür wurden aus der genomischen DNA der vier Erreger Bibliotheken konstruiert und durch wiederholte Selektion und Amplifikation, dem sogenannten Panning, immunogene Proteine isoliert. Für alle Erreger bis auf C. difficile wurden immunogene Proteine aus den jeweiligen Bibliotheken isoliert. Die identifizierten Proteine von N. meningitidis und B. burgdorferi waren größtenteils bekannt, konnten aber in dieser Arbeit durch Phage Display verifiziert werden. Für N. gonorrhoeae wurden 21 potentiell immunogene Oligopeptide isoliert, von denen sechs Proteine als neue zuvor unbeschriebene Proteine mit immunogenem Charakter identifiziert wurden. Von den Phagen-präsentierten Oligopeptide der 21 immunogenen Proteine wurden Epitopmappings mit verschiedenen polyklonalen Antikörpern durchgeführt, um immunogene Bereiche näher zu identifizieren und zu charakterisieren. Bei zehn Proteinen wurden lineare Epitope eindeutig mit drei polyklonalen Antikörpern identifiziert, von fünf weiteren Proteinen waren Epitope mit mindestens einem Antikörper detektierbar. Für eine weitere Charakterisierung der ermittelten Epitope wurden Alaninscans durchgeführt, die eine detaillierte Auskunft über kritische Aminosäuren für die Bindung des Antikörpers an das Epitop geben.
Ausgehend von dem neu identifizierten Protein mit immunogenem Charakter NGO1634 wurden 26 weitere Proteine aufgrund ihrer funktionellen Ähnlichkeit ausgewählt und mithilfe bioinformatischer Analysen auf ihre Eignung zur Entwicklung einer diagnostischen Anwendung analysiert. Durch Ausschluss der meisten Proteine aufgrund ihrer Lokalisation, Membrantopologie oder unspezifischen Proteinsequenz wurden scFv-Antikörper gegen acht Proteine mittels Phage Display generiert und anschließend als scFv-Fc-Fusionsantikörper produziert und charakterisiert.
Die hier identifizierten Proteine und linearen Epitope könnten einen Ansatzpunkt für die Entwicklung einer diagnostischen oder therapeutischen Anwendung bieten. Lineare Epitopsequenzen werden häufig für die Impfstoffentwicklung eingesetzt, sodass vor allem die in dieser Arbeit bestimmten Epitope von Membranproteinen interessante Kandidaten für weitere Untersuchungen in diese Richtung sind. Durch weitere Untersuchungen könnten möglicherweise unbekannte Virulenzfaktoren entdeckt werden, deren Inhibierung einen entscheidenden Einfluss auf Infektionen haben könnten.
Diese Arbeit zu Grunde liegenden Forschung zielte darauf ab, neue schmelzbare Acrylnitril-Copolymere zu entwickeln. Diese sollten im Anschluss über ein Schmelzspinnverfahren zur Chemiefaser geformt und im letzten Schritt zur Carbonfaser konvertiert werden. Zu diesem Zweck wurden zunächst orientierende Untersuchungen an unterschiedlichen Copolymeren des Acrylnitril aus Lösungspolymerisation durchgeführt. Die Untersuchungen zeigten, dass elektrostatische Wechselwirkungen besser als sterische Abschirmung dazu geeignet sind, Schmelzbarkeit unterhalb der Zersetzungstemperatur von Polyacrylnitril zu bewirken. Aus der Vielzahl untersuchter Copolymere stellten sich jene mit Methoxyethylacrylat (MEA) als am effektivsten heraus. Für diese Copolymere wurden sowohl die Copolymerisationsparameter bestimmt als auch die grundlegende Kinetik der Lösungspolymerisation untersucht. Die Copolymere mit MEA wurden über Schmelzspinnen zur Faser umgeformt und diese dann untersucht. Hierbei wurden auch Einflüsse verschiedener Parameter, wie z.B. die der Molmasse, auf die Fasereigenschaften und -herstellung untersucht. Zuletzt wurde ein Heterophasenpolymerisationsverfahren zur Herstellung von Copolymeren aus AN/MEA entwickelt; dadurch konnten die Materialeigenschaften weiter verbessert werden. Zur Unterdrückung der thermoplastischen Eigenschaften der Fasern wurde ein geeignetes Verfahren entwickelt und anschließend die Konversion zu Carbonfasern durchgeführt.
Intermontane valley fills
(2016)
Sedimentary valley fills are a widespread characteristic of mountain belts around the world. They transiently store material over time spans ranging from thousands to millions of years and therefore play an important role in modulating the sediment flux from the orogen to the foreland and to oceanic depocenters. In most cases, their formation can be attributed to specific fluvial conditions, which are closely related to climatic and tectonic processes. Hence, valley-fill deposits constitute valuable archives that offer fundamental insight into landscape evolution, and their study may help to assess the impact of future climate change on sediment dynamics.
In this thesis I analyzed intermontane valley-fill deposits to constrain different aspects of the climatic and tectonic history of mountain belts over multiple timescales. First, I developed a method to estimate the thickness distribution of valley fills using artificial neural networks (ANNs). Based on the assumption of geometrical similarity between exposed and buried parts of the landscape, this novel and highly automated technique allows reconstructing fill thickness and bedrock topography on the scale of catchments to entire mountain belts.
Second, I used the new method for estimating the spatial distribution of post-glacial sediments that are stored in the entire European Alps. A comparison with data from exploratory drillings and from geophysical surveys revealed that the model reproduces the measurements with a root mean squared error (RMSE) of 70m and a coefficient of determination (R2) of 0.81. I used the derived sediment thickness estimates in combination with a model of the Last Glacial Maximum (LGM) icecap to infer the lithospheric response to deglaciation, erosion and deposition, and deduce their relative contribution to the present-day rock-uplift rate. For a range of different lithospheric and upper mantle-material properties, the results suggest that the long-wavelength uplift signal can be explained by glacial isostatic adjustment with a small erosional contribution and a substantial but localized tectonic component exceeding 50% in parts of the Eastern Alps and in the Swiss Rhône Valley. Furthermore, this study reveals the particular importance of deconvolving the potential components of rock uplift when interpreting recent movements along active orogens and how this can be used to constrain physical properties of the Earth’s interior.
In a third study, I used the ANN approach to estimate the sediment thickness of alluviated reaches of the Yarlung Tsangpo River, upstream of the rapidly uplifting Namche Barwa massif. This allowed my colleagues and me to reconstruct the ancient river profile of the Yarlung Tsangpo, and to show that in the past, the river had already been deeply incised into the eastern margin of the Tibetan Plateau. Dating of basal sediments from drill cores that reached the paleo-river bed to 2–2.5 Ma are consistent with mineral cooling ages from the Namche Barwa massif, which indicate initiation of rapid uplift at ~4 Ma. Hence, formation of the Tsangpo gorge and aggradation of the voluminous valley fill was most probably a consequence of rapid uplift of the Namche Barwa massif and thus tectonic activity.
The fourth and last study focuses on the interaction of fluvial and glacial processes at the southeastern edge of the Karakoram. Paleo-ice-extent indicators and remnants of a more than 400-m-thick fluvio-lacustrine valley fill point to blockage of the Shyok River, a main tributary of the upper Indus, by the Siachen Glacier, which is the largest glacier in the Karakoram Range. Field observations and 10Be exposure dating attest to a period of recurring lake formation and outburst flooding during the penultimate glaciation prior to ~110 ka. The interaction of Rivers and Glaciers all along the Karakorum is considered a key factor in landscape evolution and presumably promoted headward erosion of the Indus-Shyok drainage system into the western margin of the Tibetan Plateau.
The results of this thesis highlight the strong influence of glaciation and tectonics on valley-fill formation and how this has affected the evolution of different mountain belts. In the Alps valley-fill deposition influenced the magnitude and pattern of rock uplift since ice retreat approximately 17,000 years ago. Conversely, the analyzed valley fills in the Himalaya are much older and reflect environmental conditions that prevailed at ~110 ka and ~2.5 Ma, respectively. Thus, the newly developed method has proven useful for inferring the role of sedimentary valley-fill deposits in landscape evolution on timescales ranging from 1,000 to 10,000,000 years.
Die Empirie des beginnenden 21. Jahrhunderts weist mehr autoritäre Regime aus als am Ende des 20. Jahrhunderts angenommen. Die gegenwärtige Autoritarismusforschung versucht die Fortdauer dieses Regimetyps in Hinblick auf die politischen Institutionen zu erklären – dabei bleiben politische Akteure, die nicht zum Herrschaftszentrum gehören, außen vor.
Das vorliegende Projekt untersucht die Rolle und Funktion politischer Opposition in autoritären Regimen. Es wird davon ausgegangen, dass sich an der Opposition eine signifikante Charakteristik autoritärer Regime manifestiert. Das akteurszentrierte Projekt ist der qualitativ orientierten Politikwissenschaft zuzurechnen und verknüpft das Autoritarismuskonzept von Juan Linz mit klassischen Ansätzen der Oppositionsforschung und macht diese Theorien für die gegenwärtige Autoritarismusforschung nutzbar.
Die eigens entwickelte elitenorientierte Oppositionstypologie wird am Beispiel Kenias im Zeitraum 1990-2005 angewendet. Die Oppositionsgruppen werden im Institutionengefüge autoritärer Regime verortet und ihr politisches Agieren in den Dimensionen Handlungsstatus, Handlungsüberzeugung und Handlungsstrategie analysiert. Unter Beachtung der historisch gewachsenen regionalen und kulturellen Spezifika wird angenommen, dass generelle, Regionen übergreifende Aussagen zur Opposition in autoritären Regimen getroffen werden können: Kein Oppositionstyp kann allein einen Herrschaftswechsel bewirken. Der Wechsel bzw. die Fortdauer der Herrschaft hängt von der Dominanz bestimmter Oppositionstypen im Oppositionsgeflecht sowie der gleichzeitigen Schwäche anderer Oppositionstypen ab.
Durch die konzeptionelle Beschäftigung mit Opposition sowie deren empirische Erschließung soll ein substantieller Beitrag für die notwendige Debatte um autoritäre Regime im 21. Jahrhundert geleistet werden.
The Earth’s shallow subsurface with sedimentary cover acts as a waveguide to any incoming wavefield. Within the framework of my thesis, I focused on the characterization of this shallow subsurface within tens to few hundreds of meters of sediment cover. I imaged the seismic 1D shear wave velocity (and possibly the 1D compressional wave velocity). This information is not only required for any seismic risk assessment, geotechnical engineering or microzonation activities, but also for exploration and global seismology where site effects are often neglected in seismic waveform modeling.
First, the conventional frequency-wavenumber (f - k) technique is used to derive the dispersion characteristic of the propagating surface waves recorded using distinct arrays of seismometers in 1D and 2D configurations. Further, the cross-correlation technique is applied to seismic array data to estimate the Green’s function between receivers pairs combination assuming one is the source and the other the receiver. With the consideration of a 1D media, the estimated cross-correlation Green’s functions are sorted with interstation distance in a virtual 1D active seismic experiment. The f - k technique is then used to estimate the dispersion curves. This integrated analysis is important for the interpretation of a large bandwidth of the phase velocity dispersion curves and therefore improving the resolution of the estimated 1D Vs profile.
Second, the new theoretical approach based on the Diffuse Field Assumption (DFA) is used for the interpretation of the observed microtremors H/V spectral ratio. The theory is further extended in this research work to include not only the interpretation of the H/V measured at the surface, but also the H/V measured at depths and in marine environments. A modeling and inversion of synthetic H/V spectral ratio curves on simple predefined geological structures shows an almost perfect recovery of the model parameters (mainly Vs and to a lesser extent Vp). These results are obtained after information from a receiver at depth has been considered in the inversion.
Finally, the Rayleigh wave phase velocity information, estimated from array data, and the H/V(z, f) spectral ratio, estimated from a single station data, are combined and inverted for the velocity profile information. Obtained results indicate an improved depth resolution in comparison to estimations using the phase velocity dispersion curves only. The overall estimated sediment thickness is comparable to estimations obtained by inverting the full micortremor H/V spectral ratio.
We study the interplay between analysis on manifolds with singularities and complex analysis and develop new structures of operators based on the Mellin transform and tools for iterating the calculus for higher singularities. We refer to the idea of interpreting boundary value problems (BVPs) in terms of pseudo-differential operators with a principal symbolic hierarchy, taking into account that BVPs are a source of cone and edge operator algebras. The respective cone and edge pseudo-differential algebras in turn are the starting point of higher corner theories. In addition there are deep relationships between corner operators and complex analysis. This will be illustrated by the Mellin symbolic calculus.
Variations in the distribution of mass within an orogen may lead to transient sediment storage, which in turn might affect the state of stress and the level of fault activity. Distinguishing between different forcing mechanisms causing variations of sediment flux and tectonic activity, is therefore one of the most challenging tasks in understanding the spatiotemporal evolution of active mountain belts.
The Himalayan mountain belt is one of the most significant Cenozoic collisional mountain belt, formed due to collision between northward-bound Indian Plate and the Eurasian Plate during the last 55-50 Ma. Ongoing convergence of these two tectonic plates is accommodated by faulting and folding within the Himalayan arc-shaped orogen and the continued lateral and vertical growth of the Tibetan Plateau and mountain belts adjacent to the plateau as well as regions farther north. Growth of the Himalayan orogen is manifested by the development of successive south-vergent thrust systems. These thrust systems divide the orogen into different morphotectonic domains. From north to south these thrusts are the Main Central Thrust (MCT), the Main Boundary Thrust (MBT) and the Main Frontal Thrust (MFT). The growing topography interacts with moisture-bearing monsoonal winds, which results in pronounced gradients in rainfall, weathering, erosion and sediment transport toward the foreland and beyond. However, a fraction of this sediment is trapped and transiently stored within the intermontane valleys or ‘dun’s within the lower-elevation foothills of the range. Improved understanding of the spatiotemporal evolution of these sediment archives could provide a unique opportunity to decipher the triggers of variations in sediment production, delivery and storage in an actively deforming mountain belt and support efforts to test linkages between sediment volumes in intermontane basins and changes in the shallow crustal stress field. As sediment redistribution in mountain belts on timescales of 102-104 years can effect cultural characteristics and infrastructure in the intermontane valleys and may even impact the seismotectonics of a mountain belt, there is a heightened interest in understanding sediment-routing processes and causal relationships between tectonism, climate and topography. It is here at the intersection between tectonic processes and superposed climatic and sedimentary processes in the Himalayan orogenic wedge, where my investigation is focused on. The study area is the intermontane Kangra Basin in the northwestern Sub-Himalaya, because the characteristics of the different Himalayan morphotectonic provinces are well developed, the area is part of a region strongly influenced by monsoonal forcing, and the existence of numerous fluvial terraces provides excellent strain markers to assess deformation processes within the Himalayan orogenic wedge. In addition, being located in front of the Dhauladhar Range the region is characterized by pronounced gradients in past and present-day erosion and sediment processes associated with repeatedly changing climatic conditions. In light of these conditions I analysed climate-driven late Pleistocene-Holocene sediment cycles in this tectonically active region, which may be responsible for triggering the tectonic re-organization within the Himalayan orogenic wedge, leading to out-of-sequence thrusting, at least since early Holocene.
The Kangra Basin is bounded by the MBT and the Sub-Himalayan Jwalamukhi Thrust (JMT) in the north and south, respectively and transiently stores sediments derived from the Dhauladhar Range. The Basin contains ~200-m-thick conglomerates reflecting two distinct aggradation phases; following aggradation, several fluvial terraces were sculpted into these fan deposits. 10Be CRN surface exposure dating of these terrace levels provides an age of 53.4±3.2 ka for the highest-preserved terrace (AF1); subsequently, this surface was incised until ~15 ka, when the second fan (AF2) began to form. AF2 fan aggradation was superseded by episodic Holocene incision, creating at least four terrace levels. We find a correlation between variations in sediment transport and ∂18O records from regions affected by the Indian Summer Monsoon (ISM). During strengthened ISMs sand post-LGM glacial retreat, aggradation occurred in the Kangra Basin, likely due to high sediment flux, whereas periods of a weakened ISM coupled with lower sediment supply coincided with renewed re-incision.
However, the evolution of fluvial terraces along Sub-Himalayan streams in the Kangra sector is also forced by tectonic processes. Back-tilted, folded terraces clearly document tectonic activity of the JMT. Offset of one of the terrace levels indicates a shortening rate of 5.6±0.8 to 7.5±1.0 mm.a-1 over the last ~10 ka. Importantly, my study reveals that late Pleistocene/Holocene out-of-sequence thrusting accommodates 40-60% of the total 14±2 mm.a-1 shortening partitioned throughout the Sub-Himalaya. Importantly, the JMT records shortening at a lower rate over longer timescales hints towards out-of-sequence activity within the Sub-Himalaya. Re-activation of the JMT could be related to changes in the tectonic stress field caused by large-scale sediment removal from the basin. I speculate that the deformation processes of the Sub-Himalaya behave according to the predictions of critical wedge model and assume the following: While >200m of sediment aggradation would trigger foreland-ward propagation of the deformation front, re-incision and removal of most of the stored sediments (nearly 80-85% of the optimum basin-fill) would again create a sub-critical condition of the wedge taper and trigger the retreat of the deformation front.
While tectonism is responsible for the longer-term processes of erosion associated with steepening hillslopes, sediment cycles in this environment are mainly the result of climatic forcing. My new 10Be cosmogenic nuclide exposure dates and a synopsis of previous studies show the late Pleistocene to Holocene alluvial fills and fluvial terraces studied here record periodic fluctuations of sediment supply and transport capacity on timescales of 1000-100000 years. To further evaluate the potential influence of climate change on these fluctuations, I compared the timing of aggradation and incision phases recorded within remnant alluvial fans and terraces with continental climate archives such as speleothems in neighboring regions affected by monsoonal precipitation. Together with previously published OSL ages yielding the timing of aggradation, I find a correlation between variations in sediment transport with oxygen-isotope records from regions affected by the Indian Summer Monsoon (ISM). Accordingly, during periods of increased monsoon intensity (transitions from dry and cold to wet and warm periods – MIS4 to MIS3 and MIS2 to MIS1) (MIS=marine isotope stage) and post-Last Glacial Maximum glacial retreat, aggradation occurred in the Kangra Basin, likely due to high sediment flux. Conversely, periods of weakened monsoon intensity or lower sediment supply coincide with re-incision of the existing basin-fill.
Finally, my study entails part of a low-temperature thermochronology study to assess the youngest exhumation history of the Dhauladhar Range. Zircon helium (ZHe) ages and existing low-temperature data sets (ZHe, apatite fission track (AFT)) across this range, together with 3D thermokinematic modeling (PECUBE) reveals constraints on exhumation and activity of the range-bounding Main Boundary Thrust (MBT) since at least mid-Miocene time. The modeling results indicate mean slip rates on the MBT-fault ramp of ~2 – 3 mm.a-1 since its activation. This has lead to the growth of the >5-km-high frontal Dhauladhar Range and continuous deep-seated exhumation and erosion. The obtained results also provide interesting constraints of deformation patterns and their variation along strike. The results point towards the absence of the time-transient ‘mid-crustal ramp’ in the basal decollement and
duplexing of the Lesser Himalayan sequence, unlike the nearby regions or even the central Nepal domain. A fraction of convergence (~10-15%) is accommodated along the deep-seated MBT-ramp, most likely merging into the MHT. This finding is crucial for a rigorous assessment of the overall level of tectonic activity in the Himalayan morphotectonic provinces as it contradicts recently-published geodetic shortening estimates. In these studies, it has been proposed that the total Himalayan shortening in the NW Himalaya is accommodated within the Sub-Himalaya whereas no tectonic activity is assigned to the MBT.
This thesis is focused on the study and the exact simulation of two classes of real-valued Brownian diffusions: multi-skew Brownian motions with constant drift and Brownian diffusions whose drift admits a finite number of jumps.
The skew Brownian motion was introduced in the sixties by Itô and McKean, who constructed it from the reflected Brownian motion, flipping its excursions from the origin with a given probability. Such a process behaves as the original one except at the point 0, which plays the role of a semipermeable barrier. More generally, a skew diffusion with several semipermeable barriers, called multi-skew diffusion, is a diffusion everywhere except when it reaches one of the barriers, where it is partially reflected with a probability depending on that particular barrier. Clearly, a multi-skew diffusion can be characterized either as solution of a stochastic differential equation involving weighted local times (these terms providing the semi-permeability) or by its infinitesimal generator as Markov process.
In this thesis we first obtain a contour integral representation for the transition semigroup of the multiskew Brownian motion with constant drift, based on a fine analysis of its complex properties. Thanks to this representation we write explicitly the transition densities of the two-skew Brownian motion with constant drift as an infinite series involving, in particular, Gaussian functions and their tails.
Then we propose a new useful application of a generalization of the known rejection sampling method. Recall that this basic algorithm allows to sample from a density as soon as one finds an - easy to sample - instrumental density verifying that the ratio between the goal and the instrumental densities is a bounded function. The generalized rejection sampling method allows to sample exactly from densities for which indeed only an approximation is known. The originality of the algorithm lies in the fact that one finally samples directly from the law without any approximation, except the machine's.
As an application, we sample from the transition density of the two-skew Brownian motion with or without constant drift. The instrumental density is the transition density of the Brownian motion with constant drift, and we provide an useful uniform bound for the ratio of the densities. We also present numerical simulations to study the efficiency of the algorithm.
The second aim of this thesis is to develop an exact simulation algorithm for a Brownian diffusion whose drift admits several jumps. In the literature, so far only the case of a continuous drift (resp. of a drift with one finite jump) was treated. The theoretical method we give allows to deal with any finite number of discontinuities. Then we focus on the case of two jumps, using the transition densities of the two-skew Brownian motion obtained before. Various examples are presented and the efficiency of our approach is discussed.
Understanding the role of natural climate variability under the pressure of human induced changes of climate and landscapes, is crucial to improve future projections and adaption strategies. This doctoral thesis aims to reconstruct Holocene climate and environmental changes in NE Germany based on annually laminated lake sediments. The work contributes to the ICLEA project (Integrated CLimate and Landscape Evolution Analyses). ICLEA intends to compare multiple high-resolution proxy records with independent chronologies from the N central European lowlands, in order to disentangle the impact of climate change and human land use on landscape development during the Lateglacial and Holocene. In this respect, two study sites in NE Germany are investigated in this doctoral project, Lake Tiefer See and palaeolake Wukenfurche. While both sediment records are studied with a combination of high-resolution sediment microfacies and geochemical analyses (e.g. µ-XRF, carbon geochemistry and stable isotopes), detailed proxy understanding mainly focused on the continuous 7.7 m long sediment core from Lake Tiefer See covering the last ~6000 years. Three main objectives are pursued at Lake Tiefer See: (1) to perform a reliable and independent chronology, (2) to establish microfacies and geochemical proxies as indicators for climate and environmental changes, and (3) to trace the effects of climate variability and human activity on sediment deposition.
Addressing the first aim, a reliable chronology of Lake Tiefer See is compiled by using a multiple-dating concept. Varve counting and tephra findings form the chronological framework for the last ~6000 years. The good agreement with independent radiocarbon dates of terrestrial plant remains verifies the robustness of the age model. The resulting reliable and independent chronology of Lake Tiefer See and, additionally, the identification of nine tephras provide a valuable base for detailed comparison and synchronization of the Lake Tiefer See data set with other climate records. The sediment profile of Lake Tiefer See exhibits striking alternations between well-varved and non-varved sediment intervals. The combination of microfacies, geochemical and microfossil (i.e. Cladocera and diatom) analyses indicates that these changes of varve preservation are caused by variations of lake circulation in Lake Tiefer See. An exception is the well-varved sediment deposited since AD 1924, which is mainly influenced by human-induced lake eutrophication. Well-varved intervals before the 20th century are considered to reflect phases of reduced lake circulation and, consequently, stronger anoxic conditions. Instead, non-varved intervals indicate increased lake circulation in Lake Tiefer See, leading to more oxygenated conditions at the lake ground. Furthermore, lake circulation is not only influencing sediment deposition, but also geochemical processes in the lake. As, for example, the proxy meaning of δ13COM varies in time in response to changes of the oxygen regime in the lake hypolinion. During reduced lake circulation and stronger anoxic conditions δ13COM is influenced by microbial carbon cycling. In contrast, organic matter degradation controls δ13COM during phases of intensified lake circulation and more oxygenated conditions. The varve preservation indicates an increasing trend of lake circulation at Lake Tiefer See after ~4000 cal a BP. This trend is superimposed by decadal to centennial scale variability of lake circulation intensity. Comparison to other records in Central Europe suggests that the long-term trend is probably related to gradual changes in Northern Hemisphere orbital forcing, which induced colder and windier conditions in Central Europe and, therefore, reinforced lake circulation. Decadal to centennial scale periods of increased lake circulation coincide with settlement phases at Lake Tiefer See, as inferred from pollen data of the same sediment record. Deforestation reduced the wind shelter of the lake, which probably increased the sensitivity of lake circulation to wind stress. However, results of this thesis also suggest that several of these phases of increased lake circulation are additionally reinforced by climate changes. A first indication is provided by the comparison to the Baltic Sea record, which shows striking correspondence between major non-varved intervals at Lake Tiefer See and bioturbated sediments in the Baltic Sea. Furthermore, a preliminary comparison to the ICLEA study site Lake Czechowskie (N central Poland) shows a coincidence of at least three phases of increased lake circulation in both lakes, which concur with periods of known climate changes (2.8 ka event, ’Migration Period’ and ’Little Ice Age’). These results suggest an additional over-regional climate forcing also on short term increased of lake circulation in Lake Tiefer See.
In summary, the results of this thesis suggest that lake circulation at Lake Tiefer See is driven by a combination of long-term and short-term climate changes as well as of anthropogenic deforestation phases. Furthermore, the lake circulation drives geochemical cycles in the lake affecting the meaning of proxy data. Therefore, the work presented here expands the knowledge of climate and environmental variability in NE Germany. Furthermore, the integration of the Lake Tiefer See multi-proxy record in a regional comparison with another ICLEA side, Lake Czechowskie, enabled to better decipher climate changes and human impact on the lake system. These first results suggest a huge potential for further detailed regional comparisons to better understand palaeoclimate dynamics in N central Europe.
Services that operate over the Internet are under constant threat of being exposed to fraudulent use. Maintaining good user experience for legitimate users often requires the classification of entities as malicious or legitimate in order to initiate countermeasures. As an example, inbound email spam filters decide for spam or non-spam. They can base their decision on both the content of each email as well as on features that summarize prior emails received from the sending server. In general, discriminative classification methods learn to distinguish positive from negative entities. Each decision for a label may be based on features of the entity and related entities. When labels of related entities have strong interdependencies---as can be assumed e.g. for emails being delivered by the same user---classification decisions should not be made independently and dependencies should be modeled in the decision function. This thesis addresses the formulation of discriminative classification problems that are tailored for the specific demands of the following three Internet security applications. Theoretical and algorithmic solutions are devised to protect an email service against flooding of user inboxes, to mitigate abusive usage of outbound email servers, and to protect web servers against distributed denial of service attacks.
In the application of filtering an inbound email stream for unsolicited emails, utilizing features that go beyond each individual email's content can be valuable. Information about each sending mail server can be aggregated over time and may help in identifying unwanted emails. However, while this information will be available to the deployed email filter, some parts of the training data that are compiled by third party providers may not contain this information. The missing features have to be estimated at training time in order to learn a classification model. In this thesis an algorithm is derived that learns a decision function that integrates over a distribution of values for each missing entry. The distribution of missing values is a free parameter that is optimized to learn an optimal decision function.
The outbound stream of emails of an email service provider can be separated by the customer IDs that ask for delivery. All emails that are sent by the same ID in the same period of time are related, both in content and in label. Hijacked customer accounts may send batches of unsolicited emails to other email providers, which in turn might blacklist the sender's email servers after detection of incoming spam emails. The risk of being blocked from further delivery depends on the rate of outgoing unwanted emails and the duration of high spam sending rates. An optimization problem is developed that minimizes the expected cost for the email provider by learning a decision function that assigns a limit on the sending rate to customers based on the each customer's email stream.
Identifying attacking IPs during HTTP-level DDoS attacks allows to block those IPs from further accessing the web servers. DDoS attacks are usually carried out by infected clients that are members of the same botnet and show similar traffic patterns. HTTP-level attacks aim at exhausting one or more resources of the web server infrastructure, such as CPU time. If the joint set of attackers cannot increase resource usage close to the maximum capacity, no effect will be experienced by legitimate users of hosted web sites. However, if the additional load raises the computational burden towards the critical range, user experience will degrade until service may be unavailable altogether. As the loss of missing one attacker depends on block decisions for other attackers---if most other attackers are detected, not blocking one client will likely not be harmful---a structured output model has to be learned. In this thesis an algorithm is developed that learns a structured prediction decoder that searches the space of label assignments, guided by a policy.
Each model is evaluated on real-world data and is compared to reference methods. The results show that modeling each classification problem according to the specific demands of the task improves performance over solutions that do not consider the constraints inherent to an application.
Der Klimawandel
(2016)
Was ist Gerechtigkeit? Wie könnten gerechte Regelungen aussehen für die Katastrophen und Leiden, die der Klimawandel auslöst bzw. auslösen wird? Diese sind häufig ungerecht, weil sie oft deutlich stärker diejenigen treffen, die am wenigsten zur Klimaveränderung beigetragen haben.
Doch was genau verstehen wir unter dem Schlagwort: ‚Klimawandel‘? Und kann dieser wirklich den Menschen direkt treffen? Ein kurzer naturwissenschaftlicher Abriss klärt hier die wichtigsten Fragen.
Da es sich hierbei um eine philosophische Arbeit handelt, muss zunächst geklärt werden, ob der Mensch überhaupt die Ursache von so etwas sein kann wie z.B. der Klimaerwärmung. Robert Spaemanns These dazu ist, dass der Mensch durch seinen freien Willen mit seinen Einzelhandlungen das Weltgeschehen verändern kann. Hans Jonas fügt dem hinzu, dass wir durch diese Fähigkeit, verantwortlich sind für die gewollten und ungewollten Folgen unserer Handlungen.
Damit wäre aus naturwissenschaftlicher Sicht (1. Teil der Arbeit) und aus philosophischer Sicht (Anfang 2. Teil) geklärt, dass der Mensch mit größter Wahrscheinlichkeit die Ursache des Klimawandels ist und diese Verursachung moralische Konsequenzen für ihn hat.
Ein philosophischer Gerechtigkeitsbegriff wird aus der Kantischen Rechts- und Moralphilosophie entwickelt, weil diese die einzige ist, die dem Menschen überhaupt ein Recht auf Rechte zusprechen kann. Diese entspringt der transzendentalen Freiheitsfähigkeit des Menschen, weshalb jedem das Recht auf Rechte absolut und immer zukommt. Gleichzeitig mündet Kants Philosophie wiederum in dem Freiheitsgedanken, indem Gerechtigkeit nur existiert, wenn alle Menschen gleichermaßen frei sein können.
Was heißt das konkret? Wie könnte Gerechtigkeit in der Realität wirklich umgesetzt werden? Die Realisierung schlägt zwei Grundrichtungen ein. John Rawls und Stefan Gosepath beschäftigen sich u.a. eingehend mit der prozeduralen Gerechtigkeit, was bedeutet, dass gerechte Verfahren gefunden werden, die das gesellschaftliche Zusammenleben regeln. Das leitende Prinzip hierfür ist vor allem: ein Mitbestimmungsrecht aller, so dass sich im Prinzip alle Bürger ihre Gesetze selbst geben und damit frei handeln.
In Bezug auf den Klimawandel steht die zweite Ausrichtung im Vordergrund – die distributive oder auch Verteilungs-Gerechtigkeit. Materielle Güter müssen so aufgeteilt werden, dass auch trotz empirischer Unterschiede alle Menschen als moralische Subjekte anerkannt werden und frei sein können.
Doch sind diese philosophischen Schlussfolgerungen nicht viel zu abstrakt, um auf ein ebenso schwer fassbares und globales Problem wie den Klimawandel angewendet zu werden? Was könnte daher eine Klimagerechtigkeit sein?
Es gibt viele Gerechtigkeitsprinzipien, die vorgeben, eine gerechte Grundlage für die Klimaprobleme zu bieten wie z.B. das Verursacherprinzip, das Fähigkeitsprinzip oder das Grandfathering-Prinzip, bei dem die Hauptverursacher nach wie vor am meisten emittieren dürfen (dieses Prinzip leitete die bisherigen internationalen Verhandlungen).
Das Ziel dieser Arbeit ist, herauszufinden, wie die Klimaprobleme gelöst werden können, so dass für alle Menschen unter allen Umständen die universellen Menschenrechte her- und sichergestellt werden und diese frei und moralisch handeln können.
Die Schlussfolgerung dieser Arbeit ist, dass Kants Gerechtigkeitsbegriff durch eine Kombination des Subsistenzemissions-Rechts, des Greenhouse-Development-Rights-Principles (GDR-Prinzip) und einer internationalen Staatlichkeit durchgesetzt werden könnte.
Durch das Subsistenzemissions-Recht hat jeder Mensch das Recht, so viel Energie zu verbrauchen und damit zusammenhängende Emissionen zu produzieren, dass er ein menschenwürdiges Leben führen kann. Das GDR-Prinzip errechnet den Anteil an der weltweiten Gesamtverantwortung zum Klimaschutz eines jeden Landes oder sogar eines jeden Weltbürgers, indem es die historischen Emissionen (Klimaschuld) zu der jetzigen finanziellen Kapazität des Landes/ des Individuums (Verantwortungsfähigkeit) hinzuaddiert. Die Implementierung von internationalen Gremien wird verteidigt, weil es ein globales, grenzüberschreitendes Problem ist, dessen Effekte und dessen Verantwortung globale Ausmaße haben.
Ein schlagendes Argument für fast alle Klimaschutzmaßnahmen ist, dass sie Synergien aufweisen zu anderen gesellschaftlichen Bereichen aufweisen wie z.B. Gesundheit und Armutsbekämpfung, in denen auch noch um die Durchsetzung unserer Menschenrechte gerungen wird.
Ist dieser Lösungsansatz nicht völlig utopisch?
Dieser Vorschlag stellt für die internationale Gemeinschaft eine große Herausforderung dar, wäre jedoch die einzig gerechte Lösung unserer Klimaprobleme. Des Weiteren wird an dem Kantischen Handlungsgrundsatz festgehalten, dass das ewige Streben auf ideale Ziele hin, die beste Verwirklichung dieser durch menschliche, fehlbare Wesen ist.
Protektiver Effekt von 6-Shogaol, Ellagsäure und Myrrhe auf die intestinale epitheliale Barriere
(2016)
Viele bioaktive Pflanzeninhaltsstoffe bzw. Pflanzenmetabolite besitzen antiinflammatorische Eigenschaften. Diese versprechen ein hohes Potential für den Einsatz in der Phytotherapie bzw. Prävention von chronisch-entzündlichen Darmerkrankungen (CED). Eine intestinale Barrieredysfunktion ist ein typisches Charakteristikum von CED Patienten, die dadurch an akuter Diarrhoe leiden.
In dieser Arbeit werden die Pflanzenkomponenten 6-Shogaol, Ellagsäure und Myrrhe an den intestinalen Kolonepithelzellmodellen HT-29/B6 und Caco-2 auf ihr Potential hin, die intestinale Barriere zu stärken bzw. eine Barrieredysfunktion zu verhindern, untersucht. Hauptschwerpunkt der Analysen ist die parazelluläre Barrierefunktion und die Regulation der dafür entscheidenden Proteinfamilie der Tight Junctions (TJs), der Claudine.
Die Barrierefunktion wird durch Messung des transepithelialen Widerstands (TER) und der Fluxmessung in der Ussing-Kammer bestimmt. Dazu werden die HT-29/B6- und Caco-2-Monolayer mit den Pflanzenkomponenten (6-Shogaol, Ellagsäure, Myrrhe), dem proinflammatorischen Zytokin TNF-α oder der Kombination von beiden Subsztanzen für 24 oder 48 h behandelt. Außerdem wurden zur weiteren Charakterisierung die Expression sowie die Lokalisation der für die parazelluläre Barriere relevanten Claudine, die TJ-Ultrastruktur und verschiedene Signalwege analysiert.
In Caco-2-Monolayern führten Ellagsäure und Myrrhe, nicht aber 6-Shogaol, allein zu einem TER-Anstieg bedingt durch eine verringerte Permeabilität für Natriumionen. Myrrhe verminderte die Expression des Kationenkanal-bildenden TJ-Proteins Claudin-2 über die Inhibierung des PI3K/Akt-Signalweges, während Ellagsäure die Expression der TJ-Proteine Claudin-4 und -7 reduzierte. Alle Pflanzenkomponenten schützten in den Caco-2-Zellen vor einer TNF-α-induzierten Barrieredysfunktion.
An den HT-29/B6-Monolayern änderte keine der Pflanzenkomponenten allein die Barrierefunktion. Die HT-29/B6-Zellen reagierten auf TNF-α mit einer deutlichen Verminderung des TER und einer erhöhten Fluoreszein-Permeabilität. Die TER-Abnahme war durch eine PI3K/Akt-vermittelte gesteigerte Claudin-2-Expression sowie eine NFκB-vermittelte Umverteilung des abdichtenden TJ-Proteins Claudin-1 gekennzeichnet. 6-Shogaol konnte den TER-Abfall partiell hemmen sowie die PI3K/Akt-induzierte Claudin-2-Expression und die NFκB-bedingte Claudin-1-Umverteilung verhindern. Ebenso inhibierte Myrrhe, nicht aber Ellagsäure, den TNF-α-induzierten TER-Abfall. Dabei konnte Myrrhe zwar den Claudin-2-Expressionsanstieg und die Claudin-1-Umverteilung unterbinden, jedoch weder die NFκB- noch die PI3K/Akt-Aktivierung hemmen. Diese Arbeit zeigt, dass auch STAT6 an dem Claudin-2-Expressionsanstieg durch
TNF-α in HT-29/B6-Zellen beteiligt ist. So wurde durch Myrrhe die TNF-α-induzierte Phosphorylierung von STAT6 und die erhöhte Claudin-2-Expression inhibiert.
Die Ergebnisse deuten darauf hin, dass die Pflanzenkomponenten 6-Shogaol, Ellagsäure und Myrrhe mit unterschiedlichen Mechanismen stärkend auf die Barriere einwirken. Zur Behandlung von intestinalen Erkrankungen mit Barrieredysfunktion könnten daher Kombinationspräparate aus verschiedenen Pflanzen effektiver sein als Monopräparate.
Gene expression describes the process of making functional gene products (e.g. proteins or special RNAs) from instructions encoded in the genetic information (e.g. DNA). This process is heavily regulated, allowing cells to produce the appropriate gene products necessary for cell survival, adapting production as necessary for different cell environments. Gene expression is subject to regulation at several levels, including transcription, mRNA degradation, translation and protein degradation. When intact, this system maintains cell homeostasis, keeping the cell alive and adaptable to different environments. Malfunction in the system can result in disease states and cell death. In this dissertation, we explore several aspects of gene expression control by analyzing data from biological experiments. Most of the work following uses a common mathematical model framework based on Markov chain models to test hypotheses, predict system dynamics or elucidate network topology. Our work lies in the intersection between mathematics and biology and showcases the power of statistical data analysis and math modeling for validation and discovery of biological phenomena.
Das Ziel der Doktorarbeit war die Entwicklung und Evaluation eines skillsbasierten primären Präventionsprogramms (Mainzer Schultraining zur Essstörungsprävention (MaiStep)) für partielle und manifeste Essstörungen. Dabei wurde die Wirksamkeit durch einen primären (Reduktion vorhandener Essstörungssymptome) und sekundären (assoziierte Psychopathologie) Zielparameter 3 und 12 Monate nach Durchführung des Trainings überprüft. Innerhalb der randomisiert kontrollierten Studie gab es zwei Interventionsgruppe und eine aktive Kontrollgruppe. 1.654 Jugendliche (weiblich/männlich: 781/873; mittleres Alter: 13.1±0.7; BMI: 20.0±3.5) konnten für die Studie, an zufällig ausgewählten Schulen in Rheinland-Pfalz, rekrutiert werden. Die Entwicklung des Präventionsprogramms basiert auf einem systematischen Literaturreview von 63 wissenschaftlichen Studien über die Prävention von Essstörungen im Kindes- und Jugendalter. Eine Interventionsgruppe wurde durch Psychologinnen/Psychologen und eine zweite durch Lehrkräfte angeleitet. Das in der aktiven Kontrollgruppe durchgeführte Sucht- bzw. Stresspräventionsprogramm wurde durch Lehrkräfte geleitet. MaiStep zeigte zur 3-Monatskatamnese keine signifikanten Effekte im Vergleich zur aktiven Kontrollgruppe. Allerdings zeigten sich nach 12 Monaten multiple signifikante Effekte zwischen den Interventions- und der aktiven Kontrollgruppe. Im Rahmen der Analyse des primären Parameters wurden in den Interventionsgruppen signifikant weniger Jugendliche mit einer partiellen Anorexia nervosa (CHI²(2) = 8.74, p = .01**) und/oder partiellen Bulimia nervosa (CHI²(2) = 7.25, p = .02*) gefunden. Im Rahmen der sekundären Zielparameter zeigten sich signifikante Veränderungen in Subskalen des Eating Disorder Inventory (EDI-2) Schlankheitsstreben (F (2, 355) = 3.94, p = .02*) und Perfektionismus (F (2, 355) = 4.19, p = .01**) sowie dem Body Image Avoidance Questionnaire (BIAQ) (F (2, 525) = 18.79, p = .01**) zwischen den Interventions- und der aktiven Kontrollgruppe. MaiStep kann somit als erfolgreiches Programm zur Reduktion von partiellen Essstörungen für die Altersgruppe der 13- 15-jährigen bezeichnet werden. Trotz unterschiedlicher Wirkmechanismen zeigten sich die Lehrkräfte im Vergleich zu den Psychologinnen/Psychologen ebenso erfolgreich in der Durchführung.
Trial registration MaiStep is registered at the German Clinical Trials Register (DRKS00005050).
Das Wissen um die lokale Struktur von Seltenen Erden Elementen (SEE) in silikatischen und aluminosilikatischen Schmelzen ist von fundamentalem Interesse für die Geochemie der magmatischen Prozesse, speziell wenn es um ein umfassendes Verständnis der Verteilungsprozesse von SEE in magmatischen Systemen geht. Es ist allgemein akzeptiert, dass die SEE-Verteilungsprozesse von Temperatur, Druck, Sauerstofffugazität (im Fall von polyvalenten Kationen) und der Kristallchemie kontrolliert werden. Allerdings ist wenig über den Einfluss der Schmelzzusammensetzung selbst bekannt. Ziel dieser Arbeit ist, eine Beziehung zwischen der Variation der SEE-Verteilung mit der Schmelzzusammensetzung und der Koordinationschemie dieser SEE in der Schmelze zu schaffen.
Dazu wurden Schmelzzusammensetzungen von Prowatke und Klemme (2005), welche eine deutliche Änderung der Verteilungskoeffizienten zwischen Titanit und Schmelze ausschließlich als Funktion der Schmelzzusammensetzung zeigen, sowie haplogranitische bzw. haplobasaltische Schmelzzusammensetzungen als Vertreter magmatischer Systeme mit La, Gd, Yb und Y dotiert und als Glas synthetisiert. Die Schmelzen variierten systematisch im Aluminiumsättigungsindex (ASI), welcher bei den Prowatke und Klemme (2005) Zusammensetzungen einen Bereich von 0.115 bis 0.768, bei den haplogranitischen Zusammensetzungen einen Bereich von 0.935 bis 1.785 und bei den haplobasaltischen Zusammensetzungen einen Bereich von 0.368 bis 1.010 abdeckt. Zusätzlich wurden die haplogranitischen Zusammensetzungen mit 4 % H2O synthetisiert, um den Einfluss von Wasser auf die lokale Umgebung von SEE zu studieren. Um Informationen über die lokalen Struktur von Gd, Yb und Y zu erhalten wurde die Röntgenabsorptionsspektroskopie angewendet. Dabei liefert die Untersuchung der Feinstruktur mittels der EXAFS-Spektroskopie (engl. Extended X-Ray Absorption Fine Structure) quantitative Informationen über die lokale Umgebung, während RIXS (engl. resonant inelastic X-ray scattering), sowie die daraus extrahierte hoch aufgelöste Nahkantenstruktur, XANES (engl. X-ray absorption near edge structure) qualitative Informationen über mögliche Koordinationsänderungen von La, Gd und Yb in den Gläsern liefert. Um mögliche Unterschiede der lokalen Struktur oberhalb der Glastransformationstemperatur (TG) zur Raumtemperatur zu untersuchen, wurden exemplarisch Hochtemperatur Y-EXAFS Untersuchungen durchgeführt.
Für die Auswertung der EXAFS-Messungen wurde ein neu eingeführter Histogramm-Fit verwendet, der auch nicht-symmetrische bzw. nichtgaußförmige Paarverteilungsfunktionen beschreiben kann, wie sie bei einem hohen Grad der Polymerisierung bzw. bei hohen Temperaturen auftreten können. Die Y-EXAFS-Spektren für die Prowatke und Klemme (2005) Zusammensetzungen zeigen mit Zunahme des ASI, eine Zunahme der Asymmetrie und Breite der Y-O Paarverteilungsfunktion, welche sich in sich in der Änderung der Koordinationszahl von 6 nach 8 und einer Zunahme des Y-O Abstand um 0.13Å manifestiert. Ein ähnlicher Trend lässt sich auch für die Gd- und Yb-EXAFS-Spektren beobachten. Die hoch aufgelösten XANESSpektren für La, Gd und Yb zeigen, dass sich die strukturellen Unterschiede zumindest halb-quantitativ bestimmen lassen. Dies gilt insbesondere für Änderungen im mittleren Abstand zu den Sauerstoffatomen. Im Vergleich zur EXAFS-Spektroskopie liefert XANES jedoch keine Informationen über die Form und Breite von Paarverteilungsfunktionen. Die Hochtemperatur EXAFS-Untersuchungen von Y zeigen Änderungen der lokalen Struktur oberhalb der Glasübergangstemperatur an, welche sich vordergründig auf eine thermisch induzierte Erhöhung des mittleren Y-O Abstandes zurückführen lassen. Allerdings zeigt ein Vergleich der Y-O Abstände für Zusammensetzungen mit einem ASI von 0.115 bzw. 0.755, ermittelt bei Raumtemperatur und TG, dass der im Glas beobachtete strukturelle Unterschied entlang der Zusammensetzungsserie in der Schmelze noch stärker ausfallen kann, als bisher für die Gläser angenommen wurde.
Die direkte Korrelation der Verteilungsdaten von Prowatke und Klemme (2005) mit den strukturellen Änderungen der Schmelzen offenbart für Y eine lineare Korrelation, wohingegen Yb und Gd eine nicht lineare Beziehung zeigen. Aufgrund seines Ionenradius und seiner Ladung wird das 6-fach koordinierte SEE in den niedriger polymerisierten Schmelzen bevorzugt durch nicht-brückenbildende Sauerstoffatome koordiniert, um stabile Konfigurationen zu bilden. In den höher polymerisierten Schmelzen mit ASI-Werten in der Nähe von 1 ist 6-fache Koordination nicht möglich, da fast nur noch brückenbildende Sauerstoffatome zur Verfügung stehen. Die Überbindung von brückenbildenden Sauerstoffatomen um das SEE wird durch Erhöhung der Koordinationszahl und des mittleren SEE-O Abstandes ausgeglichen. Dies bedeutet eine energetisch günstigere Konfiguration in den stärker depolymerisierten Zusammensetzungen, aus welcher die beobachtete Variation des Verteilungskoeffizienten resultiert, welcher sich jedoch für jedes Element stark unterscheidet. Für die haplogranitischen und haplobasaltischen Zusammensetzungen wurde mit Zunahme der Polymerisierung auch eine Zunahme der Koordinationszahl und des durchschnittlichen Bindungsabstands, einhergehend mit der Zunahme der Schiefe und der Asymmetrie der Paarverteilungsfunktion, beobachtet. Dies impliziert, dass das jeweilige SEE mit Zunahme der Polymerisierung auch inkompatibler in diesen Zusammensetzungen wird. Weiterhin zeigt die Zugabe von Wasser, dass die Schmelzen depolymerisieren, was in einer symmetrischeren Paarverteilungsfunktion resultiert, wodurch die Kompatibilität wieder zunimmt.
Zusammenfassend zeigt sich, dass die Veränderungen der Schmelzzusammensetzungen in einer Änderung der Polymerisierung der Schmelzen resultieren, die dann einen signifikanten Einfluss auf die lokale Umgebung der SEE hat. Die strukturellen Änderungen lassen sich direkt mit Verteilungsdaten korrelieren, die Trends unterscheiden sich aber stark zwischen leichten, mittleren und schweren SEE. Allerdings konnte diese Studie zeigen, in welcher Größenordnung die Änderungen liegen müssen, um einen signifikanten Einfluss auf den Verteilungskoeffizenten zu haben. Weiterhin zeigt sich, dass der Einfluss der Schmelzzusammensetzung auf die Verteilung der Spurenelemente mit Zunahme der Polymerisierung steigt und daher nicht vernachlässigt werden darf.
In complement to the well-established zwitterionic monomers 3-((2-(methacryloyloxy)ethyl)dimethylammonio)propane-1-sulfonate (“SPE”) and 3-((3-methacrylamidopropyl)dimethylammonio)propane-1-sulfonate (“SPP”), the closely related sulfobetaine monomers were synthesized and polymerized by reversible addition-fragmentation chain transfer (RAFT) polymerization, using a fluorophore labeled RAFT agent. The polyzwitterions of systematically varied molar mass were characterized with respect to their solubility in water, deuterated water, and aqueous salt solutions. These poly(sulfobetaine)s show thermoresponsive behavior in water, exhibiting upper critical solution temperatures (UCST). Phase transition temperatures depend notably on the molar mass and polymer concentration, and are much higher in D2O than in H2O. Also, the phase transition temperatures are effectively modulated by the addition of salts. The individual effects can be in parts correlated to the Hofmeister series for the anions studied. Still, they depend in a complex way on the concentration and the nature of the added electrolytes, on the one hand, and on the detailed structure of the zwitterionic side chain, on the other hand. For the polymers with the same zwitterionic side chain, it is found that methacrylamide-based poly(sulfobetaine)s exhibit higher UCST-type transition temperatures than their methacrylate analogs. The extension of the distance between polymerizable unit and zwitterionic groups from 2 to 3 methylene units decreases the UCST-type transition temperatures. Poly(sulfobetaine)s derived from aliphatic esters show higher UCST-type transition temperatures than their analogs featuring cyclic ammonium cations. The UCST-type transition temperatures increase markedly with spacer length separating the cationic and anionic moieties from 3 to 4 methylene units. Thus, apparently small variations of their chemical structure strongly affect the phase behavior of the polyzwitterions in specific aqueous environments.
Water-soluble block copolymers were prepared from the zwitterionic monomers and the non-ionic monomer N-isopropylmethacrylamide (“NIPMAM”) by the RAFT polymerization. Such block copolymers with two hydrophilic blocks exhibit twofold thermoresponsive behavior in water. The poly(sulfobetaine) block shows an UCST, whereas the poly(NIPMAM) block exhibits a lower critical solution temperature (LCST). This constellation induces a structure inversion of the solvophobic aggregate, called “schizophrenic micelle”. Depending on the relative positions of the two different phase transitions, the block copolymer passes through a molecularly dissolved or an insoluble intermediate regime, which can be modulated by the polymer concentration or by the addition of salt. Whereas, at low temperature, the poly(sulfobetaine) block forms polar aggregates that are kept in solution by the poly(NIPMAM) block, at high temperature, the poly(NIPMAM) block forms hydrophobic aggregates that are kept in solution by the poly(sulfobetaine) block. Thus, aggregates can be prepared in water, which switch reversibly their “inside” to the “outside”, and vice versa.
In this thesis we use integral-field spectroscopy to detect and understand of Lyman α (Lyα) emission from high-redshift galaxies.
Intrinsically the Lyα emission at λ = 1216 Å is the strongest recombination line from galaxies. It arises from the 2p → 1s transition in hydrogen. In star-forming galaxies the line is powered by ionisation of the interstellar gas by hot O- and B- stars. Galaxies with star-formation rates of 1 - 10 Msol/year are expected to have Lyα luminosities of 42 dex - 43 dex (erg/s), corresponding to fluxes ~ -17 dex - -18 dex (erg/s/cm²) at redshifts z~3, where Lyα is easily accessible with ground-based telescopes. However, star-forming galaxies do not show these expected Lyα fluxes. Primarily this is a consequence of the high-absorption cross-section of neutral hydrogen for Lyα photons σ ~ -14 dex (cm²). Therefore, in typical interstellar environments Lyα photons have to undergo a complex radiative transfer. The exact conditions under which Lyα photons can escape a galaxy are poorly understood.
Here we present results from three observational projects. In Chapter 2, we show integral field spectroscopic observations of 14 nearby star-forming galaxies in Balmer α radiation (Hα, λ = 6562.8 Å). These observations were obtained with the Potsdam Multi Aperture Spectrophotometer at the Calar-Alto 3.5m Telescope}. Hα directly traces the intrinsic Lyα radiation field. We present Hα velocity fields and velocity dispersion maps spatially registered onto Hubble Space Telescope Lyα and Hα images. From our observations, we conjecture a causal connection between spatially resolved Hα kinematics and Lyα photometry for individual galaxies. Statistically, we find that dispersion-dominated galaxies are more likely to emit Lyα photons than galaxies where ordered gas-motions dominate. This result indicates that turbulence in actively star-forming systems favours an escape of Lyα radiation.
Not only massive stars can power Lyα radiation, but also non-thermal emission from an accreting super-massive black hole in the galaxy centre. If a galaxy harbours such an active galactic nucleus, the rate of hydrogen-ionising photons can be more than 1000 times higher than that of a typical star-forming galaxy. This radiation can potentially ionise large regions well outside the main stellar body of galaxies. Therefore, it is expected that the neutral hydrogen from these circum-galactic regions shines fluorescently in Lyα. Circum-galactic gas plays a crucial role in galaxy formation. It may act as a reservoir for fuelling star formation, and it is also subject to feedback processes that expel galactic material. If Lyα emission from this circum-galactic medium (CGM) was detected, these important processes could be studied in-situ around high-z galaxies. In Chapter 3, we show observations of five radio-quiet quasars with PMAS to search for possible extended CGM emission in the Lyα line. However, in four of the five objects, we find no significant traces of this emission. In the fifth object, there is evidence for a weak and spatially quite compact Lyα excess at several kpc outside the nucleus. The faintness of these structures is consistent with the idea that radio-quiet quasars typically reside in dark matter haloes of modest masses. While we were not able to detect Lyα CGM emission, our upper limits provide constraints for the new generation of IFS instruments at 8--10m class telescopes.
The Multi Unit Spectroscopic Explorer (MUSE) at ESOs Very Large Telescopeis such an unique instrument. One of the main motivating drivers in its construction was the use as a survey instrument for Lyα emitting galaxies at high-z. Currently, we are conducting such a survey that will cover a total area of ~100 square arcminutes with 1 hour exposures for each 1 square arcminute MUSE pointing. As a first result from this survey we present in Chapter 5 a catalogue of 831 emission-line selected galaxies from a 22.2 square arcminute region in the Chandra Deep Field South. In order to construct the catalogue, we developed and implemented a novel source detection algorithm -- LSDCat -- based on matched filtering for line emission in 3D spectroscopic datasets (Chapter 4). Our catalogue contains 237 Lyα emitting galaxies in the redshift range 3 ≲ z ≲ 6. Only four of those previously had spectroscopic redshifts in the literature. We conclude this thesis with an outlook on the construction of a Lyα luminosity function based on this unique sample (Chapter 6).
Meter and syntax have overlapping elements in music and speech domains, and individual differences have been documented in both meter perception and syntactic comprehension paradigms. Previous evidence insinuated but never fully explored the relationship that metrical structure has to syntactic comprehension, the comparability of these processes across music and language domains, and the respective role of individual differences. This dissertation aimed to investigate neurocognitive entrainment to meter in music and language, the impact that neurocognitive entrainment had on syntactic comprehension, and whether individual differences in musical expertise, temporal perception and working memory played a role during these processes.
A theoretical framework was developed, which linked neural entrainment, cognitive entrainment, and syntactic comprehension while detailing previously documented effects of individual differences on meter perception and syntactic comprehension. The framework was developed in both music and language domains and was tested using behavioral and EEG methods across three studies (seven experiments). In order to satisfy empirical evaluation of neurocognitive entrainment and syntactic aspects of the framework, original melodies and sentences were composed. Each item had four permutations: regular and irregular metricality, based on the hierarchical organization of strong and weak notes and syllables, and preferred and non-preferred syntax, based on structurally alternate endings. The framework predicted — for both music and language domains — greater neurocognitive entrainment in regular compared to irregular metricality conditions, and accordingly, better syntactic integration in regular compared to irregular metricality conditions. Individual differences among participants were expected for both entrainment and syntactic processes.
Altogether, the dissertation was able to support a holistic account of neurocognitive entrainment to musical meter and its subsequent influence on syntactic integration of melodies, with musician participants. The theoretical predictions were not upheld in the language domain with musician participants, but initial behavioral evidence in combination with previous EEG evidence suggest that perhaps non-musician language EEG data would support the framework’s predictions. Musicians’ deviation from hypothesized results in the language domain were suspected to reflect heightened perception of acoustic features stemming from musical training, which caused current ‘overly’ regular stimuli to distract the cognitive system. The individual-differences approach was vindicated by the surfacing of two factors scores, Verbal Working Memory and Time and Pitch Discrimination, which in turn correlated with multiple experimental data across the three studies.
Over the last decades, the world’s population has been growing at a faster rate, resulting in increased urbanisation, especially in developing countries. More than half of the global population currently lives in urbanised areas with an increasing tendency. The growth of cities results in a significant loss of vegetation cover, soil compaction and sealing of the soil surface which in turn results in high surface runoff during high-intensity storms and causes the problem of accelerated soil water erosion on streets and building grounds. Accelerated soil water erosion is a serious environmental problem in cities as it gives rise to the contamination of aquatic bodies, reduction of ground water recharge and increase in land degradation, and also results in damages to urban infrastructures, including drainage systems, houses and roads. Understanding the problem of water erosion in urban settings is essential for the sustainable planning and management of cities prone to water erosion. However, in spite of the vast existence of scientific literature on water erosion in rural regions, a concrete understanding of the underlying dynamics of urban erosion still remains inadequate for the urban dryland environments.
This study aimed at assessing water erosion and the associated socio-environmental determinants in a typical dryland urban area and used the city of Windhoek, Namibia, as a case study. The study used a multidisciplinary approach to assess the problem of water erosion. This included an in depth literature review on current research approaches and challenges of urban erosion, a field survey method for the quantification of the spatial extent of urban erosion in the dryland city of Windhoek, and face to face interviews by using semi-structured questionnaires to analyse the perceptions of stakeholders on urban erosion.
The review revealed that around 64% of the literatures reviewed were conducted in the developed world, and very few researches were carried out in regions with extreme climate, including dryland regions. Furthermore, the applied methods for erosion quantification and monitoring are not inclusive of urban typical features and they are not specific for urban areas. The reviewed literature also lacked aspects aimed at addressing the issues of climate change and policies regarding erosion in cities. In a field study, the spatial extent and severity of an urban dryland city, Windhoek, was quantified and the results show that nearly 56% of the city is affected by water erosion showing signs of accelerated erosion in the form of rills and gullies, which occurred mainly in the underdeveloped, informal and semi-formal areas of the city. Factors influencing the extent of erosion in Windhoek included vegetation cover and type, socio-urban factors and to a lesser extent slope estimates. A comparison of an interpolated field survey erosion map with a conventional erosion assessment tool (the Universal Soil Loss Equation) depicted a large deviation in spatial patterns, which underlines the inappropriateness of traditional non-urban erosion tools to urban settings and emphasises the need to develop new erosion assessment and management methods for urban environments. It was concluded that measures for controlling water erosion in the city need to be site-specific as the extent of erosion varied largely across the city.
The study also analysed the perceptions and understanding of stakeholders of urban water erosion in Windhoek, by interviewing 41 stakeholders using semi-structured questionnaires. The analysis addressed their understanding of water erosion dynamics, their perceptions with regards to the causes and the seriousness of erosion damages, and their attitudes towards the responsibilities for urban erosion. The results indicated that there is less awareness of the process as a phenomenon, instead there is more awareness of erosion damages and the factors contributing to the damages. About 69% of the stakeholders considered erosion damages to be ranging from moderate to very serious. However, there were notable disparities between the private householders and public authority groups. The study further found that the stakeholders have no clear understanding of their responsibilities towards the management of the control measures and payment for the damages. The private householders and local authority sectors pointed fingers at each other for the responsibilities for erosion damage payments and for putting up prevention measures. The reluctance to take responsibility could create a predicament for areas affected, specifically in the informal settlements where land management is not carried out by the local authority and land is not owned by the occupants.
The study concluded that in order to combat urban erosion, it is crucial to understand diverse dynamics aggravating the process of urbanisation from different scales. Accordingly, the study suggests that there is an urgent need for the development of urban-specific approaches that aim at: (a) incorporating the diverse socio-economic-environmental aspects influencing erosion, (b) scientifically improving natural cycles that influence water storages and nutrients for plants in urbanised dryland areas in order to increase the amount of vegetation cover, (c) making use of high resolution satellite images to improve the adopted methods for assessing urban erosion, (d) developing water erosion policies, and (e) continuously monitoring the impact of erosion and the influencing processes from local, national and international levels.
In the interest of producing functional catalysts from sustainable building-blocks, 1, 3-dicarboxylate imidazolium salts derived from amino acids were successfully modified to be suitable as N-Heterocyclic carbene (NHC) ligands within metal complexes. Complexes of Ag(I), Pd(II), and Ir(I) were successfully produced using known procedures using ligands derived from glycine, alanine, β-alanine and phenylalanine. The complexes were characterized in solid state using X-Ray crystallography, which allowed for the steric and electronic comparison of these ligands to well-known NHC ligands within analogous metal complexes.
The palladium complexes were tested as catalysts for aqueous-phase Suzuki-Miyaura cross-coupling. Water-solubility could be induced via ester hydrolysis of the N-bound groups in the presence of base. The mono-NHC–Pd complexes were seen to be highly active in the coupling of aryl bromides with phenylboronic acid; the active catalyst of which was determined to be mostly Pd(0) nanoparticles. Kinetic studies determined that reaction proceeds quickly in the coupling of bromoacetophenone, for both pre-hydrolyzed and in-situ hydrolysis catalyst dissolution. The catalyst could also be recycled for an extra run by simply re-using the aqueous layer.
The imidazolium salts were also used to produce organosilica hybrid materials. This was attempted via two methods: by post-grafting onto a commercial organosilica, and co-condensation of the corresponding organosilane. The co-condensation technique harbours potential for the production of solid-support catalysts.
Since 1998, elite athletes’ sport injuries have been monitored in single sport event, which leads to the development of first comprehensive injury surveillance system in multi-sport Olympic Games in 2008. However, injury and illness occurred in training phases have not been systematically studied due to its multi-facets, potentially interactive risk related factors. The present thesis aim to address issues of feasibility of establishing a validated measure for injury/illness, training environment and psychosocial risk factors by creating the evaluation tool namely risk of injury questionnaire (Risk-IQ) for elite athletes, which based on IOC consensus statement 2009 recommended content of preparticipation evaluation(PPE) and periodic health exam (PHE).
A total of 335 top level athletes and a total of 88 medical care providers from Germany and Taiwan participated in tow “cross-sectional plus longitudinal” Risk-IQ and MCPQ surveys respectively. Four categories of injury/illness related risk factors questions were asked in Risk-IQ for athletes while injury risk and psychological related questions were asked in MCPQ for MCP cohorts. Answers were quantified scales wise/subscales wise before analyzed with other factors/scales. In addition, adapted variables such as sport format were introduced for difference task of analysis.
Validated with 2-wyas translation and test-retest reliabilities, the Risk-IQ was proved to be in good standard which were further confirmed by analyzed results from official surveys in both Germany and Taiwan. The result of Risk-IQ revealed that elite athletes’ accumulated total injuries, in general, were multi-factor dependent; influencing factors including but not limited to background experiences, medical history, PHE and PPE medical resources as well as stress from life events. Injuries of different body parts were sport format and location specific. Additionally, medical support of PPE and PHE indicated significant difference between German and Taiwan.
The result of the present thesis confirmed that it is feasible to construct a comprehensive evalua-tion instrument for heterogeneous elite athletes cohorts’ risk factor analysis for injury/illness oc-curred during their non-competition periods. In average and with many moderators involved, Ger-man elite athletes have superior medical care support yet suffered more severe injuries than Tai-wanese counterparts. Opinions of injury related psychological issues reflected differently on vari-ous MCP groups irrespective of different nationalities. In general, influencing factors and interac-tions existed among relevant factors in both studies which implied further investigation with multiple regression analysis is needed for better understanding.
Ionothermal carbon materials
(2016)
Alternative concepts for energy storage and conversion have to be developed, optimized and employed to fulfill the dream of a fossil-independent energy economy. Porous carbon materials play a major role in many energy-related devices. Among different characteristics, distinct porosity features, e.g., specific surface area (SSA), total pore volume (TPV), and the pore size distribution (PSD), are important to maximize the performance in the final device. In order to approach the aim to synthesize carbon materials with tailor-made porosity in a sustainable fashion, the present thesis focused on biomass-derived precursors employing and developing the ionothermal carbonization.
During the ionothermal carbonization, a salt melt simultaneously serves as solvent and porogen. Typically, eutectic mixtures containing zinc chloride are employed as salt phase. The first topic of the present thesis addressed the possibility to precisely tailor the porosity of ionothermal carbon materials by an experimentally simple variation of the molar composition of the binary salt mixture. The developed pore tuning tool allowed the synthesis of glucose derived carbon materials with predictable SSAs in the range of ~ 900 to ~ 2100 m2 g-1. Moreover, the nucleobase adenine was employed as precursor introducing nitrogen functionalities in the final material. Thereby, the chemical properties of the carbon materials are varied leading to new application fields. Nitrogen doped carbons (NDCs) are able to catalyze the oxygen reduction reaction (ORR) which takes place on the cathodic site of a fuel cell. The herein developed porosity tailoring allowed the synthesis of adenine derived NDCs with outstanding SSAs of up to 2900 m2 g-1 and very large TPV of 5.19 cm3 g-1. Furthermore, the influence of the porosity on the ORR could be directly investigated enabling the precise optimization of the porosity characteristics of NDCs for this application. The second topic addressed the development of a new method to investigate the not-yet unraveled mechanism of the oxygen reduction reaction using a rotating disc electrode setup. The focus was put on noble-metal free catalysts. The results showed that the reaction pathway of the investigated catalysts is pH-dependent indicating different active species at different pH-values. The third topic addressed the expansion of the used salts for the ionothermal approach towards hydrated calcium and magnesium chloride. It was shown that hydrated salt phases allowed the introduction of a secondary templating effect which was connected to the coexistence of liquid and solid salt phases. The method enabled the synthesis of fibrous NDCs with SSAs of up to 2780 m2 g-1 and very large TPV of 3.86 cm3 g-1. Moreover, the concept of active site implementation by a facile low-temperature metalation employing the obtained NDCs as solid ligands could be shown for the first time in the context of ORR.
Overall, the thesis may pave the way towards highly porous carbon with tailor-made porosity materials prepared by an inexpensive and sustainable pathway, which can be applied in energy related field thereby supporting the needed expansion of the renewable energy sector.
Die Dissertation befasst sich mit der Organisation von humanitären Lufttransporten bei internationalen Katastrophen. Diese Flüge finden immer dann statt, wenn die eigene Hilfeleistungsfähigkeit der von Katastrophen betroffenen Regionen überfordert ist und Hilfe aus dem Ausland angefordert wird. Bei jedem der darauffolgenden Hilfseinsätze stehen Hilfsorganisationen und weitere mit der Katastrophenhilfe beteiligte Akteure erneut vor der Herausforderung, in kürzester Zeit eine logistische Kette aufzubauen, damit die Güter zum richtigen Zeitpunkt in der richtigen Menge am richtigen Ort eintreffen.
Humanitäre Lufttransporte werden in der Regel als Charterflüge organisiert und finden auf langen Strecken zu Zielen statt, die nicht selten abseits der hochfrequentierten Warenströme liegen. Am Markt ist das Angebot für derartige Transportdienstleistungen nicht gesichert verfügbar und unter Umständen müssen Hilfsorganisationen warten bis Kapazitäten mit geeigneten Flugzeugen zur Verfügung stehen. Auch qualitativ sind die Anforderungen von Hilfsorganisationen an die Hilfsgütertransporte höher als im regulären Linientransport.
Im Rahmen der Dissertation wird ein alternatives Organisationsmodell für die Beschaffung und den Betrieb sowie die Finanzierung von humanitären Lufttransporten aufgebaut. Dabei wird die gesicherte Verfügbarkeit von besonders flexibel einsetzbaren Flugzeugen in Betracht gezogen, mit deren Hilfe die Qualität und insbesondere die Planbarkeit der Hilfeleistung verbessert werden könnte.
Ein idealtypisches Modell wird hier durch die Kopplung der Kollektivgütertheorie, die der Finanzwissenschaft zuzuordnen ist, mit der Vertragstheorie als Bestandteil der Neuen Institutionenökonomik erarbeitet.
Empirische Beiträge zur Vertragstheorie bemängeln, dass es bei der Beschaffung von transaktionsspezifischen Investitionsgütern, wie etwa Flugzeugen mit besonderen Eigenschaften, aufgrund von Risiken und Umweltunsicherheiten zu ineffizienten Lösungen zwischen Vertragspartnern kommt. Die vorliegende Dissertation zeigt eine Möglichkeit auf, wie durch Aufbau einer gemeinsamen Informationsbasis ex-ante, also vor Vertragsschluss, Risiken und Umweltunsicherheiten reduziert werden können. Dies geschieht durch eine temporale Erweiterung eines empirischen Modells zur Bestimmung der Organisationsform bei transaktionsspezifischen Investitionsgütern aus der Regulierungsökonomik.
Die Arbeitet leistet darüber hinaus einen Beitrag zur Steigerung der Effizienz in der humanitären Logistik durch die fallspezifische Betrachtung von horizontalen Kooperationen und Professionalisierung der Hilfeleistung im Bereich der humanitären Luftfahrt.
It is commonly recognized that soil moisture exhibits spatial heterogeneities occurring in a wide range of scales. These heterogeneities are caused by different factors ranging from soil structure at the plot scale to land use at the landscape scale. There is an urgent need for effi-cient approaches to deal with soil moisture heterogeneity at large scales, where manage-ment decisions are usually made. The aim of this dissertation was to test innovative ap-proaches for making efficient use of standard soil hydrological data in order to assess seep-age rates and main controls on observed hydrological behavior, including the role of soil het-erogeneities.
As a first step, the applicability of a simplified Buckingham-Darcy method to estimate deep seepage fluxes from point information of soil moisture dynamics was assessed. This was done in a numerical experiment considering a broad range of soil textures and textural het-erogeneities. The method performed well for most soil texture classes. However, in pure sand where seepage fluxes were dominated by heterogeneous flow fields it turned out to be not applicable, because it simply neglects the effect of water flow heterogeneity. In this study a need for new efficient approaches to handle heterogeneities in one-dimensional water flux models was identified.
As a further step, an approach to turn the problem of soil moisture heterogeneity into a solu-tion was presented: Principal component analysis was applied to make use of the variability among soil moisture time series for analyzing apparently complex soil hydrological systems. It can be used for identifying the main controls on the hydrological behavior, quantifying their relevance, and describing their particular effects by functional averaged time series. The ap-proach was firstly tested with soil moisture time series simulated for different texture classes in homogeneous and heterogeneous model domains. Afterwards, it was applied to 57 mois-ture time series measured in a multifactorial long term field experiment in Northeast Germa-ny.
The dimensionality of both data sets was rather low, because more than 85 % of the total moisture variance could already be explained by the hydrological input signal and by signal transformation with soil depth. The perspective of signal transformation, i.e. analyzing how hydrological input signals (e.g., rainfall, snow melt) propagate through the vadose zone, turned out to be a valuable supplement to the common mass flux considerations. Neither different textures nor spatial heterogeneities affected the general kind of signal transfor-mation showing that complex spatial structures do not necessarily evoke a complex hydro-logical behavior. In case of the field measured data another 3.6% of the total variance was unambiguously explained by different cropping systems. Additionally, it was shown that dif-ferent soil tillage practices did not affect the soil moisture dynamics at all.
The presented approach does not require a priori assumptions about the nature of physical processes, and it is not restricted to specific scales. Thus, it opens various possibilities to in-corporate the key information from monitoring data sets into the modeling exercise and thereby reduce model uncertainties.
Widespread landscape changes are presently observed in the Arctic and are most likely to
accelerate in the future, in particular in permafrost regions which are sensitive to climate warming. To assess current and future developments, it is crucial to understand past
environmental dynamics in these landscapes. Causes and interactions of environmental variability can hardly be resolved by instrumental records covering modern time scales. However, long-term
environmental variability is recorded in paleoenvironmental archives. Lake sediments are important archives that allow reconstruction of local limnogeological processes as well as past environmental changes driven directly or indirectly by climate dynamics. This study aims at
reconstructing Late Quaternary permafrost and thermokarst dynamics in central-eastern Beringia,
the terrestrial land mass connecting Eurasia and North America during glacial sea-level low stands. In order to investigate development, processes and influence of thermokarst dynamics, several sediment cores from extant lakes and drained lake basins were analyzed to answer the
following research questions:
1. When did permafrost degradation and thermokarst lake development take place and what were enhancing and inhibiting environmental factors?
2. What are the dominant processes during thermokarst lake development and how are
they reflected in proxy records?
3. How did, and still do, thermokarst dynamics contribute to the inventory and properties of organic matter in sediments and the carbon cycle?
Methods applied in this study are based upon a multi-proxy approach combining
sedimentological, geochemical, geochronological, and micropaleontological analyses, as well as
analyses of stable isotopes and hydrochemistry of pore-water and ice. Modern field observations of water quality and basin morphometrics complete the environmental investigations.
The investigated sediment cores reveal permafrost degradation and thermokarst dynamics on different time scales. The analysis of a sediment core from GG basin on the northern Seward
Peninsula (Alaska) shows prevalent terrestrial accumulation of yedoma throughout the Early to
Mid Wisconsin with intermediate wet conditions at around 44.5 to 41.5 ka BP. This first wetland
development was terminated by the accumulation of a 1-meter-thick airfall tephra most likely originating from the South Killeak Maar eruption at 42 ka BP. A depositional hiatus between 22.5 and 0.23 ka BP may indicate thermokarst lake formation in the surrounding of the site which forms a yedoma upland till today. The thermokarst lake forming GG basin initiated 230 ± 30 cal a
BP and drained in Spring 2005 AD. Four years after drainage the lake talik was still unfrozen below 268 cm depth.
A permafrost core from Mama Rhonda basin on the northern Seward Peninsula preserved a
full lacustrine record including several lake phases. The first lake generation developed at 11.8 cal ka BP during the Lateglacial-Early Holocene transition; its old basin (Grandma Rhonda) is still partially preserved at the southern margin of the study basin. Around 9.0 cal ka BP a shallow and more dynamic thermokarst lake developed with actively eroding shorelines and potentially intermediate shallow water or wetland phases (Mama Rhonda). Mama Rhonda lake drainage at 1.1 cal ka BP was followed by gradual accumulation of terrestrial peat and top-down refreezing of the lake talik. A significant lower organic carbon content was measured in Grandma Rhonda deposits (mean TOC of 2.5 wt%) than in Mama Rhonda deposits (mean TOC of 7.9 wt%) highlighting the impact of thermokarst dynamics on biogeochemical cycling in different lake generations by thawing and mobilization of organic carbon into the lake system.
Proximal and distal sediment cores from Peatball Lake on the Arctic Coastal Plain of Alaska revealed young thermokarst dynamics since about 1,400 years along a depositional gradient based on reconstructions from shoreline expansion rates and absolute dating results. After its initiation as a remnant pond of a previous drained lake basin, a rapidly deepening lake with increasing oxygenation of the water column is evident from laminated sediments, and higher Fe/Ti and Fe/S ratios in the sediment. The sediment record archived characterizing shifts in depositional regimes and sediment sources from upland deposits and re-deposited sediments from drained thaw lake basins depending on the gradually changing shoreline configuration. These changes are evident from alternating organic inputs into the lake system which highlights the potential for thermokarst lakes to recycle old carbon from degrading permafrost deposits of its catchment.
The lake sediment record from Herschel Island in the Yukon (Canada) covers the full Holocene period. After its initiation as a thermokarst lake at 11.7 cal ka BP and intense thermokarst activity until 10.0 cal ka BP, the steady sedimentation was interrupted by a depositional hiatus at 1.6 cal ka BP which likely resulted from lake drainage or allochthonous slumping due to collapsing shore lines. The specific setting of the lake on a push moraine composed of marine deposits is reflected in the sedimentary record. Freshening of the maturing lake is indicated by decreasing electrical conductivity in pore-water. Alternation of marine to freshwater ostracods and foraminifera confirms decreasing salinity as well but also reflects episodical re-deposition of allochthonous marine sediments.
Based on permafrost and lacustrine sediment records, this thesis shows examples of the Late Quaternary evolution of typical Arctic permafrost landscapes in central-eastern Beringia and the complex interaction of local disturbance processes, regional environmental dynamics and global climate patterns. This study confirms that thermokarst lakes are important agents of organic matter recycling in complex and continuously changing landscapes.
The ever-increasing fat content in Western diet, combined with decreased levels of physical activity, greatly enhance the incidence of metabolic-related diseases. Cancer cachexia (CC) and Metabolic syndrome (MetS) are both multifactorial highly complex metabolism related syndromes, whose etiology is not fully understood, as the mechanisms underlying their development are not completely unveiled. Nevertheless, despite being considered “opposite sides”, MetS and CC share several common issues such as insulin resistance and low-grade inflammation. In these scenarios, tissue macrophages act as key players, due to their capacity to produce and release inflammatory mediators. One of the main features of MetS is hyperinsulinemia, which is generally associated with an attempt of the β-cell to compensate for diminished insulin sensitivity (insulin resistance). There is growing evidence that hyperinsulinemia per se may contribute to the development of insulin resistance, through the establishment of low grade inflammation in insulin responsive tissues, especially in the liver (as insulin is secreted by the pancreas into the portal circulation). The hypothesis of the present study was that insulin may itself provoke an inflammatory response culminating in diminished hepatic insulin sensitivity. To address this premise, firstly, human cell line U937 differentiated macrophages were exposed to insulin, LPS and PGE2. In these cells, insulin significantly augmented the gene expression of the pro-inflammatory mediators IL-1β, IL-8, CCL2, Oncostatin M (OSM) and microsomal prostaglandin E2 synthase (mPGES1), and of the anti-inflammatory mediator IL-10. Moreover, the synergism between insulin and LPS enhanced the induction provoked by LPS in IL-1β, IL-8, IL-6, CCL2 and TNF-α gene. When combined with PGE2, insulin enhanced the induction provoked by PGE2 in IL-1β, mPGES1 and COX2, and attenuated the inhibition induced by PGE2 in CCL2 and TNF-α gene expression contributing to an enhanced inflammatory response by both mechanisms. Supernatants of insulin-treated U937 macrophages reduced the insulin-dependent induction of glucokinase in hepatocytes by 50%. Cytokines contained in the supernatant of insulin-treated U937 macrophages also activated hepatocytes ERK1/2, resulting in inhibitory serine phosphorylation of the insulin receptor substrate. Additionally, the transcription factor STAT3 was activated by phosphorylation resulting in the induction of SOCS3, which is capable of interrupting the insulin receptor signal chain. MicroRNAs, non-coding RNAs linked to protein expression regulation, nowadays recognized as active players in the generation of several inflammatory disorders such as cancer and type II diabetes are also of interest. Considering that in cancer cachexia, patients are highly affected by insulin resistance and inflammation, control, non-cachectic and cachectic cancer patients were selected and the respective circulating levels of pro-inflammatory mediators and microRNA-21-5p, a posttranscriptional regulator of STAT3 expression, assessed and correlated. Cachectic patients circulating cytokines IL-6 and IL-8 levels were significantly higher than those of non-cachectic and controls, and the expression of microRNA-21-5p was significantly lower. Additionally, microRNA-21-5p reduced expression correlated negatively with IL-6 plasma levels. These results indicate that hyperinsulinemia per se might contribute to the low grade inflammation prevailing in MetS patients and thereby promote the development
of insulin resistance particularly in the liver. Diminished MicroRNA-21-5p expression may enhance inflammation and STAT3 expression in cachectic patients, contributing to the development of insulin resistance.
BACKGROUND: The etiology of low back pain (LBP), one of the most prevalent and costly diseases of our time, is accepted to be multi-causal, placing functional factors in the focus of research. Thereby, pain models suggest a centrally controlled strategy of trunk stiffening in LBP. However, supporting biomechanical evidence is mostly limited to static measurements during maximum voluntary contractions (MVC), probably influenced by psychological factors in LBP. Alternatively, repeated findings indicate that the neuromuscular efficiency (NME), characterized by the strength-to-activation relationship (SAR), of lower back muscles is impaired in LBP. Therefore, a dynamic SAR protocol, consisting of normalized trunk muscle activation recordings during submaximal loads (SMVC) seems to be relevant. This thesis aimed to investigate the influence of LBP on the NME and activation pattern of trunk muscles during dynamic trunk extensions.
METHODS: The SAR protocol consisted of an initial MVC reference trial (MVC1), followed by SMVCs at 20, 40, 60 and 80% of MVC1 load. An isokinetic trunk dynamometer (Con-Trex TP, ROM: 45° flexion to 10° extension, velocity: 45°/s) and a trunk surface EMG setup (myon, up to 12 leads) was used. Extension torque output [Nm] and muscular activation [V] were assessed in all trials. Finally, another MVC trial was performed (MVC2) for reliability analysis. For SAR evaluation the SMVC trial values were normalized [%MVC1] and compared inter- and intra-individually.
The methodical validity of the approach was tested in an isometric SAR single-case pilot study (S1a: N = 2, female LBP patient vs. healthy male). In addition, the validity of the MVC reference method was verified by comparing different contraction modes (S1b: N = 17, healthy individuals). Next, the isokinetic protocol was validated in terms of content for its applicability to display known physiological differences between sexes in a cross-sectional study (S2: each n = 25 healthy males/females). Finally, the influence of acute pain on NME was investigated longitudinally by comparing N = 8 acute LBP patients with the retest after remission of pain (S3). The SAR analysis focused on normalized agonistic extensor activation and abdominal and synergistic extensor co-activation (t-tests, ANOVA, α = .05) as well as on reliability of MVC1/2 outcomes.
RESULTS: During the methodological validation of the protocol (S1a), the isometric SAR was found to be descriptively different between individuals. Whereas torque output was highest during eccentric MVC, no relevant difference in peak EMG activation was found between contraction modes (S1b). The isokinetic SAR sex comparison (S2), though showing no significant overall effects, revealed higher normalized extensor activation at moderate submaximal loads in females (13 ± 4%), primarily caused by pronounced thoracic activation. Similarly, co-activation analysis resulted in significantly higher antagonistic activation at moderate loads compared to males (33 ± 9%). During intra-individual analysis of SAR in LBP patients (S3), a significant effect of pain status on the SAR has been identified, manifesting as increased normalized EMG activation of extensors during acute LBP (11 ± 8%) particularly at high load. Abdominal co-activation tended to be elevated (27 ± 11%) just as the thoracic extensor parts seemed to take over proportions of lumbar activation. All together, the M. erector spinae behaviour during the SAR protocol was rather linear with the tendency to rise exponentially during high loads. For the level of normalized EMG activation during SMVCs, a clear increasing trend from healthy males to females over to non-acute and acute LBP patients was discovered. This was associated by elevated antagonistic activation and a shift of synergistic towards lumbar extensor activation. The MVC data revealed overall good reliability, with clearly higher variability during acute LBP.
DISCUSSION: The present thesis demonstrates that the NME of lower back muscles is impaired in LBP patients, especially during an acute pain episode. A new dynamic protocol has been developed that makes it possible to display the underlying SAR using normalized trunk muscle EMG during submaximal isokinetic loads. The protocol shows promise as a biomechanical tool for diagnostic analysis of NME in LBP patients and monitoring of rehabilitation progress. Furthermore, reliability not of maximum strength but rather of peak EMG of MVC measurements seems to be decreased in LBP patients. Meanwhile, the findings of this thesis largely substantiate the assumptions made by the recently presented ‘motor adaptation to pain’ model, suggesting a pain-related intra- and intermuscular activation redistribution affecting movement and stiffness of the trunk. Further research is needed to distinguish the grade of NME impairment between LBP subgroups.
In experiments investigating sentence processing, eye movement measures such as fixation durations and regression proportions while reading are commonly used to draw conclusions about processing difficulties. However, these measures are the result of an interaction of multiple cognitive levels and processing strategies and thus are only indirect indicators of processing difficulty. In order to properly interpret an eye movement response, one has to understand the underlying principles of adaptive processing such as trade-off mechanisms between reading speed and depth of comprehension that interact with task demands and individual differences. Therefore, it is necessary to establish explicit models of the respective mechanisms as well as their causal relationship with observable behavior. There are models of lexical processing and eye movement control on the one side and models on sentence parsing and memory processes on the other. However, no model so far combines both sides with explicitly defined linking assumptions.
In this thesis, a model is developed that integrates oculomotor control with a parsing mechanism and a theory of cue-based memory retrieval. On the basis of previous empirical findings and independently motivated principles, adaptive, resource-preserving mechanisms of underspecification are proposed both on the level of memory access and on the level of syntactic parsing. The thesis first investigates the model of cue-based retrieval in sentence comprehension of Lewis & Vasishth (2005) with a comprehensive literature review and computational modeling of retrieval interference in dependency processing. The results reveal a great variability in the data that is not explained by the theory. Therefore, two principles, 'distractor prominence' and 'cue confusion', are proposed as an extension to the theory, thus providing a more adequate description of systematic variance in empirical results as a consequence of experimental design, linguistic environment, and individual differences. In the remainder of the thesis, four interfaces between parsing and eye movement control are defined: Time Out, Reanalysis, Underspecification, and Subvocalization. By comparing computationally derived predictions with experimental results from the literature, it is investigated to what extent these four interfaces constitute an appropriate elementary set of assumptions for explaining specific eye movement patterns during sentence processing. Through simulations, it is shown how this system of in itself simple assumptions results in predictions of complex, adaptive behavior.
In conclusion, it is argued that, on all levels, the sentence comprehension mechanism seeks a balance between necessary processing effort and reading speed on the basis of experience, task demands, and resource limitations. Theories of linguistic processing therefore need to be explicitly defined and implemented, in particular with respect to linking assumptions between observable behavior and underlying cognitive processes. The comprehensive model developed here integrates multiple levels of sentence processing that hitherto have only been studied in isolation. The model is made publicly available as an expandable framework for future studies of the interactions between parsing, memory access, and eye movement control.
Personal fabrication tools, such as 3D printers, are on the way of enabling a future in which non-technical users will be able to create custom objects. However, while the hardware is there, the current interaction model behind existing design tools is not suitable for non-technical users. Today, 3D printers are operated by fabricating the object in one go, which tends to take overnight due to the slow 3D printing technology. Consequently, the current interaction model requires users to think carefully before printing as every mistake may imply another overnight print. Planning every step ahead, however, is not feasible for non-technical users as they lack the experience to reason about the consequences of their design decisions.
In this dissertation, we propose changing the interaction model around personal fabrication tools to better serve this user group. We draw inspiration from personal computing and argue that the evolution of personal fabrication may resemble the evolution of personal computing: Computing started with machines that executed a program in one go before returning the result to the user. By decreasing the interaction unit to single requests, turn-taking systems such as the command line evolved, which provided users with feedback after every input. Finally, with the introduction of direct-manipulation interfaces, users continuously interacted with a program receiving feedback about every action in real-time. In this dissertation, we explore whether these interaction concepts can be applied to personal fabrication as well.
We start with fabricating an object in one go and investigate how to tighten the feedback-cycle on an object-level: We contribute a method called low-fidelity fabrication, which saves up to 90% fabrication time by creating objects as fast low-fidelity previews, which are sufficient to evaluate key design aspects. Depending on what is currently being tested, we propose different conversions that enable users to focus on different parts: faBrickator allows for a modular design in the early stages of prototyping; when users move on WirePrint allows quickly testing an object's shape, while Platener allows testing an object's technical function. We present an interactive editor for each technique and explain the underlying conversion algorithms.
By interacting on smaller units, such as a single element of an object, we explore what it means to transition from systems that fabricate objects in one go to turn-taking systems. We start with a 2D system called constructable: Users draw with a laser pointer onto the workpiece inside a laser cutter. The drawing is captured with an overhead camera. As soon as the the user finishes drawing an element, such as a line, the constructable system beautifies the path and cuts it--resulting in physical output after every editing step. We extend constructable towards 3D editing by developing a novel laser-cutting technique for 3D objects called LaserOrigami that works by heating up the workpiece with the defocused laser until the material becomes compliant and bends down under gravity. While constructable and LaserOrigami allow for fast physical feedback, the interaction is still best described as turn-taking since it consists of two discrete steps: users first create an input and afterwards the system provides physical output.
By decreasing the interaction unit even further to a single feature, we can achieve real-time physical feedback: Input by the user and output by the fabrication device are so tightly coupled that no visible lag exists. This allows us to explore what it means to transition from turn-taking interfaces, which only allow exploring one option at a time, to direct manipulation interfaces with real-time physical feedback, which allow users to explore the entire space of options continuously with a single interaction. We present a system called FormFab, which allows for such direct control. FormFab is based on the same principle as LaserOrigami: It uses a workpiece that when warmed up becomes compliant and can be reshaped. However, FormFab achieves the reshaping not based on gravity, but through a pneumatic system that users can control interactively. As users interact, they see the shape change in real-time.
We conclude this dissertation by extrapolating the current evolution into a future in which large numbers of people use the new technology to create objects. We see two additional challenges on the horizon: sustainability and intellectual property. We investigate sustainability by demonstrating how to print less and instead patch physical objects. We explore questions around intellectual property with a system called Scotty that transfers objects without creating duplicates, thereby preserving the designer's copyright.
The human immunodeficiency virus (HIV) has resisted nearly three decades of efforts targeting a cure. Sustained suppression of the virus has remained a challenge, mainly due
to the remarkable evolutionary adaptation that the virus exhibits by the accumulation of drug-resistant mutations in its genome. Current therapeutic strategies aim at achieving and maintaining a low viral burden and typically involve multiple drugs. The choice of optimal combinations of these drugs is crucial, particularly in the background of treatment failure having occurred previously with certain other drugs. An understanding of the dynamics of viral mutant genotypes aids in the assessment of treatment failure with a certain drug
combination, and exploring potential salvage treatment regimens.
Mathematical models of viral dynamics have proved invaluable in understanding the viral life cycle and the impact of antiretroviral drugs. However, such models typically use simplified and coarse-grained mutation schemes, that curbs the extent of their application to drug-specific clinical mutation data, in order to assess potential next-line therapies. Statistical
models of mutation accumulation have served well in dissecting mechanisms of resistance evolution by reconstructing mutation pathways under different drug-environments. While these models perform well in predicting treatment outcomes by statistical learning, they do not incorporate drug effect mechanistically. Additionally, due to an inherent lack of
temporal features in such models, they are less informative on aspects such as predicting mutational abundance at treatment failure. This limits their application in analyzing the
pharmacology of antiretroviral drugs, in particular, time-dependent characteristics of HIV therapy such as pharmacokinetics and pharmacodynamics, and also in understanding the impact of drug efficacy on mutation dynamics.
In this thesis, we develop an integrated model of in vivo viral dynamics incorporating drug-specific mutation schemes learned from clinical data. Our combined modelling
approach enables us to study the dynamics of different mutant genotypes and assess mutational abundance at virological failure. As an application of our model, we estimate in vivo
fitness characteristics of viral mutants under different drug environments. Our approach also extends naturally to multiple-drug therapies. Further, we demonstrate the versatility of our model by showing how it can be modified to incorporate recently elucidated mechanisms of drug action including molecules that target host factors.
Additionally, we address another important aspect in the clinical management of HIV disease, namely drug pharmacokinetics. It is clear that time-dependent changes in in vivo
drug concentration could have an impact on the antiviral effect, and also influence decisions on dosing intervals. We present a framework that provides an integrated understanding
of key characteristics of multiple-dosing regimens including drug accumulation ratios and half-lifes, and then explore the impact of drug pharmacokinetics on viral suppression.
Finally, parameter identifiability in such nonlinear models of viral dynamics is always a concern, and we investigate techniques that alleviate this issue in our setting.
Effektivität frühzeitiger Interventionen zur Prävention von Lese- und Rechtschreibschwierigkeiten
(2016)
Die vorliegende Studie beschäftigt sich mit der Förderung der Lese- und Schreibkompetenz in der Anfangsphase des Schriftspracherwerbs. Ziel der Untersuchung ist die Erprobung und Evaluierung frühzeitiger, diagnosegeleiteter Interventionen zur Prävention von Lese- und Rechtschreibschwierigkeiten. Im Unterschied zu vielen Studien in diesem Bereich werden alle Maßnahmen unter realen schulischen Bedingungen im Rahmen des Erstlese- und Schreibunterrichts durch die Klassenlehrer selbst durchgeführt, wobei diese von der Autorin unterstützt und begleitet werden. Förder- und Prozessdiagnose sowie Elemente diagnosegeleiteter Förderung werden aus Theorien und Forschungslage abgeleitet und zu einem Interventionsset verbunden. Die Effektivität der evidenzbasierten Maßnahmen wird durch Parallelgruppenvergleiche überprüft.
An der empirischen Untersuchung nahmen insgesamt 25 Schulklassen mit 560 Erstklässlern teil, geteilt in Versuchs- und Kontrollgruppe. Mit der Eingangsdiagnose am Schulbeginn wurden Voraussetzungen für den Schriftspracherwerb erhoben und mit der Evaluierungsdiagnose am Ende der ersten Schulstufe entwicklungsadäquate schriftsprachliche Kompetenzen auf der Wortebene überprüft. Zusätzlich erfasst wurden internale und externale Einflussfaktoren, deren Wirkung in der statistischen Auswertung berücksichtigt wurde. Alle Datenerhebungen wurden in Versuchs- und Kontrollgruppe durchgeführt, während die evidenzbasierten Treatments nur in der Versuchsgruppe stattfanden.
Die Auswertung bestätigt mit signifikanten Ergebnissen den engen Zusammenhang zwischen der Phonologischen Bewusstheit zu Beginn des Schriftspracherwerbs und der Lese- und Rechtschreibfähigkeit am Ende der ersten Schulstufe sowie zwischen Familiärer Literalität und Lesefertigkeit. Schriftsprachliche Vorkenntnisse weisen eine Tendenz zur Signifikanz hinsichtlich ihrer positiven Wirkung auf die Basale Lesefertigkeit auf. Eine höchst signifikante positive Wirkung auf die Basale Lesefertigkeit zeigt die Druckschrift als Ausgangsschrift.
Die Ergebnisse deuten auf eine Überlegenheit vorschulischer präliteraler Fertigkeiten hinsichtlich ihrer Wirkung auf die Lese- und Rechtschreibfertigkeit am Ende der ersten Schulstufe gegenüber Fördermaßnahmen unter realen schulischen Bedingungen hin. Die positive Wirkung einer unverbundenen Ausgangsschrift auf den Leseerwerb betont die Wichtigkeit der Wahl der Ausgangsschrift. Im frühen Schriftspracherwerb sollte die Druckschrift für das Lesen und Schreiben verwendet werden.
Computer Security deals with the detection and mitigation of threats to computer networks, data, and computing hardware. This
thesis addresses the following two computer security problems: email spam campaign and malware detection.
Email spam campaigns can easily be generated using popular dissemination tools by specifying simple grammars that serve as message templates. A grammar is disseminated to nodes of a bot net, the nodes create messages by instantiating the grammar at random. Email spam campaigns can encompass huge data volumes and therefore pose a threat to the stability of the infrastructure of email service providers that have to store them. Malware -software that serves a malicious purpose- is affecting web servers, client computers via active content, and client computers through executable files. Without the help of malware detection systems it would be easy for malware creators to collect sensitive information or to infiltrate computers.
The detection of threats -such as email-spam messages, phishing messages, or malware- is an adversarial and therefore intrinsically
difficult problem. Threats vary greatly and evolve over time. The detection of threats based on manually-designed rules is therefore
difficult and requires a constant engineering effort. Machine-learning is a research area that revolves around the analysis of data and the discovery of patterns that describe aspects of the data. Discriminative learning methods extract prediction models from data that are optimized to predict a target attribute as accurately as possible. Machine-learning methods hold the promise of automatically identifying patterns that robustly and accurately detect threats. This thesis focuses on the design and analysis of discriminative learning methods for the two computer-security problems under investigation: email-campaign and malware detection.
The first part of this thesis addresses email-campaign detection. We focus on regular expressions as a syntactic framework, because regular expressions are intuitively comprehensible by security engineers and administrators, and they can be applied as a detection mechanism in an extremely efficient manner. In this setting, a prediction model is provided with exemplary messages from an email-spam campaign. The prediction model has to generate a regular expression that reveals the syntactic pattern that underlies the entire campaign, and that a security engineers finds comprehensible and feels confident enough to use the expression to blacklist further messages at the email server. We model this problem as two-stage learning problem with structured input and output spaces which can be solved using standard cutting plane methods. Therefore we develop an appropriate loss function, and derive a decoder for the resulting optimization problem.
The second part of this thesis deals with the problem of predicting whether a given JavaScript or PHP file is malicious or benign. Recent malware analysis techniques use static or dynamic features, or both. In fully dynamic analysis, the software or script is executed and observed for malicious behavior in a sandbox environment. By contrast, static analysis is based on features that can be extracted directly from the program file. In order to bypass static detection mechanisms, code obfuscation techniques are used to spread a malicious program file in many different syntactic variants. Deobfuscating the code before applying a static classifier can be subjected to mostly static code analysis and can overcome the problem of obfuscated malicious code, but on the other hand increases the computational costs of malware detection by an order of magnitude. In this thesis we present a cascaded architecture in which a classifier first performs a static analysis of the original code and -based on the outcome of this first classification step- the code may be deobfuscated and classified again. We explore several types of features including token $n$-grams, orthogonal sparse bigrams, subroutine-hashings, and syntax-tree features and study the robustness of detection methods and feature types against the evolution of malware over time. The developed tool scans very large file collections quickly and accurately.
Each model is evaluated on real-world data and compared to reference methods. Our approach of inferring regular expressions to filter emails belonging to an email spam campaigns leads to models with a high true-positive rate at a very low false-positive rate that is an order of magnitude lower than that of a commercial content-based filter. Our presented system -REx-SVMshort- is being used by a commercial email service provider and complements content-based and IP-address based filtering.
Our cascaded malware detection system is evaluated on a high-quality data set of almost 400,000 conspicuous PHP files and a collection of more than 1,00,000 JavaScript files. From our case study we can conclude that our system can quickly and accurately process large data collections at a low false-positive rate.
The goal of the presented work is to explore the interaction between gold nanorods (GNRs) and hyper-sound waves. For the generation of the hyper-sound I have used Azobenzene-containing polymer transducers. Multilayer polymer structures with well-defined thicknesses and smooth interfaces were built via layer-by-layer deposition. Anionic polyelectrolytes with Azobenzene side groups (PAzo) were alternated with cationic polymer PAH, for the creation of transducer films. PSS/PAH multilayer were built for spacer layers, which do not absorb in the visible light range. The properties of the PAzo/PAH film as a transducer are carefully characterized by static and transient optical spectroscopy. The optical and mechanical properties of the transducer are studied on the picosecond time scale. In particular the relative change of the refractive index of the photo-excited and expanded PAH/PAzo is Δn/n = - 2.6*10‐4. Calibration of the generated strain is performed by ultrafast X-ray diffraction calibrated the strain in a Mica substrate, into which the hyper-sound is transduced. By simulating the X-ray data with a linear-chain-model the strain in the transducer under the excitation is derived to be Δd/d ~ 5*10‐4.
Additional to the investigation of the properties of the transducer itself, I have performed a series of experiments to study the penetration of the generated strain into various adjacent materials. By depositing the PAzo/PAH film onto a PAH/PSS structure with gold nanorods incorporated in it, I have shown that nanoscale impurities can be detected via the scattering of hyper-sound.
Prior to the investigation of complex structures containing GNRs and the transducer, I have performed several sets of experiments on GNRs deposited on a small buffer of PSS/PAH. The static and transient response of GNRs is investigated for different fluence of the pump beam and for different dielectric environments (GNRs covered by PSS/PAH).
A systematic analysis of sample architectures is performed in order to construct a sample with the desired effect of GNRs responding to the hyper-sound strain wave. The observed shift of a feature related to the longitudinal plasmon resonance in the transient reflection spectra is interpreted as the event of GNRs sensing the strain wave. We argue that the shift of the longitudinal plasmon resonance is caused by the viscoelastic deformation of the polymer around the nanoparticle. The deformation is induced by the out of plane difference in strain in the area directly under a particle and next to it. Simulations based on the linear chain model support this assumption. Experimentally this assumption is proven by investigating the same structure, with GNRs embedded in a PSS/PAH polymer layer.
The response of GNRs to the hyper-sound wave is also observed for the sample structure with GNRs embedded in PAzo/PAH films. The response of GNRs in this case is explained to be driven by the change of the refractive index of PAzo during the strain propagation.
The global carbon cycle is closely linked to Earth’s climate. In the context of continuously unchecked anthropogenic CO₂ emissions, the importance of natural CO₂ bond and carbon storage is increasing. An important biogenic mechanism of natural atmospheric CO₂ drawdown is the photosynthetic carbon fixation in plants and the subsequent longterm deposition of plant detritus in sediments.
The main objective of this thesis is to identify factors that control mobilization and transport of plant organic matter (pOM) through rivers towards sedimentation basins. I investigated this aspect in the eastern Nepalese Arun Valley. The trans-Himalayan Arun River is characterized by a strong elevation gradient (205 − 8848 m asl) that is accompanied by strong changes in ecology and climate ranging from wet tropical conditions in the Himalayan forelad to high alpine tundra on the Tibetan Plateau. Therefore, the Arun is an excellent natural laboratory, allowing the investigation of the effect of vegetation cover, climate, and topography on plant organic matter mobilization and export in tributaries along the gradient.
Based on hydrogen isotope measurements of plant waxes sampled along the Arun River and its tributaries, I first developed a model that allows for an indirect quantification of pOM contributed to the mainsetm by the Arun’s tributaries. In order to determine the role of climatic and topographic parameters of sampled tributary catchments, I looked for significant statistical relations between the amount of tributary pOM export and tributary characteristics (e.g. catchment size, plant cover, annual precipitation or runoff, topographic measures). On one hand, I demonstrated that pOMsourced from the Arun is not uniformly derived from its entire catchment area. On the other, I showed that dense vegetation is a necessary, but not sufficient, criterion for high tributary pOM export. Instead, I identified erosion and rainfall and runoff as key factors controlling pOM sourcing in the Arun Valley. This finding is supported by terrestrial cosmogenic nuclide concentrations measured on river sands along the Arun and its tributaries in order to quantify catchment wide denudation rates. Highest denudation rates corresponded well with maximum pOM mobilization and export also suggesting the link between erosion and pOM sourcing.
The second part of this thesis focusses on the applicability of stable isotope records such as plant wax n-alkanes in sediment archives as qualitative and quantitative proxy for the variability of past Indian Summer Monsoon (ISM) strength. First, I determined how ISM strength affects the hydrogen and oxygen stable isotopic composition (reported as δD and δ18O values vs. Vienna Standard Mean Ocean Water) of precipitation in the Arun Valley and if this amount effect (Dansgaard, 1964) is strong enough to be recorded in potential paleo-ISM isotope proxies. Second, I investigated if potential isotope records across the Arun catchment reflect ISM strength dependent precipitation δD values only, or if the ISM isotope signal is superimposed by winter precipitation or glacial melt. Furthermore, I tested if δD values of plant waxes in fluvial deposits reflect δD values of environmental waters in the respective catchments.
I showed that surface water δD values in the Arun Valley and precipitation δD from south of the Himalaya both changed similarly during two consecutive years (2011 & 2012) with distinct ISM rainfall amounts (~20% less in 2012). In order to evaluate the effect of other water sources (Winter-Westerly precipitation, glacial melt) and evapotranspiration in the Arun Valley, I analysed satellite remote sensing data of rainfall distribution (TRMM 3B42V7), snow cover (MODIS MOD10C1), glacial coverage (GLIMSdatabase, Global Land Ice Measurements from Space), and evapotranspiration (MODIS MOD16A2). In addition to the predominant ISM in the entire catchment I found through stable isotope analysis of surface waters indications for a considerable amount of glacial melt derived from high altitude tributaries and the Tibetan Plateau. Remotely sensed snow cover data revealed that the upper portion of the Arun also receives considerable winter precipitation, but the effect of snow melt on the Arun Valley hydrology could not be evaluated as it takes place in early summer, several months prior to our sampling campaigns. However, I infer that plant wax records and other potential stable isotope proxy archives below the snowline are well-suited for qualitative, and potentially quantitative, reconstructions of past changes of ISM strength.
Geospatial data has become a natural part of a growing number of information systems and services in the economy, society, and people's personal lives. In particular, virtual 3D city and landscape models constitute valuable information sources within a wide variety of applications such as urban planning, navigation, tourist information, and disaster management. Today, these models are often visualized in detail to provide realistic imagery. However, a photorealistic rendering does not automatically lead to high image quality, with respect to an effective information transfer, which requires important or prioritized information to be interactively highlighted in a context-dependent manner.
Approaches in non-photorealistic renderings particularly consider a user's task and camera perspective when attempting optimal expression, recognition, and communication of important or prioritized information. However, the design and implementation of non-photorealistic rendering techniques for 3D geospatial data pose a number of challenges, especially when inherently complex geometry, appearance, and thematic data must be processed interactively. Hence, a promising technical foundation is established by the programmable and parallel computing architecture of graphics processing units.
This thesis proposes non-photorealistic rendering techniques that enable both the computation and selection of the abstraction level of 3D geospatial model contents according to user interaction and dynamically changing thematic information. To achieve this goal, the techniques integrate with hardware-accelerated rendering pipelines using shader technologies of graphics processing units for real-time image synthesis. The techniques employ principles of artistic rendering, cartographic generalization, and 3D semiotics—unlike photorealistic rendering—to synthesize illustrative renditions of geospatial feature type entities such as water surfaces, buildings, and infrastructure networks. In addition, this thesis contributes a generic system that enables to integrate different graphic styles—photorealistic and non-photorealistic—and provide their seamless transition according to user tasks, camera view, and image resolution.
Evaluations of the proposed techniques have demonstrated their significance to the field of geospatial information visualization including topics such as spatial perception, cognition, and mapping. In addition, the applications in illustrative and focus+context visualization have reflected their potential impact on optimizing the information transfer regarding factors such as cognitive load, integration of non-realistic information, visualization of uncertainty, and visualization on small displays.
Change points in time series are perceived as heterogeneities in the statistical or dynamical characteristics of the observations. Unraveling such transitions yields essential information for the understanding of the observed system’s intrinsic evolution and potential external influences. A precise detection of multiple changes is therefore of great importance for various research disciplines, such as environmental sciences, bioinformatics and economics. The primary purpose of the detection approach introduced in this thesis is the investigation of transitions underlying direct or indirect climate observations. In order to develop a diagnostic approach capable to capture such a variety of natural processes, the generic statistical features in terms of central tendency and dispersion are employed in the light of Bayesian inversion. In contrast to established Bayesian approaches to multiple changes, the generic approach proposed in this thesis is not formulated in the framework of specialized partition models of high dimensionality requiring prior specification, but as a robust kernel-based approach of low dimensionality employing least informative prior distributions.
First of all, a local Bayesian inversion approach is developed to robustly infer on the location and the generic patterns of a single transition. The analysis of synthetic time series comprising changes of different observational evidence, data loss and outliers validates the performance, consistency and sensitivity of the inference algorithm. To systematically investigate time series for multiple changes, the Bayesian inversion is extended to a kernel-based inference approach. By introducing basic kernel measures, the weighted kernel inference results are composed into a proxy probability to a posterior distribution of multiple transitions. The detection approach is applied to environmental time series from the Nile river in Aswan and the weather station Tuscaloosa, Alabama comprising documented changes. The method’s performance confirms the approach as a powerful diagnostic tool to decipher multiple changes underlying direct climate observations.
Finally, the kernel-based Bayesian inference approach is used to investigate a set of complex terrigenous dust records interpreted as climate indicators of the African region of the Plio-Pleistocene period. A detailed inference unravels multiple transitions underlying the indirect climate observations, that are interpreted as conjoint changes. The identified conjoint changes coincide with established global climate events. In particular, the two-step transition associated to the establishment of the modern Walker-Circulation contributes to the current discussion about the influence of paleoclimate changes on the environmental conditions in tropical and subtropical Africa at around two million years ago.
Die Elbe und ihr Einzugsgebiet sind vom Klimawandel betroffen. Um die Wirkkette von projizierten Klimaveränderungen auf den Wasserhaushalt und die daraus resultierenden Nährstoffeinträge und -frachten für große Einzugsgebiete wie das der Elbe zu analysieren, können integrierte Umweltmodellsysteme eingesetzt werden. Fallstudien, die mit diesen Modellsystemen ad hoc durchgeführt werden, repräsentieren den Istzustand von Modellentwicklungen und -unsicherheiten und sind damit statisch.
Diese Arbeit beschreibt den Einstieg in die Dynamisierung von Klimafolgenanalysen im Elbegebiet. Dies umfasst zum einen eine Plausibilitätsprüfung von Auswirkungsrechnungen, die mit Szenarien des statistischen Szenariengenerators STARS durchgeführt wurden, durch den Vergleich mit den Auswirkungen neuerer Klimaszenarien aus dem ISI-MIP Projekt, die dem letzten Stand der Klimamodellierung entsprechen. Hierfür wird ein integriertes Modellsystem mit "eingefrorenem Entwicklungsstand" verwendet. Die Klimawirkungsmodelle bleiben dabei unverändert. Zum anderen wird ein Bestandteil des integrierten Modellsystems – das ökohydrologische Modell SWIM – zu einer "live"-Version weiterentwickelt. Diese wird durch punktuelle Testung an langjährigen Versuchsreihen eines Lysimeterstandorts sowie an aktuellen Abflussreihen validiert und verbessert.
Folgende Forschungsfragen werden bearbeitet: (i) Welche Effekte haben unterschiedliche Klimaszenarien auf den Wasserhaushalt im Elbegebiet und ist eine Neubewertung der Auswirkung des Klimawandels auf den Wasserhaushalt notwendig?, (ii) Was sind die Auswirkungen des Klimawandels auf die Nährstoffeinträge und -frachten im Elbegebiet sowie die Wirksamkeit von Maßnahmen zur Reduktion der Nährstoffeinträge?, (iii) Ist unter der Nutzung (selbst einer sehr geringen Anzahl) verfügbarer tagesaktueller Witterungsdaten in einem stark heterogenen Einzugsgebiet eine valide Ansprache der aktuellen ökohydrologischen Situation des Elbeeinzugsgebiets möglich?
Die aktuellen Szenarien bestätigen die Richtung, jedoch nicht das Ausmaß der Klimafolgen: Die Rückgänge des mittleren jährlichen Gesamtabflusses und der monatlichen Abflüsse an den Pegeln bis Mitte des Jahrhunderts betragen für das STARS-Szenario ca. 30 %. Die Rückgänge bei den auf dem ISI-MIP-Szenario basierenden Modellstudien liegen hingegen nur bei ca. 10 %. Hauptursachen für diese Divergenz sind die Unterschiede in den Niederschlagsprojektionen sowie die Unterschiede in der jahreszeitlichen Verteilung der Erwärmung. Im STARS-Szenario gehen methodisch bedingt die Niederschläge zurück und der Winter erwärmt sich stärker als der Sommer. In dem ISI-MIP-Szenario bleiben die Niederschläge nahezu stabil und die Erwärmung im Sommer und Winter unterscheidet sich nur geringfügig.
Generell nehmen die Nährstoffeinträge und -frachten mit den Abflüssen in beiden Szenarien unterproportional ab, wobei die Frachten jeweils stärker als die Einträge zurückgehen. Die konkreten Effekte der Abflussänderungen sind gering und liegen im einstelligen Prozentbereich. Gleiches gilt für die Unterschiede zwischen den Szenarien. Der Effekt von zwei ausgewählten Maßnahmen zur Reduktion der Nährstoffeinträge und -frachten unterscheidet sich bei verschiedenen Abflussverhältnissen, repräsentiert durch unterschiedliche Klimaszenarien in unterschiedlich feuchter Ausprägung, ebenfalls nur geringfügig.
Die Beantwortung der ersten beiden Forschungsfragen zeigt, dass die Aktualisierung von Klimaszenarien in einem ansonsten "eingefrorenen" Verbund von ökohydrologischen Daten und Modellen eine wichtige Prüfoption für die Plausibilisierung von Klimafolgenanalysen darstellt. Sie bildet die methodische Grundlage für die Schlussfolgerung, dass bei der Wassermenge eine Neubewertung der Klimafolgen notwendig ist, während dies bei den Nährstoffeinträgen und -frachten nicht der Fall ist.
Die zur Beantwortung der dritten Forschungsfrage mit SWIM-live durchgeführten Validierungsstudien ergeben Diskrepanzen am Lysimeterstandort und bei den Abflüssen aus den Teilgebieten Saale und Spree. Sie lassen sich zum Teil mit der notwendigen Interpolationsweite der Witterungsdaten und dem Einfluss von Wasserbewirtschaftungsmaßnahmen erklären. Insgesamt zeigen die Validierungsergebnisse, dass schon die Pilotversion von SWIM-live für eine ökohydrologische Ansprache des Gebietswasserhaushaltes im Elbeeinzugsgebiet genutzt werden kann. SWIM-live ermöglicht eine unmittelbare Betrachtung und Beurteilung simulierter Daten. Dadurch werden Unsicherheiten bei der Modellierung direkt offengelegt und können infolge dessen reduziert werden. Zum einen führte die Verdichtung der meteorologischen Eingangsdaten durch die Verwendung von nun ca. 700 anstatt 19 Klima- bzw. Niederschlagstationen zu einer Verbesserung der Ergebnisse. Zum anderen wurde SWIM-live beispielhaft für einen Zyklus aus punktueller Modellverbesserung und flächiger Überprüfung der Simulationsergebnisse genutzt.
Die einzelnen Teilarbeiten tragen jeweils zur Dynamisierung von Klimafolgenanalysen im Elbegebiet bei. Der Anlass hierfür war durch die fehlerhaften methodischen Grundlagen von STARS gegeben. Die Sinnfälligkeit der Dynamisierung ist jedoch nicht an diesen konkreten Anlass gebunden, sondern beruht auf der grundlegenden Einsicht, dass Ad-hoc-Szenarienanalysen immer auch pragmatische Vereinfachungen zugrunde liegen, die fortlaufend überprüft werden müssen.
Für alle Organismen ist die Aufrechterhaltung ihres energetischen Gleichgewichts unter fluktuierenden Umweltbedingungen lebensnotwendig. In Eukaryoten steuern evolutionär konservierte Proteinkinasen, die in Pflanzen als SNF1-RELATED PROTEIN KINASE1 (SnRK1) bezeichnet werden, die Adaption an Stresssignale aus der Umwelt und an die Limitierung von Nährstoffen und zellulärer Energie. Die Aktivierung von SnRK1 bedingt eine umfangreiche transkriptionelle Umprogrammierung, die allgemein zu einer Repression energiekonsumierender Prozesse wie beispielsweise Zellteilung und Proteinbiosynthese und zu einer Induktion energieerzeugender, katabolischer Stoffwechselwege führt. Wie unterschiedliche Signale zu einer generellen sowie teilweise gewebe- und stressspezifischen SnRK1-vermittelten Antwort führen ist bisher noch nicht ausreichend geklärt, auch weil bislang nur wenige Komponenten der SnRK1-Signaltransduktion identifiziert wurden. In dieser Arbeit konnte ein Protein-Protein-Interaktionsnetzwerk um die SnRK1αUntereinheiten aus Arabidopsis AKIN10/AKIN11 etabliert werden. Dadurch wurden zunächst Mitglieder der pflanzenspezifischen DUF581-Proteinfamilie als Interaktionspartner der SnRK1α-Untereinheiten identifiziert. Diese Proteine sind über ihre konservierte DUF581Domäne, in der ein Zinkfinger-Motiv lokalisiert ist, fähig mit AKIN10/AKIN11 zu interagieren. In planta Ko-Expressionsanalysen zeigten, dass die DUF581-Proteine eine Verschiebung der nucleo-cytoplasmatischen Lokalisierung von AKIN10 hin zu einer nahezu ausschließlichen zellkernspezifischen Lokalisierung begünstigen sowie die Ko-Lokalisierung von AKIN10 und DUF581-Proteinen im Nucleus. In Bimolekularen Fluoreszenzkomplementations-Analysen konnte die zellkernspezifische Interaktion von DUF581-Proteinen mit SnRK1α-Untereinheiten in planta bestätigt werden. Außerhalb der DUF581-Domäne weisen die Proteine einander keine große Sequenzähnlichkeit auf. Aufgrund ihrer Fähigkeit mit SnRK1 zu interagieren, dem Fehlen von SnRK1Phosphorylierungsmotiven sowie ihrer untereinander sehr variabler gewebs-, entwicklungs- und stimulusspezifischer Expression wurde für DUF581-Proteine eine Funktion als Adaptoren postuliert, die unter bestimmten physiologischen Bedingungen spezifische Substratproteine in den SnRK1-Komplex rekrutieren. Auf diese Weise könnten DUF581Proteine die Interaktion von SnRK1 mit deren Zielproteinen modifizieren und eine Feinjustierung der SnRK1-Signalweiterleitung ermöglichen. Durch weiterführende Interaktionsstudien konnten DUF581-interagierende Proteine darunter Transkriptionsfaktoren, Proteinkinasen sowie regulatorische Proteine gefunden werden, die teilweise ebenfalls Wechselwirkungen mit SnRK1α-Untereinheiten aufzeigten. Im Rahmen dieser Arbeit wurde eines dieser Proteine für das eine Beteiligung an der SnRK1Signalweiterleitung als Transkriptionsregulator vermutet wurde näher charakterisiert. STKR1 (STOREKEEPER RELATED 1), ein spezifischer Interaktionspartner von DUF581-18, gehört zu einer pflanzenspezifischen Leucin-Zipper-Transkriptionsfaktorfamilie und interagiert in Hefe sowie in planta mit SnRK1. Die zellkernspezifische Interaktion von STKR1 und AKIN10 in Pflanzen unterstützt die Vermutung der kooperativen Regulation von Zielgenen. Weiterhin stabilisierte die Anwesenheit von AKIN10 die Proteingehalte von STKR1, das wahrscheinlich über das 26S Proteasom abgebaut wird. Da es sich bei STKR1 um ein Phosphoprotein mit SnRK1-Phosphorylierungsmotiv handelt, stellt es sehr wahrscheinlich ein SnRK1-Substrat dar. Allerdings konnte eine SnRK1-vermittelte Phosphorylierung von STKR1 in dieser Arbeit nicht gezeigt werden. Der Verlust von einer Phosphorylierungsstelle beeinflusste die Homo- und Heterodimerisierungsfähigkeit von STKR1 in Hefeinteraktionsstudien, wodurch eine erhöhte Spezifität der Zielgenregulation ermöglicht werden könnte. Außerdem wurden Arabidopsis-Pflanzen mit einer veränderten STKR1-Expression phänotypisch, physiologisch und molekularbiologisch charakterisiert. Während der Verlust der STKR1-Expression zu Pflanzen führte, die sich kaum von Wildtyp-Pflanzen unterschieden, bedingte die konstitutive Überexpression von STKR1 ein stark vermindertes Pflanzenwachstum sowie Entwicklungsverzögerungen hinsichtlich der Blühinduktion und Seneszenz ähnlich wie sie auch bei SnRK1α-Überexpression beschrieben wurden. Pflanzen dieser Linien waren nicht in der Lage Anthocyane zu akkumulieren und enthielten geringere Gehalte an Chlorophyll und Carotinoiden. Neben einem erhöhten nächtlichen Stärkeumsatz waren die Pflanzen durch geringere Saccharosegehalte im Vergleich zum Wildtyp gekennzeichnet. Eine Transkriptomanalyse ergab, dass in den STKR1-überexprimierenden Pflanzen unter Energiemangelbedingungen, hervorgerufen durch eine verlängerte Dunkelphase, eine größere Anzahl an Genen im Vergleich zum Wildtyp differentiell reguliert war als während der Lichtphase. Dies spricht für eine Beteiligung von STKR1 an Prozessen, die während der verlängerten Dunkelphase aktiv sind. Ein solcher ist beispielsweise die SnRK1-Signaltransduktion, die unter energetischem Stress aktiviert wird. Die STKR1Überexpression führte zudem zu einer verstärkten transkriptionellen Induktion von Abwehrassoziierten Genen sowie NAC- und WRKY-Transkriptionsfaktoren nach verlängerter Dunkelphase. Die Transkriptomdaten deuteten auf eine stimulusunabhängige Induktion von Abwehrprozessen hin und konnten eine Erklärung für die phänotypischen und physiologischen Auffälligkeiten der STKR1-Überexprimierer liefern.
Thermophony in real gases
(2016)
A thermophone is an electrical device for sound generation. The advantages of thermophones over conventional sound transducers such as electromagnetic, electrostatic or piezoelectric transducers are their operational principle which does not require any moving parts, their resonance-free behavior, their simple construction and their low production costs.
In this PhD thesis, a novel theoretical model of thermophonic sound generation in real gases has been developed. The model is experimentally validated in a frequency range from 2 kHz to 1 MHz by testing more then fifty thermophones of different materials, including Carbon nano-wires, Titanium, Indium-Tin-Oxide, different sizes and shapes for sound generation in gases such as air, argon, helium, oxygen, nitrogen and sulfur hexafluoride.
Unlike previous approaches, the presented model can be applied to different kinds of thermophones and various gases, taking into account the thermodynamic properties of thermophone materials and of adjacent gases, degrees of freedom and the volume occupied by the gas atoms and molecules, as well as sound attenuation effects, the shape and size of the thermophone surface and the reduction of the generated acoustic power due to photonic emission. As a result, the model features better prediction accuracy than the existing models by a factor up to 100. Moreover, the new model explains previous experimental findings on thermophones which can not be explained with the existing models.
The acoustic properties of the thermophones have been tested in several gases using unique, highly precise experimental setups comprising a Laser-Doppler-Vibrometer combined with a thin polyethylene film which acts as a broadband and resonance-free sound-pressure detector. Several outstanding properties of the thermophones have been demonstrated for the first time, including the ability to generate arbitrarily shaped acoustic signals, a greater acoustic efficiency compared to conventional piezoelectric and electrostatic airborne ultrasound transducers, and applicability as powerful and tunable sound sources with a bandwidth up to the megahertz range and beyond.
Additionally, new applications of thermophones such as the study of physical properties of gases, the thermo-acoustic gas spectroscopy, broad-band characterization of transfer functions of sound and ultrasound detection systems, and applications in non-destructive materials testing are discussed and experimentally demonstrated.
Proteins are amphiphilic and adsorb at liquid interfaces. Therefore, they can be efficient stabilizers of foams and emulsions. β-lactoglobulin (BLG) is one of the most widely studied proteins due to its major industrial applications, in particular in food technology.
In the present work, the influence of different bulk concentration, solution pH and ionic strength on the dynamic and equilibrium pressures of BLG adsorbed layers at the solution/tetradecane (W/TD) interface has been investigated. Dynamic interfacial pressure (Π) and interfacial dilational elastic modulus (E’) of BLG solutions for various concentrations at three different pH values of 3, 5 and 7 at a fixed ionic strength of 10 mM and for a selected fixed concentration at three different ionic strengths of 1 mM, 10 mM and 100 mM are measured by Profile Analysis Tensiometer PAT-1 (SINTERFACE Technologies, Germany). A quantitative data analysis requires additional consideration of depletion due to BLG adsorption at the interface at low protein bulk concentrations. This fact makes experiments more efficient when oil drops are studied in the aqueous protein solutions rather than solution drops formed in oil. On the basis of obtained experimental data, concentration dependencies and the effect of solution pH on the protein surface activity was qualitatively analysed. In the presence of 10 mM buffer, we observed that generally the adsorbed amount is increasing with increasing BLG bulk concentration for all three pH values. The adsorption kinetics at pH 5 result in the highest Π values at any time of adsorption while it exhibits a less active behaviour at pH 3.
Since the experimental data have not been in a good agreement with the classical diffusion controlled model due to the conformational changes which occur when the protein molecules get in contact with the hydrophobic oil phase in order to adapt to the interfacial environment, a new theoretical model is proposed here. The adsorption kinetics data were analysed with the newly proposed model, which is the classical diffusion model but modified by assuming an additional change in the surface activity of BLG molecules when adsorbing at the interface. This effect can be expressed through the adsorption activity constant in the corresponding equation of state. The dilational visco-elasticity of the BLG adsorbed interfacial layers is determined from measured dynamic interfacial tensions during sinusoidal drop area variations. The interfacial tension responses to these harmonic drop oscillations are interpreted with the same thermodynamic model which is used for the corresponding adsorption isotherm.
At a selected BLG concentration of 2×10-6 mol/l, the influence of the ionic strength using different buffer concentration of 1, 10 and 100 mM on the interfacial pressure was studied. It is affected weakly at pH 5, whereas it has a strong impact by increasing buffer concentration at pH 3 and 7. In conclusion, the structure formation of BLG adsorbed layer in the early stage of adsorption at the W/TD interface is similar to those of the solution/air (W/A) surface. However, the equation of state at the W/TD interface provides an adsorption activity constant which is almost two orders of magnitude higher than that for the solution/air surface.
At the end of this work, a new experimental tool called Drop and Bubble Micro Manipulator DBMM (SINTERFACE Technologies, Germany) has been introduced to study the stability of protein covered bubbles against coalescence. Among the available protocols the lifetime between the moment of contact and coalescence of two contacting bubble is determined for different BLG concentrations. The adsorbed amount of BLG is determined as a function of time and concentration and correlates with the observed coalescence behaviour of the contacting bubbles.
The main objective of this dissertation is to analyse prerequisites, expectations, apprehensions, and attitudes of students studying computer science, who are willing to gain a bachelor degree. The research will also investigate in the students’ learning style according to the Felder-Silverman model. These investigations fall in the attempt to make an impact on reducing the “dropout”/shrinkage rate among students, and to suggest a better learning environment.
The first investigation starts with a survey that has been made at the computer science department at the University of Baghdad to investigate the attitudes of computer science students in an environment dominated by women, showing the differences in attitudes between male and female students in different study years. Students are accepted to university studies via a centrally controlled admission procedure depending mainly on their final score at school. This leads to a high percentage of students studying subjects they do not want. Our analysis shows that 75% of the female students do not regret studying computer science although it was not their first choice. And according to statistics over previous years, women manage to succeed in their study and often graduate on top of their class. We finish with a comparison of attitudes between the freshman students of two different cultures and two different university enrolment procedures (University of Baghdad, in Iraq, and the University of Potsdam, in Germany) both with opposite gender majority.
The second step of investigation took place at the department of computer science at the University of Potsdam in Germany and analyzes the learning styles of students studying the three major fields of study offered by the department (computer science, business informatics, and computer science teaching). Investigating the differences in learning styles between the students of those study fields who usually take some joint courses is important to be aware of which changes are necessary to be adopted in the teaching methods to address those different students. It was a two stage study using two questionnaires; the main one is based on the Index of Learning Styles Questionnaire of B. A. Solomon and R. M. Felder, and the second questionnaire was an investigation on the students’ attitudes towards the findings of their personal first questionnaire. Our analysis shows differences in the preferences of learning style between male and female students of the different study fields, as well as differences between students with the different specialties (computer science, business informatics, and computer science teaching).
The third investigation looks closely into the difficulties, issues, apprehensions and expectations of freshman students studying computer science. The study took place at the computer science department at the University of Potsdam with a volunteer sample of students. The goal is to determine and discuss the difficulties and issues that they are facing in their study that may lead them to think in dropping-out, changing the study field, or changing the university. The research continued with the same sample of students (with business informatics students being the majority) through more than three semesters. Difficulties and issues during the study were documented, as well as students’ attitudes, apprehensions, and expectations. Some of the professors and lecturers opinions and solutions to some students’ problems were also documented. Many participants had apprehensions and difficulties, especially towards informatics subjects. Some business informatics participants began to think of changing the university, in particular when they reached their third semester, others thought about changing their field of study. Till the end of this research, most of the participants continued in their studies (the study they have started with or the new study they have changed to) without leaving the higher education system.
In the context of an increasing population of aging people and a shift of medical paradigm towards an individualized medicine in health care, nanostructured lanthanides doped sodium yttrium fluoride (NaYF4) represents an exciting class of upconversion nanomaterials (UCNM) which are suitable to bring forward developments in biomedicine and -biodetection. Despite the fact that among various fluoride based upconversion (UC) phosphors lanthanide doped NaYF4 is one of the most studied upconversion nanomaterial, many open questions are still remaining concerning the interplay of the population routes of sensitizer and activator electronic states involved in different luminescence upconversion photophysics as well as the role of phonon coupling. The collective work aims to explore a detailed understanding of the upconversion mechanism in nanoscaled NaYF4 based materials co-doped with several lanthanides, e.g. Yb3+ and Er3+ as the "standard" type upconversion nanoparticles (UCNP) up to advanced UCNP with Gd3+ and Nd3+. Especially the impact of the crystal lattice structure as well as the resulting lattice phonons on the upconversion luminescence was investigated in detail based on different mixtures of cubic and hexagonal NaYF4 nanoscaled crystals. Three synthesis methods, depending on the attempt of the respective central spectroscopic questions, could be accomplished in the following work. NaYF4 based upconversion nanoparticles doped with several combination of lanthanides (Yb3+, Er3+, Gd3+ and Nd3+) were synthesized successfully using a hydrothermal synthesis method under mild conditions as well as a co-precipitation and a high temperature co-precipitation technique. Structural information were gathered by means of X-ray diffraction (XRD), electron microscopy (TEM), dynamic light scattering (DLS), Raman spectroscopy and inductively coupled plasma atomic emission spectrometry (ICP-OES). The results were discussed in detail with relation to the spectroscopic results. A variable spectroscopic setup was developed for multi parameter upconversion luminescence studies at various temperature 4 K to 328 K. Especially, the study of the thermal behavior of upconversion luminescence as well as time resolved area normalized emission spectra were a prerequisite for the detailed understanding of intramolecular deactivation processes, structural changes upon annealing or Gd3+ concentration, and the role of phonon coupling for the upconversion efficiency. Subsequently it became possible to synthesize UCNP with tailored upconversion luminescence properties. In the end, the potential of UCNP for life science application should be enunciated in context of current needs and improvements of a nanomaterial based optical sensors, whereas the "standard" UCNP design was attuned according to the special conditions in the biological matrix. In terms of a better biocompatibility due to a lower impact on biological tissue and higher penetrability for the excitation light. The first step into this direction was to use Nd3+ ions as a new sensitizer in tridoped NaYF4 based UCNP, whereas the achieved absolute and relative temperature sensitivity is comparable to other types of local temperature sensors in the literature.
The lakes in the Kenyan Rift Valley offer the unique opportunity to study a wide range of hydrochemical environmental conditions, ranging from freshwater to highly saline and alkaline lakes. Because little is known about the hydro- and biogeochemical conditions in the underlying lake sediments, it was the aim of this study to extend the already existing data sets with data from porewater and biomarker analyses. Additionally, reduced sulphur compounds and sulphate reduction rates in the sediment were determined. The new data was used to examine the anthropogenic and microbial influence on the lakes sediments as well as the influence of the water chemistry on the degradation and preservation of organic matter in the sediment column. The lakes discussed in this study are: Logipi, Eight (a small crater lake in the region of Kangirinyang), Baringo, Bogoria, Naivasha, Oloiden, and Sonachi.
The biomarker compositions were similar in all studied lake sediments; nevertheless, there were some differences between the saline and freshwater lakes. One of those differences is the occurrence of a molecule related to β-carotene, which was only found in the saline lakes. This molecule most likely originates from cyanobacteria, single-celled organisms which are commonly found in saline lakes. In the two freshwater lakes, stigmasterol, a sterol characteristic for freshwater algae, was found. In this study, it was shown that Lakes Bogoria and Sonachi can be used for environmental reconstructions with biomarkers, because the absence of oxygen at the lake bottoms slowed the degradation process. Other lakes, like for example Lake Naivasha, cannot be used for such reconstructions, because of the large anthropogenic influence. But the biomarkers proved to be a useful tool to study those anthropogenic influences. Additionally, it was observed that horizons with a high concentration of elemental sulphur can be used as temporal markers. Those horizons were deposited during times when the lake levels were very low. The sulphur was deposited by microorganisms which are capable of anoxygenic photosynthesis or sulphide oxidation.
In this thesis, a route to temperature-, pH-, solvent-, 1,2-diol-, and protein-responsive sensors made of biocompatible and low-fouling materials is established. These sensor devices are based on the sensitivemodulation of the visual band gap of a photonic crystal (PhC), which is induced by the selective binding of analytes, triggering a volume phase transition.
The PhCs introduced by this work show a high sensitivity not only for small biomolecules, but also for large analytes, such as glycopolymers or proteins. This enables the PhC to act as a sensor that detects analytes without the need of complex equipment.
Due to their periodical dielectric structure, PhCs prevent the propagation of specific wavelengths. A change of the periodicity parameters is thus indicated by a change in the reflected wavelengths. In the case explored, the PhC sensors are implemented as periodically structured responsive hydrogels in formof an inverse opal.
The stimuli-sensitive inverse opal hydrogels (IOHs) were prepared using a sacrificial opal template of monodispersed silica particles. First, monodisperse silica particles were assembled with a hexagonally packed structure via vertical deposition onto glass slides. The obtained silica crystals, also named colloidal crystals (CCs), exhibit structural color. Subsequently, the CCs templates were embedded in polymer matrix with low-fouling properties. The polymer matrices were composed of oligo(ethylene glycol) methacrylate derivatives (OEGMAs) that render the hydrogels thermoresponsive. Finally, the silica particles were etched, to produce highly porous hydrogel replicas of the CC. Importantly, the inner structure and thus the ability for light diffraction of the IOHs formed was maintained.
The IOH membrane was shown to have interconnected pores with a diameter as well as interconnections between the pores of several hundred nanometers. This enables not only the detection of small analytes, but also, the detection of even large analytes that can diffuse into the nanostructured IOH membrane. Various recognition unit – analyte model systems, such as benzoboroxole – 1,2-diols, biotin – avidin and mannose – concanavalin A, were studied by incorporating functional
comonomers of benzoboroxole, biotin and mannose into the copolymers. The incorporated recognition units specifically bind to certain low and highmolar mass biomolecules, namely to certain saccharides, catechols, glycopolymers or proteins.
Their specific binding strongly changes the overall hydrophilicity, thus modulating the swelling of the IOH matrices, and in consequence, drastically changes their internal periodicity. This swelling is amplified by the thermoresponsive properties of the polymer matrix. The shift of the interference band gap due to the specific molecular recognition is easily visible by the naked eye (up to 150 nm shifts). Moreover, preliminary trial were attempted to detect even larger entities. Therefore anti-bodies were immobilized on hydrogel platforms via polymer-analogous esterification. These platforms incorporate comonomers made of tri(ethylene glycol) methacrylate end-functionalized with a carboxylic acid. In these model systems, the bacteria analytes are too big to penetrate into the IOH membranes, but can only interact with their surfaces. The selected model bacteria, as Escherichia coli, show a specific affinity to anti-body-functionalized hydrogels. Surprisingly in the case functionalized IOHs, this study produced weak color shifts, possibly opening a path to detect directly living organism, which will need further investigations.
This dissertation examines the impact of the type of referring expression on the acquisition of word order variation in German-speaking preschoolers. A puzzle in the area of language acquisition concerns the production-comprehension asymmetry for non-canonical sentences like "Den Affen fängt die Kuh." (“The monkey, the cow chases.”), that is, preschoolers usually have difficulties in accurately understanding non-canonical sentences approximately until age six (e.g., Dittmar et al., 2008) although they produce non-canonical sentences already around age three (e.g., Poeppel & Wexler, 1993; Weissenborn, 1990). This dissertation investigated the production and comprehension of non-canonical sentences to address this issue.
Three corpus analyses were conducted to investigate the impact of givenness, topic status and the type of referring expression on word order in the spontaneous speech of two- to four-year-olds and the child-directed speech produced by their mothers. The positioning of the direct object in ditransitive sentences was examined; in particular, sentences in which the direct object occurred before or after the indirect object in the sentence-medial positions and sentences in which it occurred in the sentence-initial position. The results reveal similar ordering patterns for children and adults. Word order variation was to a large extent predictable from the type of referring expression, especially with respect to the word order involving the sentence-medial positions. Information structure (e.g., topic status) had an additional impact only on word order variation that involved the sentence-initial position.
Two comprehension experiments were conducted to investigate whether the type of referring expression and topic status influences the comprehension of non-canonical transitive sentences in four- and five-year-olds. In the first experiment, the topic status of the one of the sentential arguments was established via a preceding context sentence, and in the second experiment, the type of referring expression for the sentential arguments was additionally manipulated by using either a full lexical noun phrase (NP) or a personal pronoun. The results demonstrate that children’s comprehension of non-canonical sentences improved when the topic argument was realized as a personal pronoun and this improvement was independent of the grammatical role of the arguments. However, children’s comprehension was not improved when the topic argument was realized as a lexical NP.
In sum, the results of both production and comprehension studies support the view that referring expressions may be seen as a sentence-level cue to word order and to the information status of the sentential arguments. The results highlight the important role of the type of referring expression on the acquisition of word order variation and indicate that the production-comprehension asymmetry is reduced when the type of referring expression is considered.
Light-triggered release of bioactive compounds from HA/PLL multilayer films for stimulation of cells
(2016)
The concept of targeting cells and tissues by controlled delivery of molecules is essential in the field of biomedicine. The layer-by-layer (LbL) technology for the fabrication of polymer multilayer films is widely implemented as a powerful tool to assemble tailor-made materials for controlled drug delivery. The LbL films can as well be engineered to act as mimics of the natural cellular microenvironment. Thus, due to the myriad possibilities such as controlled cellular adhesion and drug delivery offered by LbL films, it becomes easily achievable to direct the fate of cells by growing them on the films.
The aim of this work was to develop an approach for non-invasive and precise control of the presentation of bioactive molecules to cells. The strategy is based on employment of the LbL films, which function as support for cells and at the same time as reservoirs for bioactive molecules to be released in a controlled manner. UV light is used to trigger the release of the stored ATP with high spatio-temporal resolution. Both physico-chemical (competitive intermolecular interactions in the film) and biological aspects (cellular response and viability) are addressed in this study.
Biopolymers hyaluronic acid (HA) and poly-L-lysine (PLL) were chosen as the building blocks for the LbL film assembly. Poor cellular adhesion to native HA/PLL films as well as significant degradation by cells within a few days were shown. However, coating the films with gold nanoparticles not only improved cellular adhesion and protected the films from degradation, but also formed a size-exclusion barrier with adjustable cut-off in the size range of a few tens of kDa.
The films were shown to have high reservoir capacity for small charged molecules (reaching mM levels in the film). Furthermore, they were able to release the stored molecules in a sustained manner. The loading and release are explained by a mechanism based on interactions between charges of the stored molecules and uncompensated charges of the biopolymers in the film. Charge balance and polymer dynamics in the film play the pivotal role.
Finally, the concept of light-triggered release from the films has been proven using caged ATP loaded into the films from which ATP was released on demand. ATP induces a fast cellular response, i.e. increase in intracellular [Ca2+], which was monitored in real-time. Limitations of the cellular stimulation by the proposed approach are highlighted by studying the stimulation as a function of irradiation parameters (time, distance, light power). Moreover, caging molecules bind to the film stronger than ATP does, which opens new perspectives for the use of the most diverse chemical compounds as caging molecules.
Employment of HA/PLL films as a nouvelle support for cellular growth and hosting of bioactive molecules, along with the possibility to stimulate individual cells using focused light renders this approach highly efficient and unique in terms of precision and spatio-temporal resolution among those previously described. With its high potential, the concept presented herein provides the foundation for the design of new intelligent materials for single cell studies, with the focus on tissue engineering, diagnostics, and other cell-based applications.
Recently, due to an increasing demand on functionality and flexibility, beforehand isolated systems have become interconnected to gain powerful adaptive Systems of Systems (SoS) solutions with an overall robust, flexible and emergent behavior. The adaptive SoS comprises a variety of different system types ranging from small embedded to adaptive cyber-physical systems. On the one hand, each system is independent, follows a local strategy and optimizes its behavior to reach its goals. On the other hand, systems must cooperate with each other to enrich the overall functionality to jointly perform on the SoS level reaching global goals, which cannot be satisfied by one system alone. Due to difficulties of local and global behavior optimizations conflicts may arise between systems that have to be solved by the adaptive SoS.
This thesis proposes a modeling language that facilitates the description of an adaptive SoS by considering the adaptation capabilities in form of feedback loops as first class entities. Moreover, this thesis adopts the Models@runtime approach to integrate the available knowledge in the systems as runtime models into the modeled adaptation logic. Furthermore, the modeling language focuses on the description of system interactions within the adaptive SoS to reason about individual system functionality and how it emerges via collaborations to an overall joint SoS behavior. Therefore, the modeling language approach enables the specification of local adaptive system behavior, the integration of knowledge in form of runtime models and the joint interactions via collaboration to place the available adaptive behavior in an overall layered, adaptive SoS architecture.
Beside the modeling language, this thesis proposes analysis rules to investigate the modeled adaptive SoS, which enables the detection of architectural patterns as well as design flaws and pinpoints to possible system threats. Moreover, a simulation framework is presented, which allows the direct execution of the modeled SoS architecture. Therefore, the analysis rules and the simulation framework can be used to verify the interplay between systems as well as the modeled adaptation effects within the SoS. This thesis realizes the proposed concepts of the modeling language by mapping them to a state of the art standard from the automotive domain and thus, showing their applicability to actual systems. Finally, the modeling language approach is evaluated by remodeling up to date research scenarios from different domains, which demonstrates that the modeling language concepts are powerful enough to cope with a broad range of existing research problems.
The increase in atmospheric methane concentration, which is determined by an imbalance between its sources and sinks, has led to investigations of the methane cycle in various environments. Aquatic environments are of an exceptional interest due to their active involvement in methane cycling worldwide and in particular in areas sensitive to climate change. Furthermore, being connected with each other aquatic environments form networks that can be spread on vast areas involving marine, freshwater and terrestrial ecosystems. Thus, aquatic systems have a high potential to translate local or regional environmental and subsequently ecosystem changes to a bigger scale. Many studies neglect this connectivity and focus on individual aquatic or terrestrial ecosystems.
The current study focuses on environmental controls of the distribution and aerobic oxidation of methane at the example of two aquatic ecosystems. These ecosystems are Arctic fresh water bodies and the Elbe estuary which represent interfaces between freshwater-terrestrial and freshwater-marine environments, respectively.
Arctic water bodies are significant atmospheric sources of methane. At the same time the methane cycle in Arctic water bodies is strongly affected by the surrounding permafrost environment, which is characterized by high amounts of organic carbon. The results of this thesis indicate that the methane concentrations in Arctic lakes and streams substantially vary between each other being regulated by local landscape features (e.g. floodplain area) and the morphology of the water bodies (lakes, streams and river). The highest methane concentrations were detected in the lake outlets and in a floodplain lake complex. In contrast, the methane concentrations measured at different sites of the Lena River did not vary substantially. The lake complexes in comparison to the Lena River, thus, appear as more individual and heterogeneous systems with a pronounced imprint of the surrounding soil environment. Furthermore, connected with each other Arctic aquatic environments have a large potential to transport methane from methane-rich water bodies such as streams and floodplain lakes to aquatic environments relatively poor in methane such as the Lena River.
Estuaries represent hot spots of oceanic methane emissions. Also, estuaries are intermediate zones between methane-rich river water and methane depleted marine water. Substantiated through this thesis at the example of the Elbe estuary, the methane distribution in estuaries, however, cannot entirely be described by the conservative mixing model i.e. gradual decrease from the freshwater end-member to the marine water end-member. In addition to the methane-rich water from the Elbe River mouth substantial methane input occurs from tidal flats, areas of significant interaction between aquatic and terrestrial environments. Thus, this study demonstrates the complex interactions and their consequences for the methane distribution within estuaries. Also it reveals how important it is to investigate estuaries at larger spatial scales.
Methane oxidation (MOX) rates are commonly correlated with methane concentrations. This was shown in previous research studies and was substantiated by the present thesis. In detail, the highest MOX rates in the Arctic water bodies were detected in methane-rich streams and in the floodplain area while in the Elbe estuary the highest MOX rates were observed at the coastal stations. However, in these bordering environments, MOX rates are affected not only via the regulation through methane concentrations. The MOX rates in the Arctic lakes were shown to be also dependent on the abundance and community composition of methane-oxidising bacteria (MOB), that in turn are controlled by local landscape features (regardless of the methane concentrations) and by the transport of MOB between neighbouring environments. In the Elbe estuary, the MOX rates in addition to the methane concentrations are largely affected by the salinity, which is in turn regulated by the mixing of fresh- and marine waters. The magnitude of the salinity impact on MOX rates thereby depends on the MOB community composition and on the rate of the salinity change.
This study extends our knowledge of environmental controls of methane distribution and aerobic methane oxidation in aquatic environments. It also illustrates how important it is to investigate complex ecosystems rather than individual ecosystems to better understand the functioning of whole biomes.
The energy sector is both affected by climate change and a key sector for climate protection measures. Energy security is the backbone of our modern society and guarantees the functioning of most critical infrastructure. Thus, decision makers and energy suppliers of different countries should be familiar with the factors that increase or decrease the susceptibility of their electricity sector to climate change. Susceptibility means socioeconomic and structural characteristics of the electricity sector that affect the demand for and supply of electricity under climate change. Moreover, the relevant stakeholders are supposed to know whether the given national energy and climate targets are feasible and what needs to be done in order to meet these targets. In this regard, a focus should be on the residential building sector as it is one of the largest energy consumers and therefore emitters of anthropogenic CO 2 worldwide.
This dissertation addresses the first aspect, namely the susceptibility of the electricity sector, by developing a ranked index which allows for quantitative comparison of the electricity sector susceptibility of 21 European countries based on 14 influencing factors. Such a ranking has not been completed to date. We applied a sensitivity analysis to test the relative effect of each influencing factor on the susceptibility index ranking. We also discuss reasons for the ranking position and thus the susceptibility of selected countries. The second objective, namely the impact of climate change on the energy demand of buildings, is tackled by means of a new model with which the heating and cooling energy demand of residential buildings can be estimated. We exemplarily applied the model to Germany and the Netherlands. It considers projections of future changes in population, climate and the insulation standards of buildings, whereas most of the existing studies only take into account fewer than three different factors that influence the future energy demand of buildings. Furthermore, we developed a comprehensive retrofitting algorithm with which the total residential building stock can be modeled for the first time for each year in the past and future.
The study confirms that there is no correlation between the geographical location of a country and its position in the electricity sector susceptibility ranking. Moreover, we found no pronounced pattern of susceptibility influencing factors between countries that ranked higher or lower in the index. We illustrate that Luxembourg, Greece, Slovakia and Italy are the countries with the highest electricity sector susceptibility. The electricity sectors of Norway, the Czech Republic, Portugal and Denmark were found to be least susceptible to climate change. Knowledge about the most important factors for the poor and good ranking positions of these countries is crucial for finding adequate adaptation measures to reduce the susceptibility of the electricity sector. Therefore, these factors are described within this study.
We show that the heating energy demand of residential buildings will strongly decrease in both Germany and the Netherlands in the future. The analysis for the Netherlands focused on the regional level and a finer temporal resolution which revealed strong variations in the future heating energy demand changes by province and by month. In the German study, we additionally investigated the future cooling energy demand and could demonstrate that it will only slightly increase up to the middle of this century. Thus, increases in the cooling energy demand are not expected to offset reductions in heating energy demand. The main factor for substantial heating energy demand reductions is the retrofitting of buildings. We are the first to show that the given German and Dutch energy and climate targets in the building sector can only be met if the annual retrofitting rates are substantially increased. The current rate of only about 1 % of the total building stock per year is insufficient for reaching a nearly zero-energy demand of all residential buildings by the middle of this century. To reach this target, it would need to be at least tripled. To sum up, this thesis emphasizes that country-specific characteristics are decisive for the electricity sector susceptibility of European countries. It also shows for different scenarios how much energy is needed in the future to heat and cool residential buildings. With this information, existing climate mitigation and adaptation measures can be justified or new actions encouraged.
Das biogene Amin Serotonin (5-Hydroxytryptamin, 5-HT) agiert als wichtiger chemischer Botenstoff bei einer Vielzahl von Organismen. Das durch 5 HT vermittelte Signal wird dabei durch spezifische Rezeptoren wahrgenommen und in eine zelluläre Reaktion umgesetzt. Diese 5 HT Rezeptoren gehören überwiegend zur Familie der G Protein gekoppelten Rezeptoren (GPCRs). Die Honigbiene Apis mellifera bietet unter anderem aufgrund ihrer eusozialen Lebensweise vielfältige Ansatzpunkte zur Erforschung der Funktionen des serotonergen Systems in Insekten. Bei A. mellifera wurden bereits vier 5-HT-Rezeptor-Subtypen beschrieben und molekular sowie pharmakologisch charakterisiert: Am5 HT1A, Am5 HT2α, Am5 HT2β und Am5 HT7. Ziel dieser Arbeit war es, gewebespezifische sowie alters- und tageszeitabhängige Expressionsmuster der 5 HT Rezeptor-Subtypen zu untersuchen, um zu einem umfassenden Verständnis des serotonergen Systems der Honigbiene beizutragen und eine Basis zur Hypothesenentwicklung für mögliche physiologische Funktionen zu schaffen.
Es wurde die Expression der 5 HT Rezeptorgene sowohl im zentralen Nervensystem, als auch in Teilen des Verdauungs-, Exkretions- und Speicheldrüsensystems gemessen. Dabei konnte gezeigt werden, dass die untersuchten 5-HT-Rezeptor-Subtypen generell weit im Organismus der Honigbiene verbreitet sind. Interessanterweise unterschieden sich die untersuchten Gewebe hinsichtlich der mRNA-Expressionsmuster der untersuchten Rezeptoren. Während beispielsweise im Gehirn Am5 ht1A und Am5 ht7 stärker als Am5 ht2α und Am5 ht2β exprimiert wurden, zeigte sich in Darmgewebe ein umgekehrtes Muster.
Es war bereits bekannt, dass es bei der Expression der Am5-ht2-Gene zu alternativem Spleißen kommt. Dies führt zur Entstehung der verkürzten mRNA-Varianten Am5 ht2αΔIII und Am5 ht2βΔII. Die daraus resultierenden Proteine können nicht als funktionelle GPCRs agieren. Es konnte gezeigt werden, dass diese verkürzten Spleißvarianten dennoch ubiquitär in der Honigbiene exprimiert werden. Bemerkenswerterweise wurden gewebeübergreifende Ähnlichkeiten der Expressionsmuster der Spleißvarianten gegenüber deren zugehörigen Volllängenvarianten festgestellt, welche auf Funktionen der verkürzten Varianten in vivo hindeuten.
Im Hinblick auf die bei A. mellifera hauptsächlich altersbedingte Arbeitsteilung wurde die Expression der 5 HT Rezeptor-Subtypen in Gehirnen von unterschiedlich alten Arbeiterinnen mit unterschiedlichen sozialen Rollen verglichen. Während auf mRNA-Ebene keines der vier 5 HT Rezeptor-Subtypen eine altersabhängig unterschiedliche Expression zeigte, konnte für das Am5-HT1A-Protein eine höhere Konzentration in den Gehirnen älterer Tiere gefunden werden. Dies deutet auf eine posttranskriptionale Regulation der 5 HT1A Rezeptorexpression hin, welche im Zusammenhang mit der Arbeitsteilung stehen könnte.
Es erfolgte die Untersuchung tageszeitlicher Änderungen sowohl der Expression der 5 HT Rezeptor-Subtypen, als auch des biogenen Amins 5 HT selbst. Während es in den Gehirnen von Arbeiterinnen, welche unter natürlichen Bedingungen gehalten wurden, zu keiner tageszeitabhängigen Veränderung des 5 HT-Titers kam, zeigte die mRNA-Expression von Am5-ht2α und Am5-ht2β eine periodische Oszillation mit Zunahme während des Tages und Abnahme während der Nacht. Diese Regulation wird durch externe Faktoren hervorgerufen und ist nicht auf einen endogenen circadianen Rhythmus zurückzuführen. Dies ging aus der Wiederholung der Expressionsmessungen an Gehirnen von Bienen, welche unter konstanten Laborbedingungen gehalten wurden, hervor.
Weiterhin wurde die Beteiligung des serotonergen Systems an der Steuerung von Aspekten des circadianen lokomotorischen Aktivitätsrhythmus anhand von Verhaltensexperimenten untersucht. Mit 5 HT gefütterte Arbeiterinnen zeigten dabei unter konstanten Bedingungen eine längere Periode des Aktivitätsrhythmus als Kontrolltiere. Dies deutet auf einen Einfluss von 5 HT auf die Modulation der Synchronisation der inneren Uhr hin.
Die vorliegenden Ergebnisse tragen wesentlich zum tieferen Verständnis des serotonergen Systems der Honigbiene bei und bieten Ansatzpunkte für weitergehende Studien zur Funktion von 5 HT im Zusammenhang mit der Modulation von physiologischen Prozessen, Arbeitsteilung und circadianen Rhythmen.
Die sensorisch einwandfreie, konstant gute Qualität von Backprodukten, die beim Verbraucher einen hohen Stellenwert hat, wird maßgeblich durch den Gehalt endogener Getreideenzyme beeinflusst. Seit dem Auftreten züchtungsbedingter Enzymdefizite ist der Einsatz technischer Enzyme zur Gewährleistung dieser geforderten Qualität eine feste Größe in der Backwarenindustrie. Lebensmittelrechtlich werden technische Enzyme nicht als Zutat betrachtet, da sie theoretisch während des Backprozesses umgesetzt werden und im Endprodukt keine technologische Wirkung mehr zeigen. Vor allem in gebackenen Produkten bedarf es der Prüfung, dass die eingesetzten technischen Enzyme nicht mehr als Zutat vorliegen und sich somit einer potentiellen Deklarationspflicht entziehen. Zur Gewährleistung der Wirtschaftlichkeit muss der quantitative Einsatz technischer Enzyme in der Backwarenindustrie gesteuert werden, um optimale Effekte zu erzielen und Kosten zu sparen. Ziel dieser Arbeit war daher die Entwicklung eines Analysenverfahrens, das den simultanen Nachweis verschiedener technischer Enzyme und deren Quantifizierung im Spurenbereich auch in gebackenen Produkten ermöglicht.
Für die Einschätzung der Wirkung der technischen Enzyme Fungamyl (Novozymes), Amylase TXL (ASA Spezialenzyme GmbH) sowie Lipase FE-01 (ASA Spezialenzyme GmbH) wurden Backversuche durchgeführt, die zeigten, dass Fungamyl und Amylase TXL zu einer verbesserten Brotqualität (Volumenausbeute, Feuchtegehalt, Sensorik) beitrugen. Die Zugabe der Lipase FE-01 führte zu einer vermehrten Bildung freier Fettsäuren und wirkte sich negativ auf die sensorische Brotqualität aus. Dieser bisher nicht beschriebene Effekt konnte auf die Nutzung eines Spezialöls als Backzutat zurückgeführt werden, welches ausschließlich aus gesättigten Fettsäuren besteht. Dies bestätigt die Bedeutung der Auswahl eines geeigneten Fettes beim Zusatz technischer Lipase zum Backprozess.
Um die in Fungamyl und Lipase FE-01 enthaltenen Enzyme zu identifizieren, wurden SDS-PAGE und anschließender In-Gel-Verdau angewendet um die Analyse proteolytisch gespaltener Proteine mit MALDI-TOF-MS zu ermöglichen. Es konnte gezeigt werden, dass Fungamyl ein Gemisch aus 9,8 % alpha-Amylase (Aspergillus oryzae) und 5,2 % Endo-1,4-Xylanase (Thermomyces lanuginosus) enthält. Lipase FE-01 besteht aus der Lipase (Thermomyces lanuginosus), Amylase TXL wurde als alpha-Amylase (Aspergillus oryzae) identifiziert.
Zur Analyse der technischen Enzyme in Backwaren wurde aufgrund seiner Robustheit und Sensitivität das Verfahren der LC-MS/MS gewählt. Die Entwicklung einer solchen Methode zur Detektion spezifischer Peptide ermöglichte den qualitativen Nachweis der 3 Enzyme alpha-Amylase (Aspergillus oryzae), Endo-1,4-Xylanase (Thermomyces lanuginosus) und Lipase (Thermomyces lanuginosus). Durch eine lineare Kalibrierung aus synthetisch hergestellten Peptiden unter Einbeziehung eines Protein-Internen-Standards sowie isotopenmarkierter Peptidstandards erfolgte darüber hinaus die quantitative Bestimmung in selbst hergestellten Referenzmaterialien (Weizenmehl, Toastbrot und Biskuitkeks). In weniger als 20 Minuten Messzeit kann das Enzym alpha-Amylase ab einer Konzentration von 2,58 mg/kg (Mehl, Keks), bzw. 7,61 mg/kg (Brot) quantitativ nachgewiesen werden. Zeitgleich können die Enzyme Endo-1,4-Xylanase ab einer Konzentration von 7,75 mg/kg (Brot), 3,64 mg/kg (Keks) bzw. 15,60 mg/kg (Mehl) sowie Lipase ab einer Konzentration von 1,26 mg/kg (Mehl, Keks), bzw. 2,68 mg/kg (Brot) quantifiziert werden. Die Methode wurde nach allgemein verwendeten Richtlinien im Zuge einer Validierung statistisch geprüft und lieferte sehr robuste und reproduzierbare quantitative Werte mit Wiederfindungsraten zwischen 50 % und 122 %. Das primäre Ziel dieser Arbeit, die Entwicklung eines quantitativen Multiparameterverfahrens zum Nachweis technischer Enzyme in Backwaren, wurde somit erfolgreich umgesetzt.
Earthquakes deform Earth's surface, building long-lasting topographic features and contributing to landscape and mountain formation.
However, seismic waves produced by earthquakes may also destabilize hillslopes, leading to large amounts of soil and bedrock moving downslope. Moreover, static deformation and shaking are suspected to damage the surface bedrock and therefore alter its future properties, affecting hydrological and erosional dynamics. Thus, earthquakes participate both in mountain building and stimulate directly or indirectly their erosion. Moreover, the impact of earthquakes on hillslopes has important implications for the amount of sediment and organic matter delivered to rivers, and ultimately to oceans, during episodic catastrophic seismic crises, the magnitude of life and property losses associated with landsliding, the perturbation and recovery of landscape properties after shaking, and the long term topographic evolution of mountain belts. Several of these aspects have been addressed recently through individual case studies but additional data compilation as well as theoretical or numerical modelling are required to tackle these issues in a more systematic and rigorous manner.
This dissertation combines data compilation of earthquake characteristics, landslide mapping, and seismological data interpretation with physically-based modeling in order to address how earthquakes impact on erosional processes and landscape evolution. Over short time scales (10-100 s) and intermediate length scales (10 km), I have attempted to improve our understanding and ability to predict the amount of landslide debris triggered by seismic shaking in epicentral areas. Over long time scales (1-100 ky) and across a mountain belt (100 km) I have modeled the competition between erosional unloading and building of topography associated with earthquakes. Finally, over intermediate time scales (1-10 y) and at the hillslope scale (0.1-1 km) I have collected geomorphological and seismological data that highlight persistent effects of earthquakes on landscape properties and behaviour.
First, I compiled a database on earthquakes that produced significant landsliding, including an estimate of the total landslide volume and area, and earthquake characteristics such as seismic moment and source depth. A key issue is the accurate conversion of landslide maps into volume estimates. Therefore I also estimated how amalgamation - when mapping errors lead to the bundling of multiple landslide into a single polygon - affects volume estimates from various earthquake-induced landslide inventories and developed an algorithm to automatically detect this artifact. The database was used to test a physically-based prediction of the total landslide area and volume caused by earthquakes, based on seismological scaling relationships and a statistical description of the landscape properties. The model outperforms empirical fits in accuracy, with 25 out of 40 cases well predicted, and allows interpretation of many outliers in physical terms. Apart from seismological complexities neglected by the model I found that exceptional rock strength properties or antecedent conditions may explain most outliers.
Second, I assessed the geomorphic effects of large earthquakes on landscape dynamics by surveying the temporal evolution of precipitation-normalized landslide rate. I found strongly elevated landslide rates following earthquakes that progressively recover over 1 to 4 years, indicating that regolith strength drops and recovers. The relaxation is clearly non-linear for at least one case, and does not seem to correlate with coseismic landslide reactivation, water table level increase or tree root-system recovery. I suggested that shallow bedrock is damaged by the earthquake and then heals on annual timescales. Such variations in ground strength must be translated into shallow subsurface seismic velocities that are increasingly surveyed with ambient seismic noise correlations. With seismic noise autocorrelation I computed the seismic velocity in the epicentral areas of three earthquakes where I constrained a change in landslide rate. We found similar recovery dynamics and timescales, suggesting that seismic noise correlation techniques could be further developed to meaningfully assess ground strength variations for landscape dynamics. These two measurements are also in good agreement with the temporal dynamics of post-seismic surface displacement measured by GPS. This correlation suggests that the surface healing mechanism may be driven by tectonic deformation, and that the surface regolith and fractured bedrock may behave as a granular media that slowly compacts as it is sheared or vibrated.
Last, I compared our model of earthquake-induced landsliding with a standard formulation of surface deformation caused by earthquakes to understand which parameters govern the competition between the building and destruction of topography caused by earthquakes. In contrast with previous studies I found that very large (Mw>8) earthquakes always increase the average topography, whereas only intermediate (Mw ~ 7) earthquakes in steep landscapes may reduce topography. Moreover, I illustrated how the net effect of earthquakes varies with depth or landscape steepness implying a complex and ambivalent role through the life of a mountain belt. Further I showed that faults producing a Gutenberg-Richter distribution of earthquake sizes, will limit topography over a larger range of fault sizes than faults producing repeated earthquakes with a characteristic size.
The relationship between climate and forest productivity is an intensively studied subject in forest science. This Thesis is embedded within the general framework of future forest growth under climate change and its implications for the ongoing forest conversion. My objective is to investigate the future forest productivity at different spatial scales (from a single specific forest stand to aggregated information across Germany) with focus on oak-pine forests in the federal state of Brandenburg. The overarching question is: how are the oak-pine forests affected by climate change described by a variety of climate scenarios. I answer this question by using a model based analysis of tree growth processes and responses to different climate scenarios with emphasis on drought events. In addition, a method is developed which considers climate change uncertainty of forest management planning.
As a first 'screening' of climate change impacts on forest productivity, I calculated the change in net primary production on the base of a large set of climate scenarios for different tree species and the total area of Germany. Temperature increases up to 3 K lead to positive effects on the net primary production of all selected tree species. But, in water-limited regions this positive net primary production trend is dependent on the length of drought periods which results in a larger uncertainty regarding future forest productivity. One of the regions with the highest uncertainty of net primary production development is the federal state of Brandenburg.
To enhance the understanding and ability of model based analysis of tree growth sensitivity to drought stress two water uptake approaches in pure pine and mixed oak-pine stands are contrasted. The first water uptake approach consists of an empirical function for root water uptake. The second approach is more mechanistic and calculates the differences of soil water potential along a soil-plant-atmosphere continuum. I assumed the total root resistance to vary at low, medium and high total root resistance levels. For validation purposes three data sets on different tree growth relevant time scales are used. Results show that, except the mechanistic water uptake approach with high total root resistance, all transpiration outputs exceeded observed values. On the other hand high transpiration led to a better match of observed soil water content. The strongest correlation between simulated and observed annual tree ring width occurred with the mechanistic water uptake approach and high total root resistance. The findings highlight the importance of severe drought as a main reason for small diameter increment, best supported by the mechanistic water uptake approach with high root resistance. However, if all aspects of the data sets are considered no approach can be judged superior to the other. I conclude that the uncertainty of future productivity of water-limited forest ecosystems under changing environmental conditions is linked to simulated root water uptake.
Finally my study aimed at the impacts of climate change combined with management scenarios on an oak-pine forest to evaluate growth, biomass and the amount of harvested timber. The pine and the oak trees are 104 and 9 years old respectively. Three different management scenarios with different thinning intensities and different climate scenarios are used to simulate the performance of management strategies which explicitly account for the risks associated with achieving three predefined objectives (maximum carbon storage, maximum harvested timber, intermediate). I found out that in most cases there is no general management strategy which fits best to different objectives. The analysis of variance in the growth related model outputs showed an increase of climate uncertainty with increasing climate warming. Interestingly, the increase of climate-induced uncertainty is much higher from 2 to 3 K than from 0 to 2 K.
Die vorliegende Untersuchung verfolgt das Ziel, kulturelle und religiöse Aspekte der Erneuerung jüdischen Lebens in Berlin seit 1989 zu erforschen. Die Entwicklungen der jüdischen Gemeinschaft in der Hauptstadt seit dem Fall der Mauer und dem Zusammenbruch der Sowjetunion führen zur Wiederannäherung eines Teils der jüdischen Bevölkerung in Deutschland an die eigene Kultur, Religion und Geschichte. Dabei kommt die Pluralität der kulturellen, literarischen und religiösen Ausdrucksformen der jüdischen Identitäten zum Vorschein. Die Arbeit verdeutlicht diese in Berlin nach 1989 einsetzende kulturelle und religiöse „Renaissance“. Vier wichtige Punkte kennzeichnen das jüdische Leben in Berlin nach 1989. Erstens gewinnt Deutschland seit der Wiedervereinigung eine neue Rolle als mögliches Einwanderungsland für Juden. Vor allem mit der massiven jüdischen Einwanderung aus den Staaten der ehemaligen Sowjetunion seit den 1990er Jahren wird Deutschland allmählich als wichtiges Zentrum in der europäischen Diaspora anerkannt. Zweitens bleibt zwar die Shoah tief verankert im Gedächtnis der jüdischen Gemeinschaft; die meisten Kinder oder Enkelkinder von Überlebenden der Shoah weigern sich jedoch, ihre jüdische Identität exklusiv durch die Shoah zu definieren. Sie gründen zur Wiederentdeckung und Forderung ihres kulturellen, religiösen und historischen Erbes jüdische Gruppen und Einrichtungen in Berlin, die in den meisten Fällen alternativ zur Jüdischen Gemeinde entstehen: Künstlergruppen, jüdische Kulturvereine, Konferenzen und Podiumsdiskussionen, religiöse Kongregationen und Lernhäuser. Damit – und dies ist der dritte Punkt – verliert zwar die offizielle Jüdische Gemeinde an Bedeutung als einzige Vertreterin der jüdischen Gemeinschaft Berlins; diese kulturelle und religiöse „Renaissance“ außerhalb der offiziellen Strukturen der Gemeinde bedeutet aber auch eine wachsende Pluralität und Diversifizierung der jüdischen Gemeinschaft in Berlin. Viertens spielt Berlin die Hauptrolle in diesem Prozess. Heute werden viele ehemalige jüdische Orte neu belebt: Synagogen werden wiederentdeckt und renoviert, Denk- und Mahnmale gebaut, Stadtführungen auf der Spur des „jüdischen Berlins“ organisiert, Rabbinerseminare neu gegründet. Die Topographie Berlins bildet auch eine Inspirationsquelle für jüdische (und nichtjüdische) Schriftsteller und Künstler. Die Analyse dieser nach 1989 entstandenen religiösen Initiativen, literarischen Werke und kulturellen Produktionen dient dazu, Aspekte der kulturellen und religiösen „Renaissance“ in Berlin näher zu verdeutlichen.
This thesis presents new approaches of SAR methods and their application to tectonically active systems and related surface deformation. With 3 publications two case studies are presented:
(1) The coseismic deformation related to the Nura earthquake (5th October 2008, magnitude Mw 6.6) at the eastern termination of the intramontane Alai valley. Located between the southern Tien Shan and the northern Pamir the coseismic surface displacements are analysed using SAR (Synthetic Aperture RADAR) data. The results show clear gradients in the vertical and horizontal directions along a complex pattern of surface ruptures and active faults. To integrate and to interpret these observations in the context of the regional active tectonics a SAR data analysis is complemented with seismological data and geological field observations. The main moment release of the Nura earthquake appears to be on the Pamir Frontal thrust, while the main surface displacements and surface rupture occurred in the footwall and along of the NE–SW striking Irkeshtam fault. With InSAR data from ascending and descending satellite tracks along with pixel offset measurements the Nura earthquake source is modelled as a segmented rupture. One fault segment corresponds to high-angle brittle faulting at the Pamir Frontal thrust and two more fault segments show moderate-angle and low-friction thrusting at the Irkeshtam fault. The integrated analysis of the coseismic deformation argues for a rupture segmentation and strain partitioning associated to the earthquake. It possibly activated an orogenic wedge in the easternmost segment of the Pamir-Alai collision zone. Further, the style of the segmentation may be associated with the presence of Paleogene evaporites.
(2) The second focus is put on slope instabilities and consequent landslides in the area of prominent topographic transition between the Fergana basin and high-relief Alai range. The Alai range constitutes an active orogenic wedge of the Pamir – Tien Shan collision zone that described as a progressively northward propagating fold-and-thrust belt. The interferometric analysis of ALOS/PALSAR radar data integrates a period of 4 years (2007-2010) based on the Small Baseline Subset (SBAS) time-series technique to assess surface deformation with millimeter surface change accuracy. 118 interferograms are analyzed to observe spatially-continuous movements with downslope velocities up to 71 mm/yr. The obtained rates indicate slow movement of the deep-seated landslides during the observation time. We correlated these movements with precipitation and seismic records. The results suggest that the deformation peaks correlate with rainfall in the 3 preceding months and with one earthquake event. In the next step, to understand the spatial pattern of landslide processes, the tectonic morphologic and lithologic settings are combined with the patterns of surface deformation. We demonstrate that the lithological and tectonic structural patterns are the main controlling factors for landslide occurrence and surface deformation magnitudes. Furthermore active contractional deformation in the front of the orogenic wedge is the main mechanism to sustain relief. Some of the slower but continuously moving slope instabilities are directly related to tectonically active faults and unconsolidated young Quaternary syn-orogenic sedimentary sequences. The InSAR observed slow moving landslides represent active deep-seated gravitational slope deformation phenomena which is first time observed in the Tien Shan mountains. Our approach offers a new combination of InSAR techniques and tectonic aspects to localize and understand enhanced slope instabilities in tectonically active mountain fronts in the Kyrgyz Tien Shan.
Age of acquisition (AOA) is a psycholinguistic variable that significantly influences behavioural measures (response times and accuracy rates) in tasks that require lexical and semantic processing. Its origin is – unlike the origin of semantic typicality (TYP), which is assumed at the semantic level – controversially discussed. Different theories propose AOA effects to originate either at the semantic level or at the link between semantics and phonology (lemma-level).
The dissertation aims at investigating the influence of AOA and its interdependence with the semantic variable TYP on particularly semantic processing in order to pinpoint the origin of AOA effects. Therefore, three studies have been conducted that considered the variables AOA and TYP in semantic processing tasks (category verifications and animacy decisions) by means of behavioural and partly electrophysiological (ERP) data and in different populations (healthy young and elderly participants and in semantically impaired individuals with aphasia (IWA)).
The behavioural and electrophysiological data of the three studies provide evidence for distinct processing levels of the variables AOA and TYP. The data further support previous assumptions on a semantic origin for TYP but question the same for AOA. The findings, however, support an origin of AOA effects at the transition between the word form (phonology) and the semantic level that can be captured at the behavioural but not at the electrophysiological level.
Background: The engagement in aggressive behavior in middle childhood is linked to the development of severe problems in later life. Thus, identifying factors and processes that con-tribute to the continuity and increase of aggression in middle childhood is essential in order to facilitate the development of intervention programs. The present PhD thesis aimed at expand-ing the understanding of the development of aggression in middle childhood by examining risk factors in the intrapersonal and interpersonal domains as well as the interplay between these factors: Maladaptive anger regulation was examined as an intrapersonal risk factor; processes that occur in the peer context (social rejection and peer socialization) were included as interpersonal risk factors. In addition, in order to facilitate the in situ assessment of anger regulation strategies, an observational measure of anger regulation was developed and validated.
Method: The research aims were addressed within the scope of four articles. Data from two measurement time points about ten months apart were available for the analyses. Participants were elementary school children aged from 6 to 10 years at T1 and 7 to 11 years at T2. The first article was based on cross-sectional analyses including only the first time point; in the remaining three articles longitudinal associations across the two time points were analyzed. The first two articles were concerned with the development and cross-sectional as well as longitudinal validation of observational measure of anger regulation in middle childhood in a sample of 599 children. Using the same sample, the third article investigated the longitudinal link between maladaptive anger regulation and aggression considering social rejection as a mediating variable. The frequency as well as different functions of aggression (reactive and proactive) were included as outcomes measures. The fourth article examined the influence of class-level aggression on the development of different forms of aggression (relational and physical) over time under consideration of differences in initial individual aggression in a sample of 1,284 children. In addition, it was analyzed if the path from aggression to social rejection varies as a function of class-level aggression.
Results: The first two articles revealed that the observational measure of anger regulation developed for the purpose of this research was cross-sectionally related to anger reactivity, aggression and social rejection as well as longitudinally related to self-reported anger regula-tion. In the third article it was found that T1 maladaptive anger regulation showed no direct link to T2 aggression, but an indirect link through T1 social rejection. This indirect link was found for the frequency of aggression as well as for reactive and proactive aggression. The fourth article revealed that with regard to relational aggression, a high level of classroom ag-gression predicted an increase of individual aggression only among children with initially low levels of aggression. For physical aggression, it was found that the overall level of aggression in the class affected all children equally. In addition, physical aggression increased the likelihood of social rejection irrespective of the class-level of aggression whereas relational aggression caused social rejection only in classes with a generally low level of relational aggression. The analyses of gender-specific effects showed that children were mainly influenced by their same-gender peers and that the effect on the opposite gender was higher if children engaged in gender-atypical forms of aggressive behavior.
Conclusion: The results provided evidence for the construct and criterion validity of the observational measure of maladaptive anger regulation that was developed within the scope of this research. Furthermore, the findings indicated that maladaptive anger regulation constitutes an important risk factor of aggression through the influence of social rejection. Finally, the results demonstrated that the level of aggression among classmates is relevant for the development of individual aggression over time and that the children´s evaluation of relationally aggressive behavior varies as a function of the normativity of relational aggression in the class. The study findings have implications for the measurement of anger regulation in middle childhood as well as for the prevention of aggression and social rejection.
Molekulare Charakterisierung von CP75, einem neuen centrosomalen Protein in Dictyostelium discoideum
(2016)
Das Centrosom ist ein Zellkern-assoziiertes Organell, das nicht von einer Membran umschlossen ist. Es spielt eine wichtige Rolle in vielen Mikrotubuli- abhängigen Prozessen wie Organellenpositionierung, Zellpolarität oder die Organisation der mitotischen Spindel. Das Centrosom von Dictyostelium besteht aus einer dreischichtigen Core-Struktur umgeben von einer Corona, die Mikrotubuli-nukleierende Komplexe enthält. Die Verdoppelung des Centrosoms in Dictyostelium findet zu Beginn der Mitose statt. In der Prophase vergrößert sich die geschichtete Core-Struktur und die Corona löst sich auf. Anschließend trennen sich die beiden äußeren Lagen der Core-Struktur und bilden in der Metaphase die beiden Spindelpole, die in der Telophase zu zwei vollständigen Centrosomen heranreifen. Das durch eine Proteom-Analyse identifizierte Protein CP75 lokalisiert am Centrosom abhängig von den Mitosephasen. Es dissoziiert von der Core-Struktur in der Prometaphase und erscheint an den Spindelpolen in der Telophase wieder. Dieses Verhalten korreliert mit dem Verhalten der mittleren Lage der Core-Struktur in der Mitose, was darauf hinweist, dass CP75 eine Komponente dieser Schicht sein könnte. Die FRAP-Experimente am Interphase- Centrosom zeigen, dass GFP-CP75 dort nicht mobil ist. Das deutet darauf hin, dass das Protein wichtige Funktionen im Strukturerhalt der centrosomalen Core- Struktur übernehmen könnte. Sowohl die C- als auch die N-terminale Domäne von CP75 enthalten centrosomale Targeting-Domäne. Als GFP-Fusionsproteine (GFP-CP75-N und -C) lokalisieren die beiden Fragmente am Centrosom in der Interphase. Während GFP-CP75-C in der Mitose am Centrosom verbleibt, verschwindet GFP-CP75-N in der Metaphase und kehrt erst in der späten Telophase zurück. GFP-CP75-C und GFP-CP75O/E kolokalisieren mit F-Aktin am Zellcortex, zeigen aber keine Interaktion mit Aktin mit der BioID-Methode. Die N-terminale Domäne von CP75 enthält eine potentielle Plk1- Phosphorylierungssequenz. Die Überexpression der nichtphosphorylierbaren Punktmutante (GFP-CP75-Plk-S143A) ruft verschiedene Phänotypen wie verlängerte oder überzählige Centrosomen, vergrößerte Zellkerne und Anreicherung von detyrosinierten Mikrotubuli hervor. Die ähnlichen Phänotypen konnten auch bei GFP-CP75-N und CP75-RNAi beobachtet werden. Der
Phänotyp der detyrosinierten Mikrotubuli bringt erstmals den Beweis dafür, dass I
in Dictyostelium posttranslationale Modifikation an Tubulinen stattfindet. Außerdem zeigten CP75-RNAi-Zellen Defekte in der Organisation der mitotischen Spindel. Mittels BioID-Methode konnten drei potentielle Interaktionspartner von CP75 identifiziert werden. Diese drei Proteine CP39, CP91 und Cep192 sind ebenfalls Bestandteile des Centrosoms.
This publications-based thesis summarizes my contribution to the scientific field of ultrafast structural dynamics. It consists of 16 publications, about the generation, detection and coupling of coherent gigahertz longitudinal acoustic phonons, also called hypersonic waves. To generate such high frequency phonons, femtosecond near infrared laser pulses were used to heat nanostructures composed of perovskite oxides on an ultrashort timescale. As a consequence the heated regions of such a nanostructure expand and a high frequency acoustic phonon pulse is generated. To detect such coherent acoustic sound pulses I use ultrafast variants of optical Brillouin and x-ray scattering. Here an incident optical or x-ray photon is scattered by the excited sound wave in the sample. The scattered light intensity measures the occupation of the phonon modes.
The central part of this work is the investigation of coherent high amplitude phonon wave packets which can behave nonlinearly, quite similar to shallow water waves which show a steepening of wave fronts or solitons well known as tsunamis. Due to the high amplitude of the acoustic wave packets in the solid, the acoustic properties can change significantly in the vicinity of the sound pulse. This may lead to a shape change of the pulse. I have observed by time-resolved Brillouin scattering, that a single cycle hypersound pulse shows a wavefront steepening. I excited hypersound pulses with strain amplitudes until 1% which I have calibrated by ultrafast x-ray diffraction (UXRD).
On the basis of this first experiment we developed the idea of the nonlinear mixing of narrowband phonon wave packets which we call "nonlinear phononics" in analogy with the nonlinear optics, which summarizes a kaleidoscope of surprising optical phenomena showing up at very high electric fields. Such phenomena are for instance Second Harmonic Generation, four-wave-mixing or solitons. But in case of excited coherent phonons the wave packets have usually very broad spectra which make it nearly impossible to look at elementary scattering processes between phonons with certain momentum and energy.
For that purpose I tested different techniques to excite narrowband phonon wave packets which mainly consist of phonons with a certain momentum and frequency. To this end epitaxially grown metal films on a dielectric substrate were excited with a train of laser pulses. These excitation pulses drive the metal film to oscillate with the frequency given by their inverse temporal displacement and send a hypersonic wave of this frequency into the substrate. The monochromaticity of these wave packets was proven by ultrafast optical Brillouin and x-ray scattering.
Using the excitation of such narrowband phonon wave packets I was able to observe the Second Harmonic Generation (SHG) of coherent phonons as a first example of nonlinear wave mixing of nanometric phonon wave packets.