Refine
Has Fulltext
- yes (2561) (remove)
Year of publication
Document Type
- Doctoral Thesis (2561) (remove)
Language
Keywords
- climate change (54)
- Klimawandel (52)
- Modellierung (34)
- Nanopartikel (28)
- machine learning (22)
- Fernerkundung (20)
- Synchronisation (19)
- remote sensing (18)
- Spracherwerb (17)
- Blickbewegungen (16)
Institute
- Institut für Physik und Astronomie (411)
- Institut für Biochemie und Biologie (391)
- Institut für Geowissenschaften (330)
- Institut für Chemie (304)
- Extern (153)
- Institut für Umweltwissenschaften und Geographie (125)
- Institut für Ernährungswissenschaft (102)
- Wirtschaftswissenschaften (97)
- Hasso-Plattner-Institut für Digital Engineering GmbH (93)
- Department Psychologie (89)
Research on novel and advanced biomaterials is an indispensable step towards their applications in desirable fields such as tissue engineering, regenerative medicine, cell culture, or biotechnology. The work presented here focuses on such a promising material: polyelectrolyte multilayer (PEM) composed of hyaluronic acid (HA) and poly(L-lysine) (PLL). This gel-like polymer surface coating is able to accumulate (bio-)molecules such as proteins or drugs and release them in a controlled manner. It serves as a mimic of the extracellular matrix (ECM) in composition and intrinsic properties. These qualities make the HA/PLL multilayers a promising candidate for multiple bio-applications such as those mentioned above. The work presented aims at the development of a straightforward approach for assessment of multi-fractional diffusion in multilayers (first part) and at control of local molecular transport into or from the multilayers by laser light trigger (second part).
The mechanism of the loading and release is governed by the interaction of bioactives with the multilayer constituents and by the diffusion phenomenon overall. The diffusion of a molecule in HA/PLL multilayers shows multiple fractions of different diffusion rate. Approaches, that are able to assess the mobility of molecules in such a complex system, are limited. This shortcoming motivated the design of a novel evaluation tool presented here.
The tool employs a simulation-based approach for evaluation of the data acquired by fluorescence recovery after photobleaching (FRAP) method. In this approach, possible fluorescence recovery scenarios are primarily simulated and afterwards compared with the data acquired while optimizing parameters of a model until a sufficient match is achieved. Fluorescent latex particles of different sizes and fluorescein in an aqueous medium are utilized as test samples validating the analysis results. The diffusion of protein cytochrome c in HA/PLL multilayers is evaluated as well.
This tool significantly broadens the possibilities of analysis of spatiotemporal FRAP data, which originate from multi-fractional diffusion, while striving to be widely applicable. This tool has the potential to elucidate the mechanisms of molecular transport and empower rational engineering of the drug release systems.
The second part of the work focuses on the fabrication of such a spatiotemporarily-controlled drug release system employing the HA/PLL multilayer. This release system comprises different layers of various functionalities that together form a sandwich structure. The bottom layer, which serves as a reservoir, is formed by HA/PLL PEM deposited on a planar glass substrate. On top of the PEM, a layer of so-called hybrids is deposited. The hybrids consist of thermoresponsive poly(N-isopropylacrylamide) (PNIPAM) -based hydrogel microparticles with surface-attached gold nanorods. The layer of hybrids is intended to serve as a gate that controls the local molecular transport through the PEM–solution-interface. The possibility of stimulating the molecular transport by near-infrared (NIR) laser irradiation is being explored.
From several tested approaches for the deposition of hybrids onto the PEM surface, the drying-based approach was identified as optimal. Experiments, that examine the functionality of the fabricated sandwich at elevated temperature, document the reversible volume phase transition of the PEM-attached hybrids while sustaining the sandwich stability. Further, the gold nanorods were shown to effectively absorb light radiation in the tissue- and cell-friendly NIR spectral region while transducing the energy of light into heat. The rapid and reversible shrinkage of the PEM-attached hybrids was thereby achieved. Finally, dextran was employed as a model transport molecule. It loads into the PEM reservoir in a few seconds with the partition constant of 2.4, while it spontaneously releases in a slower, sustained manner. The local laser irradiation of the sandwich, which contains the fluorescein isothiocyanate tagged dextran, leads to a gradual reduction of fluorescence intensity in the irradiated region.
The release system fabricated employs renowned photoresponsivity of the hybrids in an innovative setting. The results of the research are a step towards a spatially-controlled on-demand drug release system that paves the way to spatiotemporally controlled drug release.
The approaches developed in this work have the potential to elucidate the molecular dynamics in ECM and to foster engineering of multilayers with properties tuned to mimic the ECM. The work aims at spatiotemporal control over the diffusion of bioactives and their presentation to the cells.
While estimated numbers of past and future climate migrants are alarming, the growing empirical evidence suggests that the association between adverse climate-related events and migration is not universally positive. This dissertation seeks to advance our understanding of when and how climate migration emerges by analyzing heterogeneous climatic influences on migration in low- and middle-income countries. To this end, it draws on established economic theories of migration, datasets from physical and social sciences, causal inference techniques and approaches from systematic literature review. In three of its five chapters, I estimate causal effects of processes of climate change on inequality and migration in India and Sub-Saharan Africa. By employing interaction terms and by analyzing sub-samples of data, I explore how these relationships differ for various segments of the population. In the remaining two chapters, I present two systematic literature reviews. First, I undertake a comprehensive meta-regression analysis of the econometric climate migration literature to summarize general climate migration patterns and explain the conflicting findings. Second, motivated by the broad range of approaches in the field, I examine the literature from a methodological perspective to provide best practice guidelines for studying climate migration empirically. Overall, the evidence from this dissertation shows that climatic influences on human migration are highly heterogeneous. Whether adverse climate-related impacts materialize in migration depends on the socio-economic characteristics of the individual households, such as wealth, level of education, agricultural dependence or access to adaptation technologies and insurance. For instance, I show that while adverse climatic shocks are generally associated with an increase in migration in rural India, they reduce migration in the agricultural context of Sub-Saharan Africa, where the average wealth levels are much lower so that households largely cannot afford the upfront costs of moving. I find that unlike local climatic shocks which primarily enhance internal migration to cities and hence accelerate urbanization, shocks transmitted via agricultural producer prices increase migration to neighboring countries, likely due to the simultaneous decrease in real income in nearby urban areas. These findings advance our current understanding by showing when and how economic agents respond to climatic events, thus providing explicit contexts and mechanisms of climate change effects on migration in the future. The resulting collection of findings can guide policy interventions to avoid or mitigate any present and future welfare losses from climate change-related migration choices.
Die vorliegende Arbeit untersucht die Politik der Zentralbankunabhängigkeit (ZBU) am Beispiel der Türkei. Im Mittelpunkt der Arbeit stehen theoretische und empirische Fragen und Probleme, die sich im Zusammenhang mit der ZBU stellen und anhand der türkischen Geldpolitik diskutiert werden. Ein zentrales Ziel der Arbeit besteht darin, zu untersuchen, ob und inwiefern die türkische Zentralbank nach Erlangung der de jure institutionellen Unabhängigkeit tatsächlich als unabhängig und entpolitisiert eingestuft werden kann. Um diese Forschungsfrage zu beantworten, werden die institutionellen Bedingungen, die Ziele und die Regeln, nach denen sich die türkische Geldpolitik richtet, geklärt. Anschließend wird empirisch überprüft, ob die geldpolitische Praxis der CBRT sich an dem offiziell vorgegebenen Regelwerk orientiert. Die Hauptthese dieser Arbeit lautet, dass die formelle Unabhängigkeit der CBRT und die regelorientierte Geldpolitik nicht mit einer Entpolitisierung der Geldpolitik in der Türkei gleichzusetzen ist. Als Alternative schlägt die vorliegende Studie vor, den institutionellen Status der CBRT als einen der relativen Autonomie zu untersuchen. Auch eine de jure unabhängige Zentralbank kann sich nicht von politischen Eingriffen abkoppeln, wie das Fallbeispiel Türkei zeigen wird.
Landslides are frequent natural hazards in rugged terrain, when the resisting frictional force of the surface of rupture yields to the gravitational force. These forces are functions of geological and morphological factors, such as angle of internal friction, local slope gradient or curvature, which remain static over hundreds of years; whereas more dynamic triggering events, such as rainfall and earthquakes, compromise the force balance by temporarily reducing resisting forces or adding transient loads. This thesis investigates landslide distribution and orientation due to landslide triggers (e.g. rainfall) at different scales (6-4∙10^5 km^2) and aims to link rainfall movement with the landslide distribution. It additionally explores the local impacts of the extreme rainstorms on landsliding and the role of precursory stability conditions that could be induced by an earlier trigger, such as an earthquake.
Extreme rainfall is a common landslide trigger. Although several studies assessed rainfall intensity and duration to study the distribution of thus triggered landslides, only a few case studies quantified spatial rainfall patterns (i.e. orographic effect). Quantifying the regional trajectories of extreme rainfall could aid predicting landslide prone regions in Japan. To this end, I combined a non-linear correlation metric, namely event synchronization, and radial statistics to assess the general pattern of extreme rainfall tracks over distances of hundreds of kilometers using satellite based rainfall estimates. Results showed that, although the increase in rainfall intensity and duration positively correlates with landslide occurrence, the trajectories of typhoons and frontal storms were insufficient to explain landslide distribution in Japan. Extreme rainfall trajectories inclined northwestwards and were concentrated along some certain locations, such as coastlines of southern Japan, which was unnoticed in the landslide distribution of about 5000 rainfall-triggered landslides. These landslides seemed to respond to the mean annual rainfall rates.
Above mentioned findings suggest further investigation on a more local scale to better understand the mechanistic response of landscape to extreme rainfall in terms of landslides. On May 2016 intense rainfall struck southern Germany triggering high waters and landslides. The highest damage was reported at the Braunsbach, which is located on the tributary-mouth fan formed by the Orlacher Bach. Orlacher Bach is a ~3 km long creek that drains a catchment of about ~6 km^2. I visited this catchment in June 2016 and mapped 48 landslides along the creek. Such high landslide activity was not reported in the nearby catchments within ~3300 km^2, despite similar rainfall intensity and duration based on weather radar estimates. My hypothesis was that several landslides were triggered by rainfall-triggered flash floods that undercut hillslope toes along the Orlacher Bach. I found that morphometric features such as slope and curvature play an important role in landslide distribution on this micro scale study site (<10 km^2). In addition, the high number of landslides along the Orlacher Bach could also be boosted by accumulated damages on hillslopes due karst weathering over longer time scales.
Precursory damages on hillslopes could also be induced by past triggering events that effect landscape evolution, but this interaction is hard to assess independently from the latest trigger. For example, an earthquake might influence the evolution of a landscape decades long, besides its direct impacts, such as landslides that follow the earthquake. Here I studied the consequences of the 2016 Kumamoto Earthquake (MW 7.1) that triggered some 1500 landslides in an area of ~4000 km^2 in central Kyushu, Japan. Topography, i.e. local slope and curvature, both amplified and attenuated seismic waves, thus controlling the failure mechanism of those landslides (e.g. progressive). I found that topography fails in explaining the distribution and the preferred orientation of the landslides after the earthquake; instead the landslides were concentrated around the northeast of the rupture area and faced mostly normal to the rupture plane. This preferred location of the landslides was dominated mainly by the directivity effect of the strike-slip earthquake, which is the propagation of wave energy along the fault in the rupture direction; whereas amplitude variations of the seismic radiation altered the preferred orientation. I suspect that the earthquake directivity and the asymmetry of seismic radiation damaged hillslopes at those preferred locations increasing landslide susceptibility. Hence a future weak triggering event, e.g. scattered rainfall, could further trigger landslides at those damaged hillslopes.
Mit der Liberalisierung des Strommarkts, den unsicheren Aussichten in der Klimapolitik und stark schwankenden Preisen bei Brennstoffen, Emissionsrechten und Kraftwerkskomponenten hat bei Kraftwerksinvestitionen das Risikomanagement an Bedeutung gewonnen. Dies äußert sich im vermehrten Einsatz probabilistischer Verfahren. Insbesondere bei regulativen Risiken liefert der klassische, häufigkeitsbasierte Wahrscheinlichkeitsbegriff aber keine Handhabe zur Risikoquantifizierung. In dieser Arbeit werden Kraftwerksinvestitionen und -portfolien in Deutschland mit Methoden des Bayes'schen Risikomanagements bewertet. Die Bayes'sche Denkschule begreift Wahrscheinlichkeit als persönliches Maß für Unsicherheit. Wahrscheinlichkeiten können auch ohne statistische Datenanalyse allein mit Expertenbefragungen gewonnen werden. Das Zusammenwirken unsicherer Werttreiber wurde mit einem probabilistischen DCF-Modell (Discounted Cash Flow-Modell) spezifiziert und in ein Einflussdiagramm mit etwa 1200 Objekten umgesetzt. Da der Überwälzungsgrad von Brennstoff- und CO2-Kosten und damit die Höhe der von den Kraftwerken erwirtschafteten Deckungsbeiträge im Wettbewerb bestimmt werden, reicht eine einzelwirtschaftliche Betrachtung der Kraftwerke nicht aus. Strompreise und Auslastungen werden mit Heuristiken anhand der individuellen Position der Kraftwerke in der Merit Order bestimmt, d.h. anhand der nach kurzfristigen Grenzkosten gestaffelten Einsatzreihenfolge. Dazu wurden 113 thermische Großkraftwerke aus Deutschland in einer Merit Order vereinigt. Das Modell liefert Wahrscheinlichkeitsverteilungen für zentrale Größen wie Kapitalwerte von Bestandsportfolien sowie Stromgestehungskosten und Kapitalwerte von Einzelinvestitionen (Steinkohle- und Braunkohlekraftwerke mit und ohne CO2-Abscheidung sowie GuD-Kraftwerke). Der Wert der Bestandsportfolien von RWE, E.ON, EnBW und Vattenfall wird primär durch die Beiträge der Braunkohle- und Atomkraftwerke bestimmt. Erstaunlicherweise schlägt sich der Emissionshandel nicht in Verlusten nieder. Dies liegt einerseits an den Zusatzgewinnen der Atomkraftwerke, andererseits an den bis 2012 gratis zugeteilten Emissionsrechten, welche hohe Windfall-Profite generieren. Dadurch erweist sich der Emissionshandel in seiner konkreten Ausgestaltung insgesamt als gewinnbringendes Geschäft. Über die Restlaufzeit der Bestandskraftwerke resultiert ab 2008 aus der Einführung des Emissionshandels ein Barwertvorteil von insgesamt 8,6 Mrd. €. In ähnlicher Dimension liegen die Barwertvorteile aus der 2009 von der Bundesregierung in Aussicht gestellten Laufzeitverlängerung für Atomkraftwerke. Bei einer achtjährigen Laufzeitverlängerung ergäben sich je nach CO2-Preisniveau Barwertvorteile von 8 bis 15 Mrd. €. Mit höheren CO2-Preisen und Laufzeitverlängerungen von bis zu 28 Jahren würden 25 Mrd. € oder mehr zusätzlich anfallen. Langfristig erscheint fraglich, ob unter dem gegenwärtigen Marktdesign noch Anreize für Investitionen in fossile Kraftwerke gegeben sind. Zu Beginn der NAP 2-Periode noch rentable Investitionen in Braunkohle- und GuD-Kraftwerke werden mit der auslaufenden Gratiszuteilung von Emissionsrechten zunehmend unrentabler. Die Rentabilität wird durch Strommarkteffekte der erneuerbaren Energien und ausscheidender alter Gas- und Ölkraftwerke stetig weiter untergraben. Steinkohlekraftwerke erweisen sich selbst mit anfänglicher Gratiszuteilung als riskante Investition. Die festgestellten Anreizprobleme für Neuinvestitionen sollten jedoch nicht dem Emissionshandel zugeschrieben werden, sondern resultieren aus den an Grenzkosten orientierten Strompreisen. Das Anreizproblem ist allerdings bei moderaten CO2-Preisen am größten. Es gilt auch für Kraftwerke mit CO2-Abscheidung: Obwohl die erwarteten Vermeidungskosten für CCS-Kraftwerke gegenüber konventionellen Kohlekraftwerken im Jahr 2025 auf 25 €/t CO2 (Braunkohle) bzw. 38,5 €/t CO2 (Steinkohle) geschätzt werden, wird ihr Bau erst ab CO2-Preisen von 50 bzw. 77 €/t CO2 rentabel. Ob und welche Kraftwerksinvestitionen sich langfristig rechnen, wird letztlich aber politisch entschieden und ist selbst unter stark idealisierten Bedingungen kaum vorhersagbar.
The Sun is a star, which due to its proximity has a tremendous influence on Earth. Since its very first days mankind tried to "understand the Sun", and especially in the 20th century science has uncovered many of the Sun's secrets by using high resolution observations and describing the Sun by means of models. As an active star the Sun's activity, as expressed in its magnetic cycle, is closely related to the sunspot numbers. Flares play a special role, because they release large energies on very short time scales. They are correlated with enhanced electromagnetic emissions all over the spectrum. Furthermore, flares are sources of energetic particles. Hard X-ray observations (e.g., by NASA's RHESSI spacecraft) reveal that a large fraction of the energy released during a flare is transferred into the kinetic energy of electrons. However the mechanism that accelerates a large number of electrons to high energies (beyond 20 keV) within fractions of a second is not understood yet. The thesis at hand presents a model for the generation of energetic electrons during flares that explains the electron acceleration based on real parameters obtained by real ground and space based observations. According to this model photospheric plasma flows build up electric potentials in the active regions in the photosphere. Usually these electric potentials are associated with electric currents closed within the photosphere. However as a result of magnetic reconnection, a magnetic connection between the regions of different magnetic polarity on the photosphere can establish through the corona. Due to the significantly higher electric conductivity in the corona, the photospheric electric power supply can be closed via the corona. Subsequently a high electric current is formed, which leads to the generation of hard X-ray radiation in the dense chromosphere. The previously described idea is modelled and investigated by means of electric circuits. For this the microscopic plasma parameters, the magnetic field geometry and hard X-ray observations are used to obtain parameters for modelling macroscopic electric components, such as electric resistors, which are connected with each other. This model demonstrates that such a coronal electric current is correlated with large scale electric fields, which can accelerate the electrons quickly up to relativistic energies. The results of these calculations are encouraging. The electron fluxes predicted by the model are in agreement with the electron fluxes deduced from the measured photon fluxes. Additionally the model developed in this thesis proposes a new way to understand the observed double footpoint hard X-ray sources.
Die Elektrosprayionisation (ESI) ist eine der weitverbreitetsten Ionisationstechniken für flüssige Pro-ben in der Massen- und Ionenmobilitäts(IM)-Spektrometrie. Aufgrund ihrer schonenden Ionisierung wird ESI vorwiegend für empfindliche, komplexe Moleküle in der Biologie und Medizin eingesetzt. Überdies ist sie allerdings für ein sehr breites Spektrum an Substanzklassen anwendbar. Die IM-Spektrometrie wurde ursprünglich zur Detektion gasförmiger Proben entwickelt, die hauptsächlich durch radioaktive Quellen ionisiert werden. Sie ist die einzige analytische Methode, bei der Isomere in Echtzeit getrennt und über ihre charakteristische IM direkt identifiziert werden können. ESI wurde in den 90ger Jahren durch die Hill Gruppe in die IM-Spektrometrie eingeführt. Die Kombination wird bisher jedoch nur von wenigen Gruppen verwendet und hat deshalb noch ein hohes Entwick-lungspotential. Ein vielversprechendes Anwendungsfeld ist der Einsatz in der Hochleistungs-flüssigkeitschromatographie (HPLC) zur mehrdimensionalen Trennung. Heutzutage ist die HPLC die Standardmethode zur Trennung komplexer Proben in der Routineanalytik. HPLC-Trennungsgänge sind jedoch häufig langwierig und der Einsatz verschiedener Laufmittel, hoher Flussraten, von Puffern, sowie Laufmittelgradienten stellt hohe Anforderungen an die Detektoren. Die ESI-IM-Spektrometrie wurde in einigen Studien bereits als HPLC-Detektor eingesetzt, war dort bisher jedoch auf Flussratensplitting oder geringe Flussraten des Laufmittels beschränkt.
In dieser kumulativen Doktorarbeit konnte daher erstmals ein ESI IM-Spektrometer als HPLC-Detektor für den Flussratenbereich von 200-1500 μl/min entwickelt werden. Anhand von fünf Publi-kationen wurden (1) über eine umfassende Charakterisierung die Eignung des Spektrometers als HPLC-Detektor festgestellt, (2) ausgewählte komplexe Trenngänge präsentiert und (3) die Anwen-dung zum Reaktionsmonitoring und (4, 5) mögliche Weiterentwicklungen gezeigt.
Erfolgreich konnten mit dem selbst-entwickelten ESI IM-Spektrometer typische HPLC-Bedingungen wie Wassergehalte im Laufmittel von bis zu 90%, Pufferkonzentrationen von bis zu 10 mM, sowie Nachweisgrenzen von bis zu 50 nM erreicht werden. Weiterhin wurde anhand der komplexen Trennungsgänge (24 Pestizide/18 Aminosäuren) gezeigt, dass die HPLC und die IM-Spektrometrie eine hohe Orthogonalität besitzen. Eine effektive Peakkapazität von 240 wurde so realisiert. Auf der HPLC-Säule koeluierende Substanzen konnten über die Driftzeit getrennt und über ihre IM identifi-ziert werden, sodass die Gesamttrennzeiten erheblich minimiert werden konnten. Die Anwend-barkeit des ESI IM-Spektrometers zur Überwachung chemischer Synthesen wurde anhand einer dreistufigen Reaktion demonstriert. Es konnten die wichtigsten Edukte, Zwischenprodukte und Produkte aller Stufen identifiziert werden. Eine quantitative Auswertung war sowohl über eine kurze HPLC-Vortrennung als auch durch die Entwicklung eines eigenen Kalibrierverfahrens, welches die Ladungskonkurrenz bei ESI berücksichtigt, ohne HPLC möglich. Im zweiten Teil der Arbeit werden zwei Weiterentwicklungen des Spektrometers präsentiert. Eine Möglichkeit ist die Reduzierung des Drucks in den intermediären Bereich (300 - 1000 mbar) mit dem Ziel der Verringerung der benötigten Spannungen. Mithilfe von Streulichtbildern und Strom-Spannungs-Kurven wurden für geringe Drücke eine verminderte Freisetzung der Analyt-Ionen aus den Tropfen festgestellt. Die Verluste konnten jedoch über höhere elektrische Feldstärken ausgeglichen werden, sodass gleiche Nachweisgrenzen bei 500 mbar und bei 1 bar erreicht wurden. Die zweite Weiterentwicklung ist ein neuartiges Ionentors mit Pulsschaltung, welches eine Verdopplung der Auflösung auf bis zu R > 100 bei gleicher Sensitivität ermöglichte. Eine denkbare Anwendung im Bereich der Peptidanalytik wurde mit beachtlichen Auflösungen der Peptide von R = 90 gezeigt.
Die vorliegende Arbeit beschäftigt sich mit der Charakterisierung von Seismizität anhand von Erdbebenkatalogen. Es werden neue Verfahren der Datenanalyse entwickelt, die Aufschluss darüber geben sollen, ob der seismischen Dynamik ein stochastischer oder ein deterministischer Prozess zugrunde liegt und was daraus für die Vorhersagbarkeit starker Erdbeben folgt. Es wird gezeigt, dass seismisch aktive Regionen häufig durch nichtlinearen Determinismus gekennzeichent sind. Dies schließt zumindest die Möglichkeit einer Kurzzeitvorhersage ein. Das Auftreten seismischer Ruhe wird häufig als Vorläuferphaenomen für starke Erdbeben gedeutet. Es wird eine neue Methode präsentiert, die eine systematische raumzeitliche Kartierung seismischer Ruhephasen ermöglicht. Die statistische Signifikanz wird mit Hilfe des Konzeptes der Ersatzdaten bestimmt. Als Resultat erhält man deutliche Korrelationen zwischen seismischen Ruheperioden und starken Erdbeben. Gleichwohl ist die Signifikanz dafür nicht hoch genug, um eine Vorhersage im Sinne einer Aussage über den Ort, die Zeit und die Stärke eines zu erwartenden Hauptbebens zu ermöglichen.
Species respond to environmental change by dynamically adjusting their geographical ranges. Robust predictions of these changes are prerequisites to inform dynamic and sustainable conservation strategies. Correlative species distribution models (SDMs) relate species’ occurrence records to prevailing environmental factors to describe the environmental niche. They have been widely applied in global change context as they have comparably low data requirements and allow for rapid assessments of potential future species’ distributions. However, due to their static nature, transient responses to environmental change are essentially ignored in SDMs. Furthermore, neither dispersal nor demographic processes and biotic interactions are explicitly incorporated. Therefore, it has often been suggested to link statistical and mechanistic modelling approaches in order to make more realistic predictions of species’ distributions for scenarios of environmental change. In this thesis, I present two different ways of such linkage. (i) Mechanistic modelling can act as virtual playground for testing statistical models and allows extensive exploration of specific questions. I promote this ‘virtual ecologist’ approach as a powerful evaluation framework for testing sampling protocols, analyses and modelling tools. Also, I employ such an approach to systematically assess the effects of transient dynamics and ecological properties and processes on the prediction accuracy of SDMs for climate change projections. That way, relevant mechanisms are identified that shape the species’ response to altered environmental conditions and which should hence be considered when trying to project species’ distribution through time. (ii) I supplement SDM projections of potential future habitat for black grouse in Switzerland with an individual-based population model. By explicitly considering complex interactions between habitat availability and demographic processes, this allows for a more direct assessment of expected population response to environmental change and associated extinction risks. However, predictions were highly variable across simulations emphasising the need for principal evaluation tools like sensitivity analysis to assess uncertainty and robustness in dynamic range predictions. Furthermore, I identify data coverage of the environmental niche as a likely cause for contrasted range predictions between SDM algorithms. SDMs may fail to make reliable predictions for truncated and edge niches, meaning that portions of the niche are not represented in the data or niche edges coincide with data limits. Overall, my thesis contributes to an improved understanding of uncertainty factors in predictions of range dynamics and presents ways how to deal with these. Finally I provide preliminary guidelines for predictive modelling of dynamic species’ response to environmental change, identify key challenges for future research and discuss emerging developments.
מחקר זה עוסק במשנתו החדשנית של ר' צדוק הכהן מלובלין. לפי רוב החוקרים, ר' צדוק ממשיך את משנת רבו, ר' מרדכי יוסף ליינער מאיזביצה, ומציג פטליזם אקזיסטנציאליסטי: לאדם יש חופש כנגד החוק המקובל, לפי רצון האל הנגלה בליבו, אפילו בתשוקותיו, החורג מההלכה; ואולם, האדם אינו קובע את הרצון המתגלה בשורש נשמתו, אלא חושפו בלבד. בכתבי ר' צדוק ניתן למצוא ביטויים רבים, בתוכן ובצורה, לפטליזם זה; עם זאת, הוא מציג גם התבטאויות רבות בדבר הבחירה האנושית החופשית וכוחה היצירתי לכונן ולקבוע את שורשו של האדם, ולחדש ולהשפיע על העולמות האלוקיים ועל העולם-הזה. מחקר זה מתמקד בקטעים אלו שמרכזיותם הוזנחה במחקר עד כה, ומכאן הבנתו המחודשת במשנתו.
את עמדת ר' צדוק ניתן להסביר באמצעות התפיסה הפרדוקסלית, הטוענת לכוחם המלא של שני הקטבים הסותרים ואף להשפעה הדדית ביניהם, היוצרת מתח קשה אך פורה: בניגוד לתהליך הפורמלי, האניטלקטואלי והאפריורי של הפטליזם, שבו 'הידיעה' האלוהית מבטלת את 'הבחירה' האנושית החופשית, ר' צדוק מציג פטאליות, המזהה המציאות עצמה את הידיעה והרצון האלוהיים המצויים בכל. בפטליות זו הידיעה אינה מבטלת את הבחירה; להיפך: ללא הידיעה האלוקית אין לאף דבר קיום, ולכן רק מציאות הידיעה בבחירה היא המאפשרת את קיומה הממשי.
תפיסתו האונטולוגית של ר' צדוק מתקיימת לא רק בתוכן הישיר של דבריו, אלא גם בעקיפין, באופן דרשנותו ובמובן שהוא מעניק למושגים שבהם הוא דן; לכן, הפטליות מתגלה גם בתחומים נוספים, שבהם מצוי פער בין הממד המוחלט וההכרחי ('ידיעה') לממד הקונטינגנטי, הארעי והיחסי ('בחירה'): השקר, הדמיון, הרוע, החטא, הייסורים ועוד, אמנם קונטינגנטיים ויחסיים לעומת מוחלטות האמת, הטוב וכו'; אך לפי ר' צדוק, האל רוצה בהם ככאלה – כלומר יש להם ממשות שאינה מוחלטת, אלא ממשות ככאלה, כקונטינגנטיים. אך ממשות זו אינה מאשרת אותם כפי שהם, אלא יוצרת בהם-עצמם טרנספורמציה. למשל, הרוע אינו הופך לטוב מוחלט או נותר כרוע, אלא הופך ל'טוב מאד' שלפי ר' צדוק גדול מהטוב הרגיל.
מכאן עולה גם ההשפעה ההפוכה, זו של הבחירה על הידיעה. לפי ר' צדוק, משקבלה הקונטינגנטיות והיחסיות של 'הבחירה' – שהיא למעשה מהות הבריאה ומהות האדם ומעשיו – את ממשותה, בכוחה אף להוסיף כביכול על המוחלטות האלוקית הקבועה של 'הידיעה': 'נגע' החטא או הייסורים הופך בעצמו, בחילוף אותם אותיות, ל'ענג' הגדול מההנאה הרגילה; בכוח האדם להשפיע על העולמות העליונים בגזירת גזירות ובביטול גזירות אלוקיות; בכוחו אף להשפיע על העולם-הזה, בחידושו היומיומי על-ידי חידושי התושב"ע, ובקידוש החודש המסוגל לשנות את תנועת המזלות השמימיים. מצד אחד, היצירתיות האנושית של חידושי התושב"ע כלולה באמת האלוקית הגנוזה בתושב"כ האלוקית והיא רק חושפת אותה; אך לפי ר"צ, מצד אחר מתברר שמקור חידושי התושב"ע דווקא גבוה ממקור התושב"כ והם אלו היוצרים וקובעים אותה.
ר' צדוק מציג שני מושגים מרכזיים לפרדוקס שלו: בממד האונטולוגי – 'הייחוד הגמור' של האל, שבו הדואליות הקונטינגנטית (בין האל לנבראיו, והבחירה שהיא מאפשרת) של 'הייחוד התחתון' מתקיימת פרדוקסלית עם האחדות המוחלטת ו'הידיעה' האלוהית של 'הייחוד העליון'.
בממד של מצבו האקזיסטנציאלי של האדם – 'השורש הנעלם': בניגוד לר' מרדכי יוסף רבו, ר' צדוק טוען שהאדם קובע את גורלו על-ידי מעשיו הבחיריים והקונטינגנטיים; ואולם, במקביל, בדומה לרבו, הוא טוען גם שהשורש הקבוע של האדם נקבע על-ידי 'הידיעה' האלוהית והיא זו הקובעת את מעשיו לטוב או למוטב. אך מעל אלו מציג ר' צדוק שורש נוסף, 'נעלם' הגבוה מהשורש הקבוע: הוא אכן שורש מוחלט ('ידיעה'), אך הוא נקבע ומכוּנן על-ידי מעשיו הבחיריים של האדם, בדומה לבריאה האלוהית 'יש מאין'.
With recent advances in the area of information extraction, automatically extracting structured information from a vast amount of unstructured textual data becomes an important task, which is infeasible for humans to capture all information manually. Named entities (e.g., persons, organizations, and locations), which are crucial components in texts, are usually the subjects of structured information from textual documents. Therefore, the task of named entity mining receives much attention. It consists of three major subtasks, which are named entity recognition, named entity linking, and relation extraction.
These three tasks build up an entire pipeline of a named entity mining system, where each of them has its challenges and can be employed for further applications. As a fundamental task in the natural language processing domain, studies on named entity recognition have a long history, and many existing approaches produce reliable results. The task is aiming to extract mentions of named entities in text and identify their types. Named entity linking recently received much attention with the development of knowledge bases that contain rich information about entities. The goal is to disambiguate mentions of named entities and to link them to the corresponding entries in a knowledge base. Relation extraction, as the final step of named entity mining, is a highly challenging task, which is to extract semantic relations between named entities, e.g., the ownership relation between two companies.
In this thesis, we review the state-of-the-art of named entity mining domain in detail, including valuable features, techniques, evaluation methodologies, and so on. Furthermore, we present two of our approaches that focus on the named entity linking and relation extraction tasks separately.
To solve the named entity linking task, we propose the entity linking technique, BEL, which operates on a textual range of relevant terms and aggregates decisions from an ensemble of simple classifiers. Each of the classifiers operates on a randomly sampled subset of the above range. In extensive experiments on hand-labeled and benchmark datasets, our approach outperformed state-of-the-art entity linking techniques, both in terms of quality and efficiency.
For the task of relation extraction, we focus on extracting a specific group of difficult relation types, business relations between companies. These relations can be used to gain valuable insight into the interactions between companies and perform complex analytics, such as predicting risk or valuating companies. Our semi-supervised strategy can extract business relations between companies based on only a few user-provided seed company pairs. By doing so, we also provide a solution for the problem of determining the direction of asymmetric relations, such as the ownership_of relation. We improve the reliability of the extraction process by using a holistic pattern identification method, which classifies the generated extraction patterns. Our experiments show that we can accurately and reliably extract new entity pairs occurring in the target relation by using as few as five labeled seed pairs.
Throughout the last ~3 million years, the Earth's climate system was characterised by cycles of glacial and interglacial periods. The current warm period, the Holocene, is comparably stable and stands out from this long-term cyclicality. However, since the industrial revolution, the climate has been increasingly affected by a human-induced increase in greenhouse gas concentrations. While instrumental observations are used to describe changes over the past ~200 years, indirect observations via proxy data are the main source of information beyond this instrumental era. These data are indicators of past climatic conditions, stored in palaeoclimate archives around the Earth. The proxy signal is affected by processes independent of the prevailing climatic conditions. In particular, for sedimentary archives such as marine sediments and polar ice sheets, material may be redistributed during or after the initial deposition and subsequent formation of the archive. This leads to noise in the records challenging reliable reconstructions on local or short time scales. This dissertation characterises the initial deposition of the climatic signal and quantifies the resulting archive-internal heterogeneity and its influence on the observed proxy signal to improve the representativity and interpretation of climate reconstructions from marine sediments and ice cores.
To this end, the horizontal and vertical variation in radiocarbon content of a box-core from the South China Sea is investigated. The three-dimensional resolution is used to quantify the true uncertainty in radiocarbon age estimates from planktonic foraminifera with an extensive sampling scheme, including different sample volumes and replicated measurements of batches of small and large numbers of specimen. An assessment on the variability stemming from sediment mixing by benthic organisms reveals strong internal heterogeneity. Hence, sediment mixing leads to substantial time uncertainty of proxy-based reconstructions with error terms two to five times larger than previously assumed.
A second three-dimensional analysis of the upper snowpack provides insights into the heterogeneous signal deposition and imprint in snow and firn. A new study design which combines a structure-from-motion photogrammetry approach with two-dimensional isotopic data is performed at a study site in the accumulation zone of the Greenland Ice Sheet. The photogrammetry method reveals an intermittent character of snowfall, a layer-wise snow deposition with substantial contributions by wind-driven erosion and redistribution to the final spatially variable accumulation and illustrated the evolution of stratigraphic noise at the surface. The isotopic data show the preservation of stratigraphic noise within the upper firn column, leading to a spatially variable climate signal imprint and heterogeneous layer thicknesses. Additional post-depositional modifications due to snow-air exchange are also investigated, but without a conclusive quantification of the contribution to the final isotopic signature.
Finally, this characterisation and quantification of the complex signal formation in marine sediments and polar ice contributes to a better understanding of the signal content in proxy data which is needed to assess the natural climate variability during the Holocene.
In this work, some new results to exploit the recurrence properties of quasiperiodic dynamical systems are presented by means of a two dimensional visualization technique, Recurrence Plots(RPs). Quasiperiodicity is the simplest form of dynamics exhibiting nontrivial recurrences, which are common in many nonlinear systems. The concept of recurrence was introduced to study the restricted three body problem and it is very useful for the characterization of nonlinear systems. I have analyzed in detail the recurrence patterns of systems with quasiperiodic dynamics both analytically and numerically. Based on a theoretical analysis, I have proposed a new procedure to distinguish quasiperiodic dynamics from chaos. This algorithm is particular useful in the analysis of short time series. Furthermore, this approach demonstrates to be efficient in recognizing regular and chaotic trajectories of dynamical systems with mixed phase space. Regarding the application to real situations, I have shown the capability and validity of this method by analyzing time series from fluid experiments.
Lava domes are severely hazardous, mound-shaped extrusions of highly viscous lava and commonly erupt at many active stratovolcanoes around the world. Due to gradual growth and flank oversteepening, such lava domes regularly experience partial or full collapses, resulting in destructive and far-reaching pyroclastic density currents. They are also associated with cyclic explosive activity as the complex interplay of cooling, degassing, and solidification of dome lavas regularly causes gas pressurizations on the dome or the underlying volcano conduit. Lava dome extrusions can last from days to decades, further highlighting the need for accurate and reliable monitoring data.
This thesis aims to improve our understanding of lava dome processes and to contribute to the monitoring and prediction of hazards posed by these domes. The recent rise and sophistication of photogrammetric techniques allows for the extraction of observational data in unprecedented detail and creates ideal tools for accomplishing this purpose. Here, I study natural lava dome extrusions as well as laboratory-based analogue models of lava dome extrusions and employ photogrammetric monitoring by Structure-from-Motion (SfM) and Particle-Image-Velocimetry (PIV) techniques. I primarily use aerial photography data obtained by helicopter, airplanes, Unoccupied Aircraft Systems (UAS) or ground-based timelapse cameras. Firstly, by combining a long time-series of overflight data at Volcán de Colima, México, with seismic and satellite radar data, I construct a detailed timeline of lava dome and crater evolution. Using numerical model, the impact of the extrusion on dome morphology and loading stress is further evaluated and an impact on the growth direction is identified, bearing important implications for the location of collapse hazards. Secondly, sequential overflight surveys at the Santiaguito lava dome, Guatemala, reveal surface motion data in high detail. I quantify the growth of the lava dome and the movement of a lava flow, showing complex motions that occur on different timescales and I provide insight into rock properties relevant for hazard assessment inferred purely by photogrammetric processing of remote sensing data. Lastly, I recreate artificial lava dome and spine growth using analogue modelling under controlled conditions, providing new insights into lava extrusion processes and structures as well as the conditions in which they form.
These findings demonstrate the capabilities of photogrammetric data analyses to successfully monitor lava dome growth and evolution while highlighting the advantages of complementary modelling methods to explain the observed phenomena. The results presented herein further bear important new insights and implications for the hazards posed by lava domes.
This dissertation examines the integration of incongruent visual-scene and morphological-case information (“cues”) in building thematic-role representations of spoken relative clauses in German.
Addressing the mutual influence of visual and linguistic processing, the Coordinated Interplay Account (CIA) describes a mechanism in two steps supporting visuo-linguistic integration (Knoeferle & Crocker, 2006, Cog Sci). However, the outcomes and dynamics of integrating incongruent thematic-role representations from distinct sources have been investigated scarcely. Further, there is evidence that both second-language (L2) and older speakers may rely on non-syntactic cues relatively more than first-language (L1)/young speakers. Yet, the role of visual information for thematic-role comprehension has not been measured in L2 speakers, and only limitedly across the adult lifespan.
Thematically unambiguous canonically ordered (subject-extracted) and noncanonically ordered (object-extracted) spoken relative clauses in German (see 1a-b) were presented in isolation and alongside visual scenes conveying either the same (congruent) or the opposite (incongruent) thematic relations as the sentence did.
1 a Das ist der Koch, der die Braut verfolgt.
This is the.NOM cook who.NOM the.ACC bride follows
This is the cook who is following the bride.
b Das ist der Koch, den die Braut verfolgt.
This is the.NOM cook whom.ACC the.NOM bride follows
This is the cook whom the bride is following.
The relative contribution of each cue to thematic-role representations was assessed with agent identification. Accuracy and latency data were collected post-sentence from a sample of L1 and L2 speakers (Zona & Felser, 2023), and from a sample of L1 speakers from across the adult lifespan (Zona & Reifegerste, under review). In addition, the moment-by-moment dynamics of thematic-role assignment were investigated with mouse tracking in a young L1 sample (Zona, under review).
The following questions were addressed: (1) How do visual scenes influence thematic-role representations of canonical and noncanonical sentences? (2) How does reliance on visual-scene, case, and word-order cues vary in L1 and L2 speakers? (3) How does reliance on visual-scene, case, and word-order cues change across the lifespan?
The results showed reliable effects of incongruence of visually and linguistically conveyed thematic relations on thematic-role representations. Incongruent (vs. congruent) scenes yielded slower and less accurate responses to agent-identification probes presented post-sentence. The recently inspected agent was considered as the most likely agent ~300ms after trial onset, and the convergence of visual scenes and word order enabled comprehenders to assign thematic roles predictively.
L2 (vs. L1) participants relied more on word order overall. In response to noncanonical clauses presented with incongruent visual scenes, sensitivity to case predicted the size of incongruence effects better than L1-L2 grouping. These results suggest that the individual’s ability to exploit specific cues might predict their weighting.
Sensitivity to case was stable throughout the lifespan, while visual effects increased with increasing age and were modulated by individual interference-inhibition levels. Thus, age-related changes in comprehension may stem from stronger reliance on visually (vs. linguistically) conveyed meaning.
These patterns represent evidence for a recent-role preference – i.e., a tendency to re-assign visually conveyed thematic roles to the same referents in temporally coordinated utterances. The findings (i) extend the generalizability of CIA predictions across stimuli, tasks, populations, and measures of interest, (ii) contribute to specifying the outcomes and mechanisms of detecting and indexing incongruent representations within the CIA, and (iii) speak to current efforts to understand the sources of variability in sentence comprehension.
Institutionelle Bildung ist für autistische Lernende mit vielgestaltigen und spezifischen Hindernissen verbunden. Dies gilt insbesondere im Zusammenhang mit Inklusion, deren Relevanz nicht zuletzt durch das Übereinkommen der Vereinten Nationen über die Rechte von Menschen mit Behinderung gegeben ist.
Diese Arbeit diskutiert zahlreiche lernrelevante Besonderheiten im Kontext von Autismus und zeigt Diskrepanzen zu den nicht immer ausreichend angemessenen institutionellen Lehrkonzepten. Eine zentrale These ist hierbei, dass die ungewöhnlich intensive Aufmerksamkeit von Autist*innen für ihre Spezialinteressen dafür genutzt werden kann, das Lernen mit fremdgestellten Inhalten zu erleichtern. Darauf aufbauend werden Lösungsansätze diskutiert, welche in einem neuartigen Konzept für ein digitales mehrgerätebasiertes Lernspiel resultieren.
Eine wesentliche Herausforderung bei der Konzeption spielbasierten Lernens besteht in der adäquaten Einbindung von Lerninhalten in einen fesselnden narrativen Kontext. Am Beispiel von Übungen zur emotionalen Deutung von Mimik, welche für das Lernen von sozioemotionalen Kompetenzen besonders im Rahmen von Therapiekonzepten bei Autismus Verwendung finden, wird eine angemessene Narration vorgestellt, welche die störungsarme Einbindung dieser sehr speziellen Lerninhalte ermöglicht.
Die Effekte der einzelnen Konzeptionselemente werden anhand eines prototypisch entwickelten Lernspiels untersucht. Darauf aufbauend zeigt eine quantitative Studie die gute Akzeptanz und Nutzerfreundlichkeit des Spiels und belegte vor allem die
Verständlichkeit der Narration und der Spielelemente. Ein weiterer Schwerpunkt liegt in der minimalinvasiven Untersuchung möglicher Störungen des Spielerlebnisses durch den Wechsel zwischen verschiedenen Endgeräten, für die ein innovatives Messverfahren entwickelt wurde.
Im Ergebnis beleuchtet diese Arbeit die Bedeutung und die Grenzen von spielbasierten Ansätzen für autistische Lernende. Ein großer Teil der vorgestellten Konzepte lässt sich auf andersartige Lernszenarien übertragen. Das dafür entwickelte technische Framework zur Realisierung narrativer Lernpfade ist ebenfalls darauf vorbereitet, für weitere Lernszenarien, gerade auch im institutionellen Kontext, Verwendung zu finden.
This research addressed the question, if it is possible to simplify current microcontact printing systems for the production of anisotropic building blocks or patchy particles, by using common chemicals while still maintaining reproducibility, high precision and tunability of the Janus-balance
Chapter 2 introduced the microcontact printing materials as well as their defined electrostatic interactions. In particular polydimethylsiloxane stamps, silica particles and high molecular weight polyethylenimine ink were mainly used in this research. All of these components are commercially available in large quantities and affordable, which gives this approach a huge potential for further up-scaling developments. The benefits of polymeric over molecular inks was described including its flexible influence on the printing pressure. With this alteration of the µCP concept, a new method of solvent assisted particle release mechanism enabled the switch from two-dimensional surface modification to three-dimensional structure printing on colloidal silica particles, without changing printing parameters or starting materials. This effect opened the way to use the internal volume of the achieved patches for incorporation of nano additives, introducing additional physical properties into the patches without alteration of the surface chemistry.
The success of this system and its achievable range was further investigated in chapter 3 by giving detailed information about patch geometry parameters including diameter, thickness and yield. For this purpose, silica particles in a size range between 1µm and 5µm were printed with different ink concentrations to change the Janus-balance of these single patched particles. A necessary intermediate step, consisting of air-plasma treatment, for the production of trivalent particles using "sandwich" printing was discovered and comparative studies concerning the patch geometry of single and double patched particles were conducted. Additionally, the usage of structured PDMS stamps during printing was described. These results demonstrate the excellent precision of this approach and opens the pathway for even greater accuracy as further parameters can be finely tuned and investigated, e.g. humidity and temperature during stamp loading.
The performance of these synthesized anisotropic colloids was further investigated in chapter 4, starting with behaviour studies in alcoholic and aqueous dispersions. Here, the stability of the applied patches was studied in a broad pH range, discovering a release mechanism by disabling the electrostatic bonding between particle surface and polyelectrolyte ink. Furthermore, the absence of strong attractive forces between divalent particles in water was investigated using XPS measurements. These results lead to the conclusion that the transfer of small PDMS oligomers onto the patch surface is shielding charges, preventing colloidal agglomeration. However, based on this knowledge, further patch modifications for particle self-assembly were introduced including physical approaches using magnetic nano additives, chemical patch functionalization with avidin-biotin or the light responsive cyclodextrin-arylazopyrazoles coupling as well as particle surface modification for the synthesis of highly amphiphilic colloids. The successful coupling, its efficiency, stability and behaviour in different solvents were evaluated to find a suitable coupling system for future assembly experiments. Based on these results the possibility of more sophisticated structures by colloidal self-assembly is given.
Certain findings needed further analysis to understand their underlying mechanics, including the relatively broad patch diameter distribution and the decreasing patch thickness for smaller silica particles. Mathematical assumptions for both effects are introduced in chapter 5. First, they demonstrate the connection between the naturally occurring particle size distribution and the broadening of the patch diameter, indicating an even higher precision for this µCP approach. Second, explaining the increase of contact area between particle and ink surface due to higher particle packaging, leading to a decrease in printing pressure for smaller particles.
These calculations ultimately lead to the development of a new mechanical microcontact printing approach, using centrifugal forces for high pressure control and excellent parallel alignment of printing substrates. First results with this device and the comparison with previously conducted by-hand experiments conclude this research. It furthermore displays the advantages of such a device for future applications using a mechanical printing approach, especially for accessing even smaller nano particles with great precision and excellent yield.
In conclusion, this work demonstrates the successful adjustment of the µCP approach using commercially available and affordable silica particles and polyelectrolytes for high flexibility, reduced costs and higher scale-up value. Furthermore, its was possible to increase the modification potential by introducing three-dimensional patches for additional functionalization volume. While keeping a high colloidal stability, different coupling systems showed the self-assembly capabilities of this toolbox for anisotropic particles.
As land-cover conversion continues to expand into ever more remote areas in the humid tropics, montane rainforests are increasingly threatened. In the south Ecuadorian Andes, they are not only subject to man-made disturbances but also to naturally occurring landslides. I was interested in the impact of this ecosystem dynamics on a key parameter of the hydrologic cycle, the soil saturated hydraulic conductivity (synonym: permeability; Ks from here on), because it is a sensitive indicator for soil disturbances. My general objective was to quantify the effects of the regional natural and human disturbances on the saturated hydraulic conductivity and to describe the resulting spatial-temporal patterns. The main hypotheses were: 1) disturbances cause an apparent displacement of the less permeable soil layer towards the surface, either due to a loss of the permeable surface soil after land-sliding, or as a consequence of the surface soil compaction under cattle pastures; 2) ‘recovery’ from disturbance, either because of landslide re-vegetation or because of secondary succession after pasture abandonment, involves an apparent displacement of the less permeable layer back towards the original depth an 3) disturbances cause a simplification of the Ks spatial structure, i.e. the spatially dependent random variation diminishes; the subsequent recovery entails the re-establishment of the original structure. In my first study, I developed a synthesis of recent geostatistical research regarding its applicability to soil hydraulic data, including exploratory data analysis and variogram estimation techniques; I subsequently evaluated the results in terms of spatial prediction uncertainty. Concerning the exploratory data analysis, my main results were: 1) Gaussian uni- and bivariate distributions of the log-transformed data; 2) the existence of significant local trends; 3) no need for robust estimation; 4) no anisotropic variation. I found partly considerable differences in covariance parameters resulting from different variogram estimation techniques, which, in the framework of spatial prediction, were mainly reflected in the spatial connectivity of the Ks-field. Ignoring the trend component and an arbitrary use of robust estimators, however, would have the most severe consequences in this respect. Regarding variogram modeling, I encouraged restricted maximum likelihood estimation because of its accuracy and independence on the selected lags needed for experimental variograms. The second study dealt with the Ks spatial-temporal pattern in the sequences of natural and man-made disturbances characteristic for the montane rainforest study area. To investigate the disturbance effects both on global means and the spatial structure of Ks, a combined design-and model-based sampling approach was used for field-measurements at soil depths of 12.5, 20, and 50 cm (n=30-150/depth) under landslides of different ages (2 and 8 years), under actively grazed pasture, fallows following pasture abandonment (2 to 25 years of age), and under natural forest. Concerning global means, our main findings were 1) global means of the soil permeability generally decrease with increasing soil depth; 2) no significant Ks differences can be observed among landslides and compared to the natural forest; 3) a distinct permeability decrease of two orders of magnitude occurs after forest conversion to pasture at shallow soil depths, and 4) the slow regeneration process after pasture abandonment requires at least one decade. Regarding the Ks spatial structure, we found that 1) disturbances affect the Ks spatial structure in the topsoil, and 2) the largest differences in spatial patterns are associated with the subsoil permeability. In summary, the regional landslide activity seems to affect soil hydrology to a marginal extend only, which is in contrast to the pronounced drop of Ks after forest conversion. We used this spatial-temporal information combined with local rain intensities to assess the partitioning of rainfall into vertical and lateral flowpaths under undisturbed, disturbed, and regenerating land-cover types in the third study. It turned out that 1) the montane rainforest is characterized by prevailing vertical flowpaths in the topsoil, which can switch to lateral directions below 20 cm depth for a small number of rain events, which may, however, transport a high portion of the annual runoff; 2) similar hydrological flowpaths occur under the landslides except for a somewhat higher probability of impermeable layer formation in the topsoil of a young landslide, and 3) pronounced differences in runoff components can be observed for the human disturbance sequence involving the development of near-surface impeding layers for 24, 44, and 8 % of rain events for pasture, a two-year-old fallow, and a ten-year-old fallow, respectively.
Motivations and research objectives: During the passage of rain water through a forest canopy two main processes take place. First, water is redistributed; and second, its chemical properties change substantially. The rain water redistribution and the brief contact with plant surfaces results in a large variability of both throughfall and its chemical composition. Since throughfall and its chemistry influence a range of physical, chemical and biological processes at or below the forest floor the understanding of throughfall variability and the prediction of throughfall patterns potentially improves the understanding of near-surface processes in forest ecosystems. This thesis comprises three main research objectives. The first objective is to determine the variability of throughfall and its chemistry, and to investigate some of the controlling factors. Second, I explored throughfall spatial patterns. Finally, I attempted to assess the temporal persistence of throughfall and its chemical composition. Research sites and methods: The thesis is based on investigations in a tropical montane rain forest in Ecuador, and lowland rain forest ecosystems in Brazil and Panama. The first two studies investigate both throughfall and throughfall chemistry following a deterministic approach. The third study investigates throughfall patterns with geostatistical methods, and hence, relies on a stochastic approach. Results and Conclusions: Throughfall is highly variable. The variability of throughfall in tropical forests seems to exceed that of many temperate forests. These differences, however, do not solely reflect ecosystem-inherent characteristics, more likely they also mirror management practices. Apart from biotic factors that influence throughfall variability, rainfall magnitude is an important control. Throughfall solute concentrations and solute deposition are even more variable than throughfall. In contrast to throughfall volumes, the variability of solute deposition shows no clear differences between tropical and temperate forests, hence, biodiversity is not a strong predictor of solute deposition heterogeneity. Many other factors control solute deposition patterns, for instance, solute concentration in rainfall and antecedent dry period. The temporal variability of the latter factors partly accounts for the low temporal persistence of solute deposition. In contrast, measurements of throughfall volume are quite stable over time. Results from the Panamanian research site indicate that wet and dry areas outlast consecutive wet seasons. At this research site, throughfall exhibited only weak or pure nugget autocorrelation structures over the studies lag distances. A close look at the geostatistical tools at hand provided evidence that throughfall datasets, in particular those of large events, require robust variogram estimation if one wants to avoid outlier removal. This finding is important because all geostatistical throughfall studies that have been published so far analyzed their data using the classical, non-robust variogram estimator.
Die vorliegende Untersuchung verfolgt das Ziel, kulturelle und religiöse Aspekte der Erneuerung jüdischen Lebens in Berlin seit 1989 zu erforschen. Die Entwicklungen der jüdischen Gemeinschaft in der Hauptstadt seit dem Fall der Mauer und dem Zusammenbruch der Sowjetunion führen zur Wiederannäherung eines Teils der jüdischen Bevölkerung in Deutschland an die eigene Kultur, Religion und Geschichte. Dabei kommt die Pluralität der kulturellen, literarischen und religiösen Ausdrucksformen der jüdischen Identitäten zum Vorschein. Die Arbeit verdeutlicht diese in Berlin nach 1989 einsetzende kulturelle und religiöse „Renaissance“. Vier wichtige Punkte kennzeichnen das jüdische Leben in Berlin nach 1989. Erstens gewinnt Deutschland seit der Wiedervereinigung eine neue Rolle als mögliches Einwanderungsland für Juden. Vor allem mit der massiven jüdischen Einwanderung aus den Staaten der ehemaligen Sowjetunion seit den 1990er Jahren wird Deutschland allmählich als wichtiges Zentrum in der europäischen Diaspora anerkannt. Zweitens bleibt zwar die Shoah tief verankert im Gedächtnis der jüdischen Gemeinschaft; die meisten Kinder oder Enkelkinder von Überlebenden der Shoah weigern sich jedoch, ihre jüdische Identität exklusiv durch die Shoah zu definieren. Sie gründen zur Wiederentdeckung und Forderung ihres kulturellen, religiösen und historischen Erbes jüdische Gruppen und Einrichtungen in Berlin, die in den meisten Fällen alternativ zur Jüdischen Gemeinde entstehen: Künstlergruppen, jüdische Kulturvereine, Konferenzen und Podiumsdiskussionen, religiöse Kongregationen und Lernhäuser. Damit – und dies ist der dritte Punkt – verliert zwar die offizielle Jüdische Gemeinde an Bedeutung als einzige Vertreterin der jüdischen Gemeinschaft Berlins; diese kulturelle und religiöse „Renaissance“ außerhalb der offiziellen Strukturen der Gemeinde bedeutet aber auch eine wachsende Pluralität und Diversifizierung der jüdischen Gemeinschaft in Berlin. Viertens spielt Berlin die Hauptrolle in diesem Prozess. Heute werden viele ehemalige jüdische Orte neu belebt: Synagogen werden wiederentdeckt und renoviert, Denk- und Mahnmale gebaut, Stadtführungen auf der Spur des „jüdischen Berlins“ organisiert, Rabbinerseminare neu gegründet. Die Topographie Berlins bildet auch eine Inspirationsquelle für jüdische (und nichtjüdische) Schriftsteller und Künstler. Die Analyse dieser nach 1989 entstandenen religiösen Initiativen, literarischen Werke und kulturellen Produktionen dient dazu, Aspekte der kulturellen und religiösen „Renaissance“ in Berlin näher zu verdeutlichen.
This work incorporates three treatises which are commonly concerned with a stochastic theory of the Lyapunov exponents. With the help of this theory universal scaling laws are investigated which appear in coupled chaotic and disordered systems. First, two continuous-time stochastic models for weakly coupled chaotic systems are introduced to study the scaling of the Lyapunov exponents with the coupling strength (coupling sensitivity of chaos). By means of the the Fokker-Planck formalism scaling relations are derived, which are confirmed by results of numerical simulations. Next, coupling sensitivity is shown to exist for coupled disordered chains, where it appears as a singular increase of the localization length. Numerical findings for coupled Anderson models are confirmed by analytic results for coupled continuous-space Schrödinger equations. The resulting scaling relation of the localization length resembles the scaling of the Lyapunov exponent of coupled chaotic systems. Finally, the statistics of the exponential growth rate of the linear oscillator with parametric noise are studied. It is shown that the distribution of the finite-time Lyapunov exponent deviates from a Gaussian one. By means of the generalized Lyapunov exponents the parameter range is determined where the non-Gaussian part of the distribution is significant and multiscaling becomes essential.
Geometric electroelasticity
(2014)
In this work a diffential geometric formulation of the theory of electroelasticity is developed which also includes thermal and magnetic influences. We study the motion of bodies consisting of an elastic material that are deformed by the influence of mechanical forces, heat and an external electromagnetic field. To this end physical balance laws (conservation of mass, balance of momentum, angular momentum and energy) are established. These provide an equation that describes the motion of the body during the deformation. Here the body and the surrounding space are modeled as Riemannian manifolds, and we allow that the body has a lower dimension than the surrounding space. In this way one is not (as usual) restricted to the description of the deformation of three-dimensional bodies in a three-dimensional space, but one can also describe the deformation of membranes and the deformation in a curved space. Moreover, we formulate so-called constitutive relations that encode the properties of the used material. Balance of energy as a scalar law can easily be formulated on a Riemannian manifold. The remaining balance laws are then obtained by demanding that balance of energy is invariant under the action of arbitrary diffeomorphisms on the surrounding space. This generalizes a result by Marsden and Hughes that pertains to bodies that have the same dimension as the surrounding space and does not allow the presence of electromagnetic fields. Usually, in works on electroelasticity the entropy inequality is used to decide which otherwise allowed deformations are physically admissible and which are not. It is alsoemployed to derive restrictions to the possible forms of constitutive relations describing the material. Unfortunately, the opinions on the physically correct statement of the entropy inequality diverge when electromagnetic fields are present. Moreover, it is unclear how to formulate the entropy inequality in the case of a membrane that is subjected to an electromagnetic field. Thus, we show that one can replace the use of the entropy inequality by the demand that for a given process balance of energy is invariant under the action of arbitrary diffeomorphisms on the surrounding space and under linear rescalings of the temperature. On the one hand, this demand also yields the desired restrictions to the form of the constitutive relations. On the other hand, it needs much weaker assumptions than the arguments in physics literature that are employing the entropy inequality. Again, our result generalizes a theorem of Marsden and Hughes. This time, our result is, like theirs, only valid for bodies that have the same dimension as the surrounding space.
Die gewaltigen Strukturveränderungen im Bereich des Gesundheitswesens, die in den letzten Jahren bereits erfolgten und die, die noch bevorstehen, zwingen Unternehmen, mit geplanten und gesteuerten Veränderungsprozessen die Voraussetzungen für eine kontinuierliche Anpassung an die neuen Gegebenheiten zu schaffen und somit ihre Zukunftsfähigkeit sicherzustellen. Vor diesem Hintergrund wird gezeigt, wie das Excellence-Modell der EFQM als Instrument für Veränderungsprozesse eingesetzt werden kann und es geeignet ist, Veränderungsziele zu definieren und die Zielerreichung zu bewerten. Referenzobjekt der Fallstudienanalyse, die einen Zeitraum von 13 Jahren umfasst, ist der Medizinische Dienst der Krankenversicherung Rheinland-Pfalz. Neben der Analyse und Darstellung von theoretischen Grundlagen wird an einem Unternehmen des Gesundheitswesens beispielhaft gezeigt, wie die Umsetzung in der Praxis unter Einsatz des EFQM-Modells erfolgen kann. Da das EFQM-Modell mit seiner Systematik unternehmensunabhängig eingesetzt werden kann, sind Lern- und Übertragungsmöglichkeiten strukturiert möglich. Es wird der Nachweis erbracht, dass sich das EFQM-Modell im Rahmen eines Management- und Qualitätssteue-rungsmodells als Universalmethode nutzen lässt, sofern das Management über die Kompe-tenz branchenspezifischer Anpassung verfügt. Auf dem Weg zu organisatorischer Excel-lence wird gezielt an Planungs- und Prognosetechniken des strategischen Managements (SWOT, Szenario-Analyse, Portfolio-Analyse) angeknüpft und auf das VRIO-Paradigma des Resource-based View Bezug genommen. Das EFQM-Modell wird dem Stresstest des ressourcenstrategischen Ansatzes unterzogen, um so zu zeigen, dass es sich beim EFQM-Modell um ein einzigartiges, schwer imitierbares, kaum zu substituierendes, organisatorisch verankertes und kundennutzen-stiftendes Er-folgspotenzial handeln kann. Die Arbeit liefert Impulse und konkrete Anregungen, die zu einem hohen managementprakti-schen Erkenntniszuwachs für den erfolgreichen Umgang mit dem EFQM-Modell und dem Einsatz von Qualitätsmanagementsystemen führen können.
Hauptanliegen der Dissertation ist es, einen Entwurf einer praktischen Ästhetik zu lancieren, der an der Schnittstelle zwischen philosophischer Ästhetik und Kunst – genauer Performancekunst - im Zeichen der Bezugsgrösse der Verletzbarkeit steht. In jüngeren Ästhetikansätzen hat sich eine Auffassung herauskristallisiert, die nicht über, sondern mit Kunst reflektiert. Die Pointe im ‚Mit’ liegt darin, dass diese Ästhetiken die Kunst nicht erklären, sie bestimmen und damit ihre Bedeutung festlegen, sondern dass diese entlang der Kunst die Brüche, Widerstände und Zäsuren zwischen Wahrnehmen und Denken markieren und diese als produktiv bewerten. Diese Lesart etabliert ein Denken, das nicht aus der Distanz auf etwas schaut (theoria), sondern ästhetisch-reflektierend (zurückwendend, auch selbstkritisch) mit der Kunst denkt. Die Disziplin der Ästhetik - als aisthesis: Lehre der sinnlichen Wahrnehmung - nimmt innerhalb der Philosophie eine besondere Stellung ein, weil sie auf ebendiese Differenz verweist und deshalb sinnliche und nicht nur logisch-argumentatorische Denkfiguren stärkt. Als eine Möglichkeit, die Kluft, das Nicht-Einholbare, die brüchige Unzulänglichkeit des begrifflich Denkenden gegenüber ästhetischer Erfahrung zu stärken, schlage ich die Bezugsgrösse der Verletzbarkeit vor. Eine solche Ästhetik besteht aus dem Kreieren verletzbarer Orte, wobei diese auf zweierlei Weisen umkreist werden: Zum einen aus der Kunstpraxis heraus anhand der ästhetischen Figur des verletzbaren Körpes, wie er sich in der zeitgenössischen Performance zeigt. Zum anderen als ein Kreieren von Begriffen im Bewusstsein ihrer Verletzbarkeit. Ausgangspunkte sind die Denkentwürfe von Gilles Deleuze und Hans Blumenberg: Die Ästhetik von Gilles Deleuze entwirft eine konkrete Überschneidungsmöglichkeit von Kunst und Philosophie, aus der sich meine These des Mit-Kunst-Denkens entwickeln lässt. Sie kann aus der Grundvoraussetzung des Deleuzeschen Denkens heraus begründet werden, die besagt, dass nicht nur die Kunst, sondern auch die Philosophie eine schöpferische Tätigkeit ist. Beide Disziplinen beruhen auf dem Prinzip der creatio continua, durch welche die Kunst Empfindungen und die Philosophie Begriffe schöpft, wobei eben genau dieser schöpferische Prozess Kunst und Philosophie in ein produktives Verhältnis zueinander treten lässt. Wie Deleuze seine Begriffsarbeit entlang künstlerischer Praxis entwickelt, wird anhand der Analyse des bis heute wenig rezipierten Textes Ein Manifest weniger in Bezug auf das Theater von Carmelo Bene analysiert. Eine ganz anderen Zugang zum Entwurf einer praktischen Ästhetik liefert Hans Blumenberg, der eine Theorie der Unbegrifflichkeit in Aussicht stellt. Im Anschluss an seine Forderung, die Metapher wieder vermehrt in die philosophische Denkpraxis zu integrieren, radikalisiert er seine Forderung, auch das Nichtanschauliche zu berücksichtigen, indem er das gänzlich Unbegriffliche an die Seite des Begrifflichen stellt. Definitorische Schwäche zeigt sich als wahrhaftige Stärke, die in der Unbegrifflichkeit ihren Zenit erreicht. Der Schiffbruch wird von mir als zentrale Metapher – gewissermassen als Metapher der Metapher – verstanden, die das Auf-Grund-Laufen des Allwissenden veranschaulicht. Im Schiffbruch wird die produktive Kollision von Theorie und Praxis deutlich. Deleuze und Blumenberg zeigen über ‚creatio continua’ und ‚Unbegrifflichkeit’ die Grenzen des Begreifens, indem sie betonen, dass sich Ästhetik nicht nur auf künstlerische Erfahrungen bezieht, sondern selber in das Gegenwärtigmachen von Erfahrungen involviert ist. Daraus folgt, dass ästhetische Reflexion nicht nur begrifflich agieren muss. Die praktische Ästhetik animiert dazu, andere darstellerische Formen (Bilder, Töne, Körper) als differente und ebenbürtige reflexive Modi anzuerkennen und sie als verletzbarmachende Formate der Sprache an die Seite zu stellen. Diese Lesart betont den gestalterischen Aspekt der Ästhetik selber. Zur Verdeutlichung dieser Kluft zwischen (Körper-)Bild und Begriff ist der von mir mitgestaltete Film Augen blickeN der Dissertation als Kapitel beigefügt. Dieser Film zeigt Performer und Performerinnen, die sich bewusst entschieden haben, ihren ‚abweichenden’ Körper auf der Bühne zu präsentieren. Das Wort Verletzbarkeit verweist auf die paradoxe Situation, etwas Brüchiges tragfähig zu machen und dadurch auch auf eine besondere Beziehungsform und auf ein existenzielles Aufeinander-Verwiesensein der Menschen. Verletzbarkeit geht alle an, und stiftet deshalb eine Gemeinsamkeit besonderer Art. In diesem Sinne sind verletzbare Orte nicht nur ästhetische, sondern auch ethische Orte, womit die politische Dimension des Vorhabens betont wird.
Distances affect economic decision-making in numerous situations. The time at which we make a decision about future consumption has an impact on our consumption behavior. The spatial distance to employer, school or university impacts the place where we live and vice versa. The emotional closeness to other individuals influences our willingness to give money to them. This cumulative thesis aims to enrich the literature on the role of distance for economic decision-making. Thereby, each of my research projects sheds light on the impact of one kind of distance for efficient decision-making.
This thesis is concerned with the solution of the blind source separation problem (BSS). The BSS problem occurs frequently in various scientific and technical applications. In essence, it consists in separating meaningful underlying components out of a mixture of a multitude of superimposed signals. In the recent research literature there are two related approaches to the BSS problem: The first is known as Independent Component Analysis (ICA), where the goal is to transform the data such that the components become as independent as possible. The second is based on the notion of diagonality of certain characteristic matrices derived from the data. Here the goal is to transform the matrices such that they become as diagonal as possible. In this thesis we study the latter method of approximate joint diagonalization (AJD) to achieve a solution of the BSS problem. After an introduction to the general setting, the thesis provides an overview on particular choices for the set of target matrices that can be used for BSS by joint diagonalization. As the main contribution of the thesis, new algorithms for approximate joint diagonalization of several matrices with non-orthogonal transformations are developed. These newly developed algorithms will be tested on synthetic benchmark datasets and compared to other previous diagonalization algorithms. Applications of the BSS methods to biomedical signal processing are discussed and exemplified with real-life data sets of multi-channel biomagnetic recordings.
Information on the contemporary in-situ stress state of the earth’s crust is essential for geotechnical applications and physics-based seismic hazard assessment. Yet, stress data records for a data point are incomplete and their availability is usually not dense enough to allow conclusive statements. This demands a thorough examination of the in-situ stress field which is achieved by 3D geomechanicalnumerical models. However, the models spatial resolution is limited and the resulting local stress state is subject to large uncertainties that confine the significance of the findings. In addition, temporal variations of the in-situ stress field are naturally or anthropogenically induced. In my thesis I address these challenges in three manuscripts that investigate (1) the current crustal stress field orientation, (2) the 3D geomechanical-numerical modelling of the in-situ stress state, and (3) the phenomenon of injection induced temporal stress tensor rotations. In the first manuscript I present the first comprehensive stress data compilation of Iceland with 495 data records. Therefore, I analysed image logs from 57 boreholes in Iceland for indicators of the orientation of the maximum horizontal stress component. The study is the first stress survey from different kinds of stress indicators in a geologically very young and tectonically active area of an onshore spreading ridge. It reveals a distinct stress field with a depth independent stress orientation even very close to the spreading centre. In the second manuscript I present a calibrated 3D geomechanical-numerical modelling approach of the in-situ stress state of the Bavarian Molasse Basin that investigates the regional (70x70x10km³) and local (10x10x10km³) stress state. To link these two models I develop a multi-stage modelling approach that provides a reliable and efficient method to derive from the larger scale model initial and boundary conditions for the smaller scale model. Furthermore, I quantify the uncertainties in the models results which are inherent to geomechanical-numerical modelling in general and the multi-stage approach in particular. I show that the significance of the models results is mainly reduced due to the uncertainties in the material properties and the low number of available stress magnitude data records for calibration. In the third manuscript I investigate the phenomenon of injection induced temporal stress tensor rotation and its controlling factors. I conduct a sensitivity study with a 3D generic thermo-hydro-mechanical model. I show that the key control factors for the stress tensor rotation are the permeability as the decisive factor, the injection rate, and the initial differential stress. In particular for enhanced geothermal systems with a low permeability large rotations of the stress tensor are indicated. According to these findings the estimation of the initial differential stress in a reservoir is possible provided the permeability is known and the angle of stress rotation is observed. I propose that the stress tensor rotations can be a key factor in terms of the potential for induced seismicity on pre-existing faults due to the reorientation of the stress field that changes the optimal orientation of faults.
Self-adaptive data quality
(2017)
Carrying out business processes successfully is closely linked to the quality of the data inventory in an organization. Lacks in data quality lead to problems: Incorrect address data prevents (timely) shipments to customers. Erroneous orders lead to returns and thus to unnecessary effort. Wrong pricing forces companies to miss out on revenues or to impair customer satisfaction. If orders or customer records cannot be retrieved, complaint management takes longer. Due to erroneous inventories, too few or too much supplies might be reordered.
A special problem with data quality and the reason for many of the issues mentioned above are duplicates in databases. Duplicates are different representations of same real-world objects in a dataset. However, these representations differ from each other and are for that reason hard to match by a computer. Moreover, the number of required comparisons to find those duplicates grows with the square of the dataset size. To cleanse the data, these duplicates must be detected and removed. Duplicate detection is a very laborious process. To achieve satisfactory results, appropriate software must be created and configured (similarity measures, partitioning keys, thresholds, etc.). Both requires much manual effort and experience.
This thesis addresses automation of parameter selection for duplicate detection and presents several novel approaches that eliminate the need for human experience in parts of the duplicate detection process.
A pre-processing step is introduced that analyzes the datasets in question and classifies their attributes semantically. Not only do these annotations help understanding the respective datasets, but they also facilitate subsequent steps, for example, by selecting appropriate similarity measures or normalizing the data upfront. This approach works without schema information.
Following that, we show a partitioning technique that strongly reduces the number of pair comparisons for the duplicate detection process. The approach automatically finds particularly suitable partitioning keys that simultaneously allow for effective and efficient duplicate retrieval. By means of a user study, we demonstrate that this technique finds partitioning keys that outperform expert suggestions and additionally does not need manual configuration. Furthermore, this approach can be applied independently of the attribute types.
To measure the success of a duplicate detection process and to execute the described partitioning approach, a gold standard is required that provides information about the actual duplicates in a training dataset. This thesis presents a technique that uses existing duplicate detection results and crowdsourcing to create a near gold standard that can be used for the purposes above. Another part of the thesis describes and evaluates strategies how to reduce these crowdsourcing costs and to achieve a consensus with less effort.
Biofilms are complex living materials that form as bacteria get embedded in a matrix of self-produced protein and polysaccharide fibres. The formation of a network of extracellular biopolymer fibres contributes to the cohesion of the biofilm by promoting cell-cell attachment and by mediating biofilm-substrate interactions. This sessile mode of bacteria growth has been well studied by microbiologists to prevent the detrimental effects of biofilms in medical and industrial settings. Indeed, biofilms are associated with increased antibiotic resistance in bacterial infections, and they can also cause clogging of pipelines or promote bio-corrosion. However, biofilms also gained interest from biophysics due to their ability to form complex morphological patterns during growth. Recently, the emerging field of engineered living materials investigates biofilm mechanical properties at multiple length scales and leverages the tools of synthetic biology to tune the functions of their constitutive biopolymers.
This doctoral thesis aims at clarifying how the morphogenesis of Escherichia coli (E. coli) biofilms is influenced by their growth dynamics and mechanical properties. To address this question, I used methods from cell mechanics and materials science. I first studied how biological activity in biofilms gives rise to non-uniform growth patterns. In a second study, I investigated how E. coli biofilm morphogenesis and its mechanical properties adapt to an environmental stimulus, namely the water content of their substrate. Finally, I estimated how the mechanical properties of E. coli biofilms are altered when the bacteria express different extracellular biopolymers.
On nutritive hydrogels, micron-sized E. coli cells can build centimetre-large biofilms. During this process, bacterial proliferation and matrix production introduce mechanical stresses in the biofilm, which release through the formation of macroscopic wrinkles and delaminated buckles. To relate these biological and mechanical phenomena, I used time-lapse fluorescence imaging to track cell and matrix surface densities through the early and late stages of E. coli biofilm growth. Colocalization of high cell and matrix densities at the periphery precede the onset of mechanical instabilities at this annular region. Early growth is detected at this outer annulus, which was analysed by adding fluorescent microspheres to the bacterial inoculum. But only when high rates of matrix production are present in the biofilm centre, does overall biofilm spreading initiate along the solid-air interface. By tracking larger fluorescent particles for a long time, I could distinguish several kinematic stages of E. coli biofilm expansion and observed a transition from non-linear to linear velocity profiles, which precedes the emergence of wrinkles at the biofilm periphery. Decomposing particle velocities to their radial and circumferential components revealed a last kinematic stage, where biofilm movement is mostly directed towards the radial delaminated buckles, which verticalize. The resulting compressive strains computed in these regions were observed to substantially deform the underlying agar substrates. The co-localization of higher cell and matrix densities towards an annular region and the succession of several kinematic stages are thus expected to promote the emergence of mechanical instabilities at the biofilm periphery. These experimental findings are predicted to advance future modelling approaches of biofilm morphogenesis.
E. coli biofilm morphogenesis is further anticipated to depend on external stimuli from the environment. To clarify how the water could be used to tune biofilm material properties, we quantified E. coli biofilm growth, wrinkling dynamics and rigidity as a function of the water content of the nutritive substrates. Time-lapse microscopy and computational image analysis revealed that substrates with high water content promote biofilm spreading kinetics, while substrates with low water content promote biofilm wrinkling. The wrinkles observed on biofilm cross-sections appeared more bent on substrates with high water content, while they tended to be more vertical on substrates with low water content. Both wet and dry biomass, accumulated over 4 days of culture, were larger in biofilms cultured on substrates with high water content, despite extra porosity within the matrix layer. Finally, the micro-indentation analysis revealed that substrates with low water content supported the formation of stiffer biofilms. This study shows that E. coli biofilms respond to the water content of their substrate, which might be used for tuning their material properties in view of further applications.
Biofilm material properties further depend on the composition and structure of the matrix of extracellular proteins and polysaccharides. In particular, E. coli biofilms were suggested to present tissue-like elasticity due to a dense fibre network consisting of amyloid curli and phosphoethanolamine-modified cellulose. To understand the contribution of these components to the emergent mechanical properties of E. coli biofilms, we performed micro-indentation on biofilms grown from bacteria of several strains. Besides showing higher dry masses, larger spreading diameters and slightly reduced water contents, biofilms expressing both main matrix components also presented high rigidities in the range of several hundred kPa, similar to biofilms containing only curli fibres. In contrast, a lack of amyloid curli fibres provides much higher adhesive energies and more viscoelastic fluid-like material behaviour. Therefore, the combination of amyloid curli and phosphoethanolamine-modified cellulose fibres implies the formation of a composite material whereby the amyloid curli fibres provide rigidity to E. coli biofilms, whereas the phosphoethanolamine-modified cellulose rather acts as a glue. These findings motivate further studies involving purified versions of these protein and polysaccharide components to better understand how their interactions benefit biofilm functions.
All three studies depict different aspects of biofilm morphogenesis, which are interrelated. The first work reveals the correlation between non-uniform biological activities and the emergence of mechanical instabilities in the biofilm. The second work acknowledges the adaptive nature of E. coli biofilm morphogenesis and its mechanical properties to an environmental stimulus, namely water. Finally, the last study reveals the complementary role of the individual matrix components in the formation of a stable biofilm material, which not only forms complex morphologies but also functions as a protective shield for the bacteria it contains. Our experimental findings on E. coli biofilm morphogenesis and their mechanical properties can have further implications for fundamental and applied biofilm research fields.
Concerns have been raised that anthropogenic climate change could lead to large-scale singular climate events, i.e., abrupt nonlinear climate changes with repercussions on regional to global scales. One central goal of this thesis is the development of models of two representative components of the climate system that could exhibit singular behavior: the Atlantic thermohaline circulation (THC) and the Indian monsoon. These models are conceived so as to fulfill the main requirements of integrated assessment modeling, i.e., reliability, computational efficiency, transparency and flexibility. The model of the THC is an interhemispheric four-box model calibrated against data generated with a coupled climate model of intermediate complexity. It is designed to be driven by global mean temperature change which is translated into regional fluxes of heat and freshwater through a linear down-scaling procedure. Results of a large number of transient climate change simulations indicate that the reduced-form THC model is able to emulate key features of the behavior of comprehensive climate models such as the sensitivity of the THC to the amount, regional distribution and rate of change in the heat and freshwater fluxes. The Indian monsoon is described by a novel one-dimensional box model of the tropical atmosphere. It includes representations of the radiative and surface fluxes, the hydrological cycle and surface hydrology. Despite its high degree of idealization, the model satisfactorily captures relevant aspects of the observed monsoon dynamics, such as the annual course of precipitation and the onset and withdrawal of the summer monsoon. Also, the model exhibits the sensitivity to changes in greenhouse gas and sulfate aerosol concentrations that are known from comprehensive models. A simplified version of the monsoon model is employed for the identification of changes in the qualitative system behavior against changes in boundary conditions. The most notable result is that under summer conditions a saddle-node bifurcation occurs at critical values of the planetary albedo or insolation. Furthermore, the system exhibits two stable equilibria: besides the wet summer monsoon, a stable state exists which is characterized by a weak hydrological cycle. These results are remarkable insofar, as they indicate that anthropogenic perturbations of the planetary albedo such as sulfur emissions and/or land-use changes could destabilize the Indian summer monsoon. The reduced-form THC model is employed in an exemplary integrated assessment application. Drawing on the conceptual and methodological framework of the tolerable windows approach, emissions corridors (i.e., admissible ranges of CO2- emissions) are derived that limit the risk of a THC collapse while considering expectations about the socio-economically acceptable pace of emissions reductions. Results indicate, for example, a large dependency of the width of the emissions corridor on climate and hydrological sensitivity: for low values of climate and/or hydrological sensitivity, the corridor boundaries are far from being transgressed by any plausible emissions scenario for the 21st century. In contrast, for high values of both quantities low non-intervention scenarios leave the corridor already in the early decades of the 21st century. This implies that if the risk of a THC collapse is to be kept low, business-as-usual paths would need to be abandoned within the next two decades. All in all, this thesis highlights the value of reduced-form modeling by presenting a number of applications of this class of models, ranging from sensitivity and bifurcation analysis to integrated assessment. The results achieved and conclusions drawn provide a useful contribution to the scientific and policy debate about the consequences of anthropogenic climate change and the long-term goals of climate protection. --- Anmerkung: Die Autorin ist Trägerin des von der Mathematisch-Naturwissenschaftlichen Fakultät der Universität Potsdam vergebenen Michelson-Preises für die beste Promotion des Jahres 2003/2004.
The European Water Framework Directive (WFD) has identified river morphological alteration and diffuse pollution as the two main pressures affecting water bodies in Europe at the catchment scale. Consequently, river restoration has become a priority to achieve the WFD's objective of good ecological status. However, little is known about the effects of stream morphological changes, such as re-meandering, on in-stream nitrate retention at the river network scale. Therefore, catchment nitrate modeling is necessary to guide the implementation of spatially targeted and cost-effective mitigation measures. Meanwhile, Germany, like many other regions in central Europe, has experienced consecutive summer droughts from 2015-2018, resulting in significant changes in river nitrate concentrations in various catchments. However, the mechanistic exploration of catchment nitrate responses to changing weather conditions is still lacking.
Firstly, a fully distributed, process-based catchment Nitrate model (mHM-Nitrate) was used, which was properly calibrated and comprehensively evaluated at numerous spatially distributed nitrate sampling locations. Three calibration schemes were designed, taking into account land use, stream order, and mean nitrate concentrations, and they varied in spatial coverage but used data from the same period (2011–2019). The model performance for discharge was similar among the three schemes, with Nash-Sutcliffe Efficiency (NSE) scores ranging from 0.88 to 0.92. However, for nitrate concentrations, scheme 2 outperformed schemes 1 and 3 when compared to observed data from eight gauging stations. This was likely because scheme 2 incorporated a diverse range of data, including low discharge values and nitrate concentrations, and thus provided a better representation of within-catchment heterogenous. Therefore, the study suggests that strategically selecting gauging stations that reflect the full range of within-catchment heterogeneity is more important for calibration than simply increasing the number of stations.
Secondly, the mHM-Nitrate model was used to reveal the causal relations between sequential droughts and nitrate concentration in the Bode catchment (3200 km2) in central Germany, where stream nitrate concentrations exhibited contrasting trends from upstream to downstream reaches. The model was evaluated using data from six gauging stations, reflecting different levels of runoff components and their associated nitrate-mixing from upstream to downstream. Results indicated that the mHM-Nitrate model reproduced dynamics of daily discharge and nitrate concentration well, with Nash-Sutcliffe Efficiency ≥ 0.73 for discharge and Kling-Gupta Efficiency ≥ 0.50 for nitrate concentration at most stations. Particularly, the spatially contrasting trends of nitrate concentration were successfully captured by the model. The decrease of nitrate concentration in the lowland area in drought years (2015-2018) was presumably due to (1) limited terrestrial export loading (ca. 40% lower than that of normal years 2004-2014), and (2) increased in-stream retention efficiency (20% higher in summer within the whole river network). From a mechanistic modelling perspective, this study provided insights into spatially heterogeneous flow and nitrate dynamics and effects of sequential droughts, which shed light on water-quality responses to future climate change, as droughts are projected to be more frequent.
Thirdly, this study investigated the effects of stream restoration via re-meandering on in-stream nitrate retention at network-scale in the well-monitored Bode catchment. The mHM-Nitrate model showed good performance in reproducing daily discharge and nitrate concentrations, with median Kling-Gupta values of 0.78 and 0.74, respectively. The mean and standard deviation of gross nitrate retention efficiency, which accounted for both denitrification and assimilatory uptake, were 5.1 ± 0.61% and 74.7 ± 23.2% in winter and summer, respectively, within the stream network. The study found that in the summer, denitrification rates were about two times higher in lowland sub-catchments dominated by agricultural lands than in mountainous sub-catchments dominated by forested areas, with median ± SD of 204 ± 22.6 and 102 ± 22.1 mg N m-2 d-1, respectively. Similarly, assimilatory uptake rates were approximately five times higher in streams surrounded by lowland agricultural areas than in those in higher-elevation, forested areas, with median ± SD of 200 ± 27.1 and 39.1 ± 8.7 mg N m-2 d-1, respectively. Therefore, restoration strategies targeting lowland agricultural areas may have greater potential for increasing nitrate retention. The study also found that restoring stream sinuosity could increase net nitrate retention efficiency by up to 25.4 ± 5.3%, with greater effects seen in small streams. These results suggest that restoration efforts should consider augmenting stream sinuosity to increase nitrate retention and decrease nitrate concentrations at the catchment scale.
To what extent cities can be made sustainable under the mega-trends of urbanization and climate change remains a matter of unresolved scientific debate. Our inability in answering this question lies partly in the deficient knowledge regarding pivotal humanenvironment interactions. Regarded as the most well documented anthropogenic climate modification, the urban heat island (UHI) effect – the warmth of urban areas relative to the rural hinterland – has raised great public health concerns globally. Worse still, heat waves are being observed and are projected to increase in both frequency and intensity, which further impairs the well-being of urban dwellers. Albeit with a substantial increase in the number of publications on UHI in the recent decades, the diverse urban-rural definitions applied in previous studies have remarkably hampered the general comparability of results achieved. In addition, few studies have attempted to synergize the land use data and thermal remote sensing to systematically assess UHI and its contributing factors.
Given these research gaps, this work presents a general framework to systematically quantify the UHI effect based on an automated algorithm, whereby cities are defined as clusters of maximum spatial continuity on the basis of land use data, with their rural hinterland being defined analogously. By combining land use data with spatially explicit surface skin temperatures from satellites, the surface UHI intensity can be calculated in a consistent and robust manner. This facilitates monitoring, benchmarking, and categorizing UHI intensities for cities across scales. In light of this innovation, the relationship between city size and UHI intensity has been investigated, as well as the contributions of urban form indicators to the UHI intensity.
This work delivers manifold contributions to the understanding of the UHI, which have complemented and advanced a number of previous studies. Firstly, a log-linear relationship between surface UHI intensity and city size has been confirmed among the 5,000 European cities. The relationship can be extended to a log-logistic one, when taking a wider range of small-sized cities into account. Secondly, this work reveals a complex interplay between UHI intensity and urban form. City size is found to have the strongest influence on the UHI intensity, followed by the fractality and the anisometry. However, their relative contributions to the surface UHI intensity depict a pronounced regional heterogeneity, indicating the importance of considering spatial patterns of UHI while implementing UHI adaptation measures.
Lastly, this work presents a novel seasonality of the UHI intensity for individual clusters in the form of hysteresis-like curves, implying a phase shift between the time series of UHI intensity and background temperatures. Combining satellite observation and urban boundary layer simulation, the seasonal variations of UHI are assessed from both screen and skin levels. Taking London as an example, this work ascribes the discrepancies between the seasonality observed at different levels mainly to the peculiarities of surface skin temperatures associated with the incoming solar radiation. In addition, the efforts in classifying cities according to their UHI characteristics highlight the important role of regional climates in determining the UHI.
This work serves as one of the first studies conducted to systematically and statistically scrutinize the UHI. The outcomes of this work are of particular relevance for the overall spatial planning and regulation at meso- and macro levels in order to harness the benefits of rapid urbanization, while proactively minimizing its ensuing thermal stress.
Noise is ubiquitous in nature and usually results in rich dynamics in stochastic systems such as oscillatory systems, which exist in such various fields as physics, biology and complex networks. The correlation and synchronization of two or many oscillators are widely studied topics in recent years.
In this thesis, we mainly investigate two problems, i.e., the stochastic bursting phenomenon in noisy excitable systems and synchronization in a three-dimensional Kuramoto model with noise. Stochastic bursting here refers to a sequence of coherent spike train, where each spike has random number of followers due to the combined effects of both time delay and noise. Synchronization, as a universal phenomenon in nonlinear dynamical systems, is well illustrated in the Kuramoto model, a prominent model in the description of collective motion.
In the first part of this thesis, an idealized point process, valid if the characteristic timescales in the problem are well separated, is used to describe statistical properties such as the power spectral density and the interspike interval distribution. We show how the main parameters of the point process, the spontaneous excitation rate, and the probability to induce a spike during the delay action can be calculated from the solutions of a stationary and a forced Fokker-Planck equation. We extend it to the delay-coupled case and derive analytically the statistics of the spikes in each neuron, the pairwise correlations between any two neurons, and the spectrum of the total output from the network.
In the second part, we investigate the three-dimensional noisy Kuramoto model, which can be used to describe the synchronization in a swarming model with helical trajectory. In the case without natural frequency, the Kuramoto model can be connected with the Vicsek model, which is widely studied in collective motion and swarming of active matter. We analyze the linear stability of the incoherent state and derive the critical coupling strength above which the incoherent state loses stability. In the limit of no natural frequency, an exact self-consistent equation of the mean field is derived and extended straightforward to any high-dimensional case.
The plasmasphere is a dynamic region of cold, dense plasma surrounding the Earth. Its shape and size are highly susceptible to variations in solar and geomagnetic conditions. Having an accurate model of plasma density in the plasmasphere is important for GNSS navigation and for predicting hazardous effects of radiation in space on spacecraft. The distribution of cold plasma and its dynamic dependence on solar wind and geomagnetic conditions remain, however, poorly quantified. Existing empirical models of plasma density tend to be oversimplified as they are based on statistical averages over static parameters. Understanding the global dynamics of the plasmasphere using observations from space remains a challenge, as existing density measurements are sparse and limited to locations where satellites can provide in-situ observations. In this dissertation, we demonstrate how such sparse electron density measurements can be used to reconstruct the global electron density distribution in the plasmasphere and capture its dynamic dependence on solar wind and geomagnetic conditions.
First, we develop an automated algorithm to determine the electron density from in-situ measurements of the electric field on the Van Allen Probes spacecraft. In particular, we design a neural network to infer the upper hybrid resonance frequency from the dynamic spectrograms obtained with the Electric and Magnetic Field Instrument Suite and Integrated Science (EMFISIS) instrumentation suite, which is then used to calculate the electron number density. The developed Neural-network-based Upper hybrid Resonance Determination (NURD) algorithm is applied to more than four years of EMFISIS measurements to produce the publicly available electron density data set.
We utilize the obtained electron density data set to develop a new global model of plasma density by employing a neural network-based modeling approach. In addition to the location, the model takes the time history of geomagnetic indices and location as inputs, and produces electron density in the equatorial plane as an output. It is extensively validated using in-situ density measurements from the Van Allen Probes mission, and also by comparing the predicted global evolution of the plasmasphere with the global IMAGE EUV images of He+ distribution. The model successfully reproduces erosion of the plasmasphere on the night side as well as plume formation and evolution, and agrees well with data.
The performance of neural networks strongly depends on the availability of training data, which is limited during intervals of high geomagnetic activity. In order to provide reliable density predictions during such intervals, we can employ physics-based modeling. We develop a new approach for optimally combining the neural network- and physics-based models of the plasmasphere by means of data assimilation. The developed approach utilizes advantages of both neural network- and physics-based modeling and produces reliable global plasma density reconstructions for quiet, disturbed, and extreme geomagnetic conditions.
Finally, we extend the developed machine learning-based tools and apply them to another important problem in the field of space weather, the prediction of the geomagnetic index Kp. The Kp index is one of the most widely used indicators for space weather alerts and serves as input to various models, such as for the thermosphere, the radiation belts and the plasmasphere. It is therefore crucial to predict the Kp index accurately. Previous work in this area has mostly employed artificial neural networks to nowcast and make short-term predictions of Kp, basing their inferences on the recent history of Kp and solar wind measurements at L1. We analyze how the performance of neural networks compares to other machine learning algorithms for nowcasting and forecasting Kp for up to 12 hours ahead. Additionally, we investigate several machine learning and information theory methods for selecting the optimal inputs to a predictive model of Kp. The developed tools for feature selection can also be applied to other problems in space physics in order to reduce the input dimensionality and identify the most important drivers.
Research outlined in this dissertation clearly demonstrates that machine learning tools can be used to develop empirical models from sparse data and also can be used to understand the underlying physical processes. Combining machine learning, physics-based modeling and data assimilation allows us to develop novel methods benefiting from these different approaches.
Evaluation of nitrogen dynamics in high-order streams and rivers based on high-frequency monitoring
(2023)
Nutrient storage, transform and transport are important processes for achieving environmental and ecological health, as well as conducting water management plans. Nitrogen is one of the most noticeable elements due to its impacts on tremendous consequences of eutrophication in aquatic systems. Among all nitrogen components, researches on nitrate are blooming because of widespread deployments of in-situ high-frequency sensors. Monitoring and studying nitrate can become a paradigm for any other reactive substances that may damage environmental conditions and cause economic losses.
Identifying nitrate storage and its transport within a catchment are inspiring to the management of agricultural activities and municipal planning. Storm events are periods when hydrological dynamics activate the exchange between nitrate storage and flow pathways. In this dissertation, long-term high-frequency monitoring data at three gauging stations in the Selke river were used to quantify event-scale nitrate concentration-discharge (C-Q) hysteretic relationships. The Selke catchment is characterized into three nested subcatchments by heterogeneous physiographic conditions and land use. With quantified hysteresis indices, impacts of seasonality and landscape gradients on C-Q relationships are explored. For example, arable area has deep nitrate legacy and can be activated with high intensity precipitation during wetting/wet periods (i.e., the strong hydrological connectivity). Hence, specific shapes of C-Q relationships in river networks can identify targeted locations and periods for agricultural management actions within the catchment to decrease nitrate output into downstream aquatic systems like the ocean.
The capacity of streams for removing nitrate is of both scientific and social interest, which makes the quantification motivated. Although measurements of nitrate dynamics are advanced compared to other substances, the methodology to directly quantify nitrate uptake pathways is still limited spatiotemporally. The major problem is the complex convolution of hydrological and biogeochemical processes, which limits in-situ measurements (e.g., isotope addition) usually to small streams with steady flow conditions. This makes the extrapolation of nitrate dynamics to large streams highly uncertain. Hence, understanding of in-stream nitrate dynamic in large rivers is still necessary. High-frequency monitoring of nitrate mass balance between upstream and downstream measurement sites can quantitatively disentangle multi-path nitrate uptake dynamics at the reach scale (3-8 km). In this dissertation, we conducted this approach in large stream reaches with varying hydro-morphological and environmental conditions for several periods, confirming its success in disentangling nitrate uptake pathways and their temporal dynamics. Net nitrate uptake, autotrophic assimilation and heterotrophic uptake were disentangled, as well as their various diel and seasonal patterns. Natural streams generally can remove more nitrate under similar environmental conditions and heterotrophic uptake becomes dominant during post-wet seasons. Such two-station monitoring provided novel insights into reach-scale nitrate uptake processes in large streams.
Long-term in-stream nitrate dynamics can also be evaluated with the application of water quality model. This is among the first time to use a data-model fusion approach to upscale the two-station methodology in large-streams with complex flow dynamics under long-term high-frequency monitoring, assessing the in-stream nitrate retention and its responses to drought disturbances from seasonal to sub-daily scale. Nitrate retention (both net uptake and net release) exhibited substantial seasonality, which also differed in the investigated normal and drought years. In the normal years, winter and early spring seasons exhibited extensive net releases, then general net uptake occurred after the annual high-flow season at later spring and early summer with autotrophic processes dominating and during later summer-autumn low-flow periods with heterotrophy-characteristics predominating. Net nitrate release occurred since late autumn until the next early spring. In the drought years, the late-autumn net releases were not so consistently persisted as in the normal years and the predominance of autotrophic processes occurred across seasons. Aforementioned comprehensive results of nitrate dynamics on stream scale facilitate the understanding of instream processes, as well as raise the importance of scientific monitoring schemes for hydrology and water quality parameters.
Completely water-based systems are of interest for the development of novel material for various reasons: On one hand, they provide benign environment for biological systems and on the other hand they facilitate effective molecular transport in a membrane-free environment. In order to investigate the general potential of aqueous two-phase systems (ATPSs) for biomaterials and compartmentalized systems, various solid particles were applied to stabilize all-aqueous emulsion droplets. The target ATPS to be investigated should be prepared via mixing of two aqueous solutions of water-soluble polymers, which turn biphasic when exceeding a critical polymer concentration. Hydrophilic polymers with a wide range of molar mass such as dextran/poly(ethylene glycol) (PEG) can therefore be applied. Solid particles adsorbed at the interfaces can be exceptionally efficient stabilizers forming so-called Pickering emulsions, and nanoparticles can bridge the correlation length of polymer solutions and are thereby the best option for water-in-water emulsions.
The first approach towards the investigation of ATPS was conducted with all aqueous dextran-PEG emulsions in the presence of poly(dopamine) particles (PDP) in Chapter 4. The water-in-water emulsions were formed with a PEG/dextran system via utilizing PDP as stabilizers. Studies of the formed emulsions were performed via laser scanning confocal microscope (CLSM), optical microscope (OM), cryo-scanning electron microscope (SEM) and tensiometry. The stable emulsions (at least 16 weeks) were demulsified easily via dilution or surfactant addition. Furthermore, the solid PDP at the water-water interface were crosslinked in order to inhibit demulsification of the Pickering emulsion. Transmission electron microscope (TEM) and scanning electron microscope (SEM) were used to visualize the morphology of PDP before and after crosslinking. PDP stabilized water-in-water emulsions were utilized in the following Chapter 5 to form supramolecular compartmentalized hydrogels. Here, hydrogels were prepared in pre-formed water-in-water emulsions and gelled via α-cyclodextrin-PEG (α-CD-PEG) inclusion complex formation. Studies of the formed complexes were performed via X-ray powder diffraction (XRD) and the mechanical properties of the hydrogels were measured with oscillatory shear rheology. In order to verify the compartmentalized state and its triggered decomposition, hydrogels and emulsions were assessed via OM, SEM and CLSM. The last chapter broadens the investigations from the previous two systems by utilizing various carbon nitrides (CN) as different stabilizers in ATPS. CN introduces another way to trigger demulsification, namely irradiation with visible light. Therefore, emulsification and demulsification with various triggers were probed. The investigated all aqueous multi-phase systems will act as model for future fabrication of biocompatible materials, cell micropatterning as well as separation of compartmentalized systems.
Microfabricated solid-state surfaces, also called atom chip', have become a well-established technique to trap and manipulate atoms. This has simplified applications in atom interferometry, quantum information processing, and studies of many-body systems. Magnetic trapping potentials with arbitrary geommetries are generated with atom chip by miniaturized current-carrying conductors integrated on a solid substrate. Atoms can be trapped and cooled to microKelvin and even nanoKelvin temperatures in such microchip trap. However, cold atoms can be significantly perturbed by the chip surface, typically held at room temperature. The magnetic field fluctuations generated by thermal currents in the chip elements may induce spin flips of atoms and result in loss, heating and decoherence. In this thesis, we extend previous work on spin flip rates induced by magnetic noise and consider the more complex geometries that are typically encountered in atom chips: layered structures and metallic wires of finite cross-section. We also discuss a few aspects of atom chips traps built with superconducting structures that have been suggested as a means to suppress magnetic field fluctuations. The thesis describes calculations of spin flip rates based on magnetic Green functions that are computed analytically and numerically. For a chip with a top metallic layer, the magnetic noise depends essentially on the thickness of that layer, as long as the layers below have a much smaller conductivity. Based on this result, scaling laws for loss rates above a thin metallic layer are derived. A good agreement with experiments is obtained in the regime where the atom-surface distance is comparable to the skin depth of metal. Since in the experiments, metallic layers are always etched to separate wires carrying different currents, the impact of the finite lateral wire size on the magnetic noise has been taken into account. The local spectrum of the magnetic field near a metallic microstructure has been investigated numerically with the help of boundary integral equations. The magnetic noise significantly depends on polarizations above flat wires with finite lateral width, in stark contrast to an infinitely wide wire. Correlations between multiple wires are also taken into account. In the last part, superconducting atom chips are considered. Magnetic traps generated by superconducting wires in the Meissner state and the mixed state are studied analytically by a conformal mapping method and also numerically. The properties of the traps created by superconducting wires are investigated and compared to normal conducting wires: they behave qualitatively quite similar and open a route to further trap miniaturization, due to the advantage of low magnetic noise. We discuss critical currents and fields for several geometries.
This thesis aimed to investigate several fundamental and perplexing questions relating to the phloem loading and transport mechanisms of Cucurbita maxima, by combining metabolomic analysis with cell biological techniques. This putative symplastic loading species has long been used for experiments on phloem anatomy, phloem biochemistry, phloem transport physiology and phloem signalling. Symplastic loading species have been proposed to use a polymer trapping mechanism to accumulate RFO (raffinose family oligosaccharides) sugars to build up high osmotic pressure in minor veins which sustains a concentration gradient that drives mass flow. However, extensive evidence indicating a low sugar concentration in their phloem exudates is a long-known problem that conflicts with this hypothesis. Previous metabolomic analysis shows the concentration of many small molecules in phloem exudates is higher than that of leaf tissues, which indicates an active apoplastic loading step. Therefore, in the view of the phloem metabolome, a symplastic loading mechanism cannot explain how small molecules other than RFO sugars are loaded into phloem. Most studies of phloem physiology using cucurbits have neglected the possible functions of vascular architecture in phloem transport. It is well known that there are two phloem systems in cucurbits with distinctly different anatomical features: central phloem and extrafascicular phloem. However, mistaken conclusions on sources of cucurbit phloem exudation from previous reports have hindered consideration of the idea that there may be important differences between these two phloem systems. The major results are summarized as below: 1) O-linked glycans in C.maxima were structurally identified as beta-1,3 linked glucose polymers, and the composition of glycans in cucurbits was found to be species-specific. Inter-species grafting experiments proved that these glycans are phloem mobile and transported uni-directionally from scion to stock. 2) As indicated by stable isotopic labelling experiments, a considerable amount of carbon is incorporated into small metabolites in phloem exudates. However, the incorporation of carbon into RFO sugars is much faster than for other metabolites. 3) Both CO2 labelling experiments and comparative metabolomic analysis of phloem exudates and leaf tissues indicated that metabolic processes other than RFO sugar metabolism play an important role in cucurbit phloem physiology. 4) The underlying assumption that the central phloem of cucurbits continuously releases exudates after physical incision was proved wrong by rigorous experiments including direct observation by normal microscopy and combined multiple-microscopic methods. Errors in previous experimental confirmation of phloem exudation in cucurbits are critically discussed. 5) Extrafascicular phloem was proved to be functional, as indicated by phloem-mobile carboxyfluorescein tracer studies. Commissural sieve tubes interconnect phloem bundles into a complete super-symplastic network. 6) Extrafascicular phloem represents the main source of exudates following physical incision. The major transported metabolites by these extrafacicular phloem are non-sugar compounds including amino acids, O-glycans, amines. 7) Central phloem contains almost exclusively RFO sugars, the estimated amount of which is up to 1 to 2 molar. The major RFO sugar present in central phloem is stachyose. 8) Cucurbits utilize two structurally different phloem systems for transporting different group of metabolites (RFO sugars and non-RFO sugar compounds). This implies that cucurbits may use spatially separated loading mechanisms (apoplastic loading for extrafascicular phloem and symplastic loading for central phloem) for supply of nutrients to sinks. 9) Along the transport systems, RFO sugars were mainly distributed within central phloem tissues. There were only small amounts of RFO sugars present in xylem tissues (millimolar range) and trace amounts of RFO sugars in cortex and pith. The composition of small molecules in external central phloem is very different from that in internal central phloem. 10) Aggregated P-proteins were manually dissected from central phloem and analysed by both SDS-PAGE and mass spectrometry. Partial sequences of peptides were obtained by QTOF de novo sequencing from trypsin digests of three SDS-PAGE bands. None of these partial sequences shows significant homology to known cucurbit phloem proteins or other plant proteins. This proves that these central phloem proteins are a completely new group of proteins different from those in extrafascicular phloem. The extensively analysed P-proteins reported in literature to date are therefore now shown to arise from extrafascicular phloem and not central phloem, and therefore do not appear to be involved in the occlusion processes in central phloem.
In the present thesis I investigate the lattice dynamics of thin film hetero structures of magnetically ordered materials upon femtosecond laser excitation as a probing and manipulation scheme for the spin system. The quantitative assessment of laser induced thermal dynamics as well as generated picosecond acoustic pulses and their respective impact on the magnetization dynamics of thin films is a challenging endeavor. All the more, the development and implementation of effective experimental tools and comprehensive models are paramount to propel future academic and technological progress.
In all experiments in the scope of this cumulative dissertation, I examine the crystal lattice of nanoscale thin films upon the excitation with femtosecond laser pulses. The relative change of the lattice constant due to thermal expansion or picosecond strain pulses is directly monitored by an ultrafast X-ray diffraction (UXRD) setup with a femtosecond laser-driven plasma X-ray source (PXS). Phonons and spins alike exert stress on the lattice, which responds according to the elastic properties of the material, rendering the lattice a versatile sensor for all sorts of ultrafast interactions. On the one hand, I investigate materials with strong magneto-elastic properties; The highly magnetostrictive rare-earth compound TbFe2, elemental Dysprosium or the technological relevant Invar material FePt. On the other hand I conduct a comprehensive study on the lattice dynamics of Bi1Y2Fe5O12 (Bi:YIG), which exhibits high-frequency coherent spin dynamics upon femtosecond laser excitation according to the literature. Higher order standing spinwaves (SSWs) are triggered by coherent and incoherent motion of atoms, in other words phonons, which I quantified with UXRD. We are able to unite the experimental observations of the lattice and magnetization dynamics qualitatively and quantitatively. This is done with a combination of multi-temperature, elastic, magneto-elastic, anisotropy and micro-magnetic modeling.
The collective data from UXRD, to probe the lattice, and time-resolved magneto-optical Kerr effect (tr-MOKE) measurements, to monitor the magnetization, were previously collected at different experimental setups. To improve the precision of the quantitative assessment of lattice and magnetization dynamics alike, our group implemented a combination of UXRD and tr-MOKE in a singular experimental setup, which is to my knowledge, the first of its kind. I helped with the conception and commissioning of this novel experimental station, which allows the simultaneous observation of lattice and magnetization dynamics on an ultrafast timescale under identical excitation conditions. Furthermore, I developed a new X-ray diffraction measurement routine which significantly reduces the measurement time of UXRD experiments by up to an order of magnitude. It is called reciprocal space slicing (RSS) and utilizes an area detector to monitor the angular motion of X-ray diffraction peaks, which is associated with lattice constant changes, without a time-consuming scan of the diffraction angles with the goniometer. RSS is particularly useful for ultrafast diffraction experiments, since measurement time at large scale facilities like synchrotrons and free electron lasers is a scarce and expensive resource. However, RSS is not limited to ultrafast experiments and can even be extended to other diffraction techniques with neutrons or electrons.
The aim of this thesis is the quantum dynamical study of two examples of scanning tunneling microscope (STM)-controllable, Si(100)(2x1) surface-mounted switches of atomic and molecular scale. The first example considers the switching of single H-atoms between two dangling-bond chemisorption sites on a Si-dimer of the Si(100) surface (Grey et al., 1996). The second system examines the conformational switching of single 1,5-cyclooctadiene molecules chemisorbed on the Si(100) surface (Nacci et al., 2008). The temporal dynamics are provided by the propagation of the density matrix in time via an according set of equations of motion (EQM). The latter are based on the open-system density matrix theory in Lindblad form. First order perturbation theory is used to evaluate those transition rates between vibrational levels of the system part. In order to account for interactions with the surface phonons, two different dissipative models are used, namely the bilinear, harmonic and the Ohmic bath model. IET-induced vibrational transitions in the system are due to the dipole- and the resonance-mechanism. A single surface approach is used to study the influence of dipole scattering and resonance scattering in the below-threshold regime. Further, a second electronic surface was included to study the resonance-induced switching in the above-threshold regime. Static properties of the adsorbate, e.g., potentials and dipole function and potentials, are obtained from quantum chemistry and used within the established quantum dynamical models.
Die funktionelle Charakterisierung von therapeutisch relevanten Proteinen kann bereits durch die Bereitstellung des Zielproteins in adäquaten Mengen limitierend sein. Dies trifft besonders auf Membranproteine zu, die aufgrund von zytotoxischen Effekten auf die Produktionszelllinie und der Tendenz Aggregate zu bilden, in niedrigen Ausbeuten an aktivem Protein resultieren können. Der lebende Organismus kann durch die Verwendung von translationsaktiven Zelllysaten umgangen werden- die Grundlage der zellfreien Proteinsynthese. Zu Beginn der Arbeit wurde die ATP-abhängige Translation eines Lysates auf der Basis von kultivierten Insektenzellen (Sf21) analysiert. Für diesen Zweck wurde ein ATP-bindendes Aptamer eingesetzt, durch welches die Translation der Nanoluziferase reguliert werden konnte. Durch die dargestellte Applizierung von Aptameren, könnten diese zukünftig in zellfreien Systemen für die Visualisierung der Transkription und Translation eingesetzt werden, wodurch zum Beispiel komplexe Prozesse validiert werden können.
Neben der reinen Proteinherstellung können Faktoren wie posttranslationale Modifikationen sowie eine Integration in eine lipidische Membran essentiell für die Funktionalität des Membranproteins sein. Im zweiten Abschnitt konnte, im zellfreien Sf21-System, für den G-Protein-gekoppelten Rezeptor Endothelin B sowohl eine Integration in die endogen vorhandenen Endoplasmatisch Retikulum-basierten Membranstrukturen als auch Glykosylierungen, identifiziert werden.
Auf der Grundlage der erfolgreichen Synthese des ET-B-Rezeptors wurden verschiedene Methoden zur Fluoreszenzmarkierung des Adenosin-Rezeptors A2a (Adora2a) angewandt und optimiert. Im dritten Abschnitt wurde der Adora2a mit Hilfe einer vorbeladenen tRNA, welche an eine fluoreszierende Aminosäure gekoppelt war, im zellfreien Chinesischen Zwerghamster Ovarien (CHO)-System markiert. Zusätzlich konnte durch den Einsatz eines modifizierten tRNA/Aminoacyl-tRNA-Synthetase-Paares eine nicht-kanonische Aminosäure an Position eines integrierten Amber-Stopcodon in die Polypeptidkette eingebaut und die funktionelle Gruppe im Anschluss an einen Fluoreszenzfarbstoff gekoppelt werden. Aufgrund des offenen Charakters eignen sich zellfreie Proteinsynthesesysteme besonders für eine Integration von exogenen Komponenten in den Translationsprozess. Mit Hilfe der Fluoreszenzmarkierung wurde eine ligandvermittelte Konformationsänderung im Adora2a über einen Biolumineszenz-Resonanzenergietransfer detektiert. Durch die Etablierung der Amber-Suppression wurde darüber hinaus das Hormon Erythropoetin pegyliert, wodurch Eigenschaften wie Stabilität und Halbwertszeit des Proteins verändert wurden.
Zu guter Letzt wurde ein neues tRNA/Aminoacyl-tRNA-Synthetase-Paar auf Basis der Methanosarcina mazei Pyrrolysin-Synthetase etabliert, um das Repertoire an nicht-kanonischen Aminosäuren und den damit verbundenen Kopplungsreaktionen zu erweitern. Zusammenfassend wurden die Potenziale zellfreier Systeme in Bezug auf der Herstellung von komplexen Membranproteinen und der Charakterisierung dieser durch die Einbringung einer positionsspezifischen Fluoreszenzmarkierung verdeutlicht, wodurch neue Möglichkeiten für die Analyse und Funktionalisierung von komplexen Proteinen geschaffen wurden.
The mammalian brain is, with its numerous neural elements and structured complex connectivity, one of the most complex systems in nature. Recently, large-scale corticocortical connectivities, both structural and functional, have received a great deal of research attention, especially using the approach of complex networks. Here, we try to shed some light on the relationship between structural and functional connectivities by studying synchronization dynamics in a realistic anatomical network of cat cortical connectivity. We model the cortical areas by a subnetwork of interacting excitable neurons (multilevel model) and by a neural mass model (population model). With weak couplings, the multilevel model displays biologically plausible dynamics and the synchronization patterns reveal a hierarchical cluster organization in the network structure. We can identify a group of brain areas involved in multifunctional tasks by comparing the dynamical clusters to the topological communities of the network. With strong couplings of multilevel model and by using neural mass model, the dynamics are characterized by well-defined oscillations. The synchronization patterns are mainly determined by the node intensity (total input strengths of a node); the detailed network topology is of secondary importance. The biologically improved multilevel model exhibits similar dynamical patterns in the two regimes. Thus, the study of synchronization in a multilevel complex network model of cortex can provide insights into the relationship between network topology and functional organization of complex brain networks.
The Greenland Ice Sheet is the second-largest mass of ice on Earth. Being almost 2000 km long, more than 700 km wide, and more than 3 km thick at the summit, it holds enough ice to raise global sea levels by 7m if melted completely. Despite its massive size, it is particularly vulnerable to anthropogenic climate change: temperatures over the Greenland Ice Sheet have increased by more than 2.7◦C in the past 30 years, twice as much as the global mean temperature. Consequently, the ice sheet has been significantly losing mass since the 1980s and the rate of loss has increased sixfold since then. Moreover, it is one of the potential tipping elements of the Earth System, which might undergo irreversible change once a warming threshold is exceeded. This thesis aims at extending the understanding of the resilience of the Greenland Ice Sheet against global warming by analyzing processes and feedbacks relevant to its centennial to multi-millennial stability using ice sheet modeling.
One of these feedbacks, the melt-elevation-feedback is driven by the temperature rise with decreasing altitudes: As the ice sheet melts, its thickness and surface elevation decrease, exposing the ice surface to warmer air and thus increasing the melt rates even further. The glacial isostatic adjustment (GIA) can partly mitigate this melt-elevation feedback as the bedrock lifts in response to an ice load decrease, forming the negative GIA feedback. In my thesis, I show that the interaction between these two competing feedbacks can lead to qualitatively different dynamical responses of the Greenland Ice Sheet to warming – from permanent loss to incomplete recovery, depending on the feedback parameters. My research shows that the interaction of those feedbacks can initiate self-sustained oscillations of the ice volume while the climate forcing remains constant.
Furthermore, the increased surface melt changes the optical properties of the snow or ice surface, e.g. by lowering their albedo, which in turn enhances melt rates – a process known as the melt-albedo feedback. Process-based ice sheet models often neglect this melt-albedo feedback. To close this gap, I implemented a simplified version of the diurnal Energy Balance Model, a computationally efficient approach that can capture the first-order effects of the melt-albedo feedback, into the Parallel Ice Sheet Model (PISM). Using the coupled model, I show in warming experiments that the melt-albedo feedback almost doubles the ice loss until the year 2300 under the low greenhouse gas emission scenario RCP2.6, compared to simulations where the melt-albedo feedback is neglected,
and adds up to 58% additional ice loss under the high emission scenario RCP8.5. Moreover, I find that the melt-albedo feedback dominates the ice loss until 2300, compared to the melt-elevation feedback.
Another process that could influence the resilience of the Greenland Ice Sheet is the warming induced softening of the ice and the resulting increase in flow. In my thesis, I show with PISM how the uncertainty in Glen’s flow law impacts the simulated response to warming. In a flow line setup at fixed climatic mass balance, the uncertainty in flow parameters leads to a range of ice loss comparable to the range caused by different warming levels.
While I focus on fundamental processes, feedbacks, and their interactions in the first three projects of my thesis, I also explore the impact of specific climate scenarios on the sea level rise contribution of the Greenland Ice Sheet. To increase the carbon budget flexibility, some warming scenarios – while still staying within the limits of the Paris Agreement – include a temporal overshoot of global warming. I show that an overshoot by 0.4◦C increases the short-term and long-term ice loss from Greenland by several centimeters. The long-term increase is driven by the warming at high latitudes, which persists even when global warming is reversed. This leads to a substantial long-term commitment of the sea level rise contribution from the Greenland Ice Sheet.
Overall, in my thesis I show that the melt-albedo feedback is most relevant for the ice loss of the Greenland Ice Sheet on centennial timescales. In contrast, the melt-elevation feedback and its interplay with the GIA feedback become increasingly relevant on millennial timescales. All of these influence the resilience of the Greenland Ice Sheet against global warming, in the near future and on the long term.
In dieser Arbeit wurden Zusammenhänge zwischen den sexuellen Erfahrungen junger Frauen und Männer, ihren Persönlichkeitseigenschaften und ihren sexualmoralischen Einstellungen auf der einen Seite und der Einschätzung ihrer sexuellen Handlungsfähigkeit auf der anderen Seite untersucht. Die Grundlage für das Modell der sexuellen Handlungsfähigkeit bildeten die Vorstellungen der Arbeitsgruppe um Matthias Grundmann (Grundmann et al. 2006) sowie von Emirbayer und Mische (1998). Das in dieser Arbeit entwickelte Modell zur sexuellen Handlungsfähigkeit ist ein multidimensionales Konstrukt, das sich aus den Komponenten „sexuelle Kommunikation“, „sexuelle Zufriedenheit“, „sexuelle Reziprozität“ sowie „sexuelle Eigenverantwortung“ zusammensetzt. „Sexuelle Kommunikation“ beinhaltet die Fähigkeit, sexuelle Wünsche zum Ausdruck bringen zu können. „Sexuelle Zufriedenheit“ beschreibt den Grad der Zufriedenheit mit dem eigenen Sexualleben. „Sexuelle Reziprozität“ verweist auf die Fähigkeit, sexuelle Aufmerksamkeiten sowohl Annehmen als auch Geben zu können. „Sexuelle Eigenverantwortung“ betont schließlich die Einschätzung, inwieweit die eigene Sexualität selbst bestimmt gestaltet werden kann. Mit Emirbayer und Mische werden die sexuellen Erfahrungen der Frauen und Männer als Korrelate der Einschätzung der Dimensionen der sexuellen Handlungsfähigkeit betrachtet. Mit Grundmann et al. sind es zudem verschiedene Persönlichkeitseigenschaften sowie sexualmoralische Einstellungen, deren Beschaffenheiten Aussagen über die sexuelle Handlungsfähigkeit erlauben. Um die Thematik der sexuellen Handlungsfähigkeit empirisch zu betrachten, wurden im Jahr 2006 695 junge Potsdamer/innen im Alter von 19 bis 21 Jahren im Rahmen einer standardisierten Erhebung zu ihren sexuellen und Beziehungserfahrungen befragt. Die empirischen Analysen verdeutlichen eine ko-konstruktive Anschauung von der Entwicklung sexueller Handlungsfähigkeit. Diese entsteht nicht im Individuum allein, sondern innerhalb der Interaktions- und Aushandlungsprozesse des Individuums mit den Anderen seiner sozialen und sexuellen Umwelt. Von Bedeutung erweisen dabei sowohl die Erlebnisse der sexuellen Biografie als auch die Persönlichkeitsmerkmale eines jeden Einzelnen. Nur geringfügig erscheinen die erfragten sexualmoralischen Ansichten von Bedeutung.
Seismological and seismotectonic analysis of the northwestern Argentine Central Andean foreland
(2020)
After a severe M W 5.7 earthquake on October 17, 2015 in El Galpón in the province of Salta NW Argentina, I installed a local seismological network around the estimated epicenter. The network covered an area characterized by inherited Cretaceous normal faults and neotectonic faults with unknown recurrence intervals, some of which may have been reactivated normal faults. The 13 three-component seismic stations recorded data continuously for 15 months.
The 2015 earthquake took place in the Santa Bárbara System of the Andean foreland, at about 17km depth. This region is the easternmost morphostructural region of the central Andes. As a part of the broken foreland, it is bounded to the north by the Subandes fold-and-thrust belt and the Sierras Pampeanas to the south; to the east lies the Chaco-Paraná basin.
A multi-stage morphotectonic evolution with thick-skinned basement uplift and coeval thin-skinned deformation in the intermontane basins is suggested for the study area. The release of stresses associated with the foreland deformation can result in strong earthquakes, as the study area is known for recurrent and historical, destructive earthquakes. The available continuous record reaches back in time, when the strongest event in 1692 (magnitude 7 or intensity IX) destroyed the city of Esteco. Destructive earthquakes and surface deformation are thus a hallmark of this part of the Andean foreland.
With state-of-the-art Python packages (e.g. pyrocko, ObsPy), a semi-automatic approach is followed to analyze the collected continuous data of the seismological network. The resulting 1435 hypocenter locations consist of three different groups: 1.) local crustal earthquakes (nearly half of the events belong to this group), 2.) interplate activity, of regional distance in the slab of the Nazca-plate, and 3.) very deep earthquakes at about 600km depth. My major interest focused on the first event class. Those crustal events are partly aftershock events of the El Galpón earthquake and a second earthquake, in the south of the same fault. Further events can be considered as background seismicity of other faults within the study area. Strikingly, the seismogenic zone encompass the whole crust and propagates brittle deformation down, close to the Moho.
From the collected seismological data, a local seismic velocity model is estimated, using VELEST. After the execution of various stability tests, the robust minimum 1D-velocity model implies guiding values for the composition of the local, subsurface structure of the crust. Afterwards, performing a hypocenter relocation enables the assignment of individual earthquakes to aftershock clusters or extended seismotectonic structures. This allows the mapping of previously unknown seismogenic faults.
Finally, focal mechanisms are modeled for events with acurately located hypocenters, using the newly derived local velocity model. A compressive regime is attested by the majority of focal mechanisms, while the strike direction of the individual seismogenic structures is in agreement with the overall north – south orientation of the Central Andes, its mountain front, and individual mountain ranges in the southern Santa-Bárbara-System.
This work describes the realization of physically crosslinked networks based on gelatin by the introduction of functional groups enabling specific supramolecular interactions. Molecular models were developed in order to predict the material properties and permit to establish a knowledge-based approach to material design. The effect of additional supramolecular interactions with hydroxyapaptite was then studied in composite materials. The calculated properties are compared to experimental results to validate the models. The models are then further used for the study of physically crosslinked networks. Gelatin was functionalized with desaminotyrosine (DAT) and desaminotyrosyl-tyrosine (DATT) side groups, derived from the natural amino acid tyrosine. These group can potentially undergo to π-π and hydrogen bonding interactions also under physiological conditions. Molecular dynamics (MD) simulations were performed on models with 0.8 wt.-% or 25 wt.-% water content, using the second generation forcefield CFF91. The validation of the models was obtained by the comparison with specific experimental data such as, density, peptide conformational angles and X-ray scattering spectra. The models were then used to predict the supramolecular organization of the polymer chain, analyze the formation of physical netpoints and calculate the mechanical properties. An important finding of simulation was that with the increase of aromatic groups also the number of observed physical netpoints increased. The number of relatively stable physical netpoints, on average zero 0 for natural gelatin, increased to 1 and 6 for DAT and DATT functionalized gelatins respectively. A comparison with the Flory-Rehner model suggested reduced equilibrium swelling by factor 6 of the DATT-functionalized materials in water. The functionalized gelatins could be synthesized by chemoselective coupling of the free carboxylic acid groups of DAT and DATT to the free amino groups of gelatin. At 25 wt.-% water content, the simulated and experimentally determined elastic mechanical properties (e.g. Young Modulus) were both in the order of GPa and were not influenced by the degree of aromatic modification. The experimental equilibrium degree of swelling in water decreased with increasing the number of inserted aromatic functions (from 2800 vol.-% for pure gelatin to 300 vol.-% for the DATT modified gelatin), at the same time, Young’s modulus, elongation at break, and maximum tensile strength increased. It could be show that the functionalization with DAT and DATT influences the chain organization of gelatin based materials together with a controlled drying condition. Functionalization with DAT and DATT lead to a drastic reduction of helical renaturation, that could be more finely controlled by the applied drying conditions. The properties of the materials could then be influenced by application of two independent methods. Composite materials of DAT and DATT functionalized gelatins with hydroxyapatite (HAp) show a drastic reduction of swelling degree. In tensile tests and rheological measurements, the composites equilibrated in water had increased Young’s moduli (from 200 kPa up to 2 MPa) and tensile strength (from 57 kPa up to 1.1 MPa) compared to the natural polymer matrix without affecting the elongation at break. Furthermore, an increased thermal stability from 40 °C to 85 °C of the networks could be demonstrated. The differences of the behaviour of the functionalized gelatins to pure gelatin as matrix suggested an additional stabilizing bond between the incorporated aromatic groups to the hydroxyapatite.
This thesis focuses on the study of marked Gibbs point processes, in particular presenting some results on their existence and uniqueness, with ideas and techniques drawn from different areas of statistical mechanics: the entropy method from large deviations theory, cluster expansion and the Kirkwood--Salsburg equations, the Dobrushin contraction principle and disagreement percolation.
We first present an existence result for infinite-volume marked Gibbs point processes. More precisely, we use the so-called entropy method (and large-deviation tools) to construct marked Gibbs point processes in R^d under quite general assumptions. In particular, the random marks belong to a general normed space S and are not bounded. Moreover, we allow for interaction functionals that may be unbounded and whose range is finite but random. The entropy method relies on showing that a family of finite-volume Gibbs point processes belongs to sequentially compact entropy level sets, and is therefore tight.
We then present infinite-dimensional Langevin diffusions, that we put in interaction via a Gibbsian description. In this setting, we are able to adapt the general result above to show the existence of the associated infinite-volume measure. We also study its correlation functions via cluster expansion techniques, and obtain the uniqueness of the Gibbs process for all inverse temperatures β and activities z below a certain threshold. This method relies in first showing that the correlation functions of the process satisfy a so-called Ruelle bound, and then using it to solve a fixed point problem in an appropriate Banach space. The uniqueness domain we obtain consists then of the model parameters z and β for which such a problem has exactly one solution.
Finally, we explore further the question of uniqueness of infinite-volume Gibbs point processes on R^d, in the unmarked setting. We present, in the context of repulsive interactions with a hard-core component, a novel approach to uniqueness by applying the discrete Dobrushin criterion to the continuum framework. We first fix a discretisation parameter a>0 and then study the behaviour of the uniqueness domain as a goes to 0. With this technique we are able to obtain explicit thresholds for the parameters z and β, which we then compare to existing results coming from the different methods of cluster expansion and disagreement percolation.
Throughout this thesis, we illustrate our theoretical results with various examples both from classical statistical mechanics and stochastic geometry.
Taking advantage of ATRP and using functionalized initiators, different functionalities were introduced in both α and ω chain-ends of synthetic polymers. These functionalized polymers could then go through modular synthetic pathways such as click cycloaddition (copper-catalyzed or copper-free) or amidation to couple synthetic polymers to other synthetic polymers, biomolecules or silica monoliths. Using this general strategy and designing these co/polymers so that they are thermoresponsive, yet bioinert and biocompatible with adjustable cloud point values (as it is the case in the present thesis), the whole generated system becomes "smart" and potentially applicable in different branches. The applications which were considered in the present thesis were in polymer post-functionalization (in situ functionalization of micellar aggregates with low and high molecular weight molecules), hydrophilic/hydrophobic tuning, chromatography and bioconjugation (enzyme thermoprecipitation and recovery, improvement of enzyme activity). Different α-functionalized co/polymers containing cholesterol moiety, aldehyde, t-Boc protected amine, TMS-protected alkyne and NHS-activated ester were designed and synthesized in this work.
Interactions and feedbacks between tectonics, climate, and upper plate architecture control basin geometry, relief, and depositional systems. The Andes is part of a longlived continental margin characterized by multiple tectonic cycles which have strongly modified the Andean upper plate architecture. In the Andean retroarc, spatiotemporal variations in the structure of the upper plate and tectonic regimes have resulted in marked along-strike variations in basin geometry, stratigraphy, deformational style, and mountain belt morphology. These along-strike variations include high-elevation plateaus (Altiplano and Puna) associated with a thin-skin fold-and-thrust-belt and thick-skin deformation in broken foreland basins such as the Santa Barbara system and the Sierras Pampeanas. At the confluence of the Puna Plateau, the Santa Barbara system and the Sierras Pampeanas, major along-strike changes in upper plate architecture, mountain belt morphology, basement exhumation, and deformation style can be recognized. I have used a source to sink approach to unravel the spatiotemporal tectonic evolution of the Andean retroarc between 26 and 28°S. I obtained a large low-temperature thermochronology data set from basement units which includes apatite fission track, apatite U-Th-Sm/He, and zircon U-Th/He (ZHe) cooling ages. Stratigraphic descriptions of Miocene units were temporally constrained by U-Pb LA-ICP-MS zircon ages from interbedded pyroclastic material.
Modeled ZHe ages suggest that the basement of the study area was exhumed during the Famatinian orogeny (550-450 Ma), followed by a period of relative tectonic quiescence during the Paleozoic and the Triassic. The basement experienced horst exhumation during the Cretaceous development of the Salta rift. After initial exhumation, deposition of thick Cretaceous syn-rift strata caused reheating of several basement blocks within the Santa Barbara system. During the Eocene-Oligocene, the Andean compressional setting was responsible for the exhumation of several disconnected basement blocks. These exhumed blocks were separated by areas of low relief, in which humid climate and low erosion rates facilitated the development of etchplains on the crystalline basement. The exhumed basement blocks formed an Eocene to Oligocene broken foreland basin in the back-bulge depozone of the Andean foreland. During the Early Miocene, foreland basin strata filled up the preexisting Paleogene topography. The basement blocks in lower relief positions were reheated; associated geothermal gradients were higher than 25°C/km. Miocene volcanism was responsible for lateral variations on the amount of reheating along the Campo-Arenal basin. Around 12 Ma, a new deformational phase modified the drainage network and fragmented the lacustrine system. As deformation and rock uplift continued, the easily eroded sedimentary cover was efficiently removed and reworked by an ephemeral fluvial system, preventing the development of significant relief. After ~6 Ma, the low erodibility of the basement blocks which began to be exposed caused relief increase, leading to the development of stable fluvial systems. Progressive relief development modified atmospheric circulation, creating a rainfall gradient. After 3 Ma, orographic rainfall and high relief lead to the development of proximal fluvial-gravitational depositional systems in the surrounding basins.
The recent discovery of an intricate and nontrivial interaction topology among the elements of a wide range of natural systems has altered the manner we understand complexity. For example, the axonal fibres transmitting electrical information between cortical regions form a network which is neither regular nor completely random. Their structure seems to follow functional principles to balance between segregation (functional specialisation) and integration. Cortical regions are clustered into modules specialised in processing different kinds of information, e.g. visual or auditory. However, in order to generate a global perception of the real world, the brain needs to integrate the distinct types of information. Where this integration happens, nobody knows. We have performed an extensive and detailed graph theoretical analysis of the cortico-cortical organisation in the brain of cats, trying to relate the individual and collective topological properties of the cortical areas to their function. We conclude that the cortex possesses a very rich communication structure, composed of a mixture of parallel and serial processing paths capable of accommodating dynamical processes with a wide variety of time scales. The communication paths between the sensory systems are not random, but largely mediated by a small set of areas. Far from acting as mere transmitters of information, these central areas are densely connected to each other, strongly indicating their functional role as integrators of the multisensory information. In the quest of uncovering the structure-function relationship of cortical networks, the peculiarities of this network have led us to continuously reconsider the stablished graph measures. For example, a normalised formalism to identify the “functional roles” of vertices in networks with community structure is proposed. The tools developed for this purpose open the door to novel community detection techniques which may also characterise the overlap between modules. The concept of integration has been revisited and adapted to the necessities of the network under study. Additionally, analytical and numerical methods have been introduced to facilitate understanding of the complicated statistical interrelations between the distinct network measures. These methods are helpful to construct new significance tests which may help to discriminate the relevant properties of real networks from side-effects of the evolutionary-growth processes.